"This is a relaxed question guys", "Every time you see this just close your eyes and pick this option guys!", "What is this answer!" Love your personality!
Question 10: Shouldn't the answer be B and then D? The question states that "during upload requests the website resizes the images to a standard size" & "must design the most operationally efficient process for image uploads". By not resizing during upload, we are essentially improving the slow upload requests by uploading the original images to amazon s3 and then using notifications to resize versus doing it all during the upload. Wouldn't C in this case NOT BE the most operationally efficient process, since it will require to configure the application to upload directly from the browser and generate a presigned URL for each user? From the question the social media company is already allowing users to upload images to its website so users being able to upload is already taken care off, its the speed which we need to take care off. Thank you appreciate your work!
B. Configure the web server to upload the original images to Amazon S3:Partially Suitable: This option could help reduce the workload on the EC2 instances, but it still keeps the web server involved in the upload process. It doesn't fully decouple the process or improve user experience significantly.
For Question 8, the answer should be D. The option that you chose to use SQS won’t work for multiple targets. Once a message is consumed by a target it will be removed from the queue
I am also a bit confused here, because what does 'multiple target systems' mean here? Is it one system with distributed processing? In which case SQS can work. Or is multiple systems where each system needs to process all the messages? In which case SQS will not work and D is the correct option. But I think here it means one system with distributed processing.
RDS Event notification doesn't constitute an event that deletes the item from the db. You can subscribe to an event category for the following resources: DB instance DB snapshot DB parameter group DB security group RDS Proxy Custom engine version
This is my first UA-cam comment! Your video greatly helped me to pass the exam and kudos to your wonderful contribution to the tech community.😊Keep up the good work!!!
This teacher is just Awesome :-) These videos are Amazing and sure shot for you to pass the Exam. I just passed the exam :-) I took test recently and almost 90% of the questions are same or very similar. It gives opportunity to practically review your preparation and the way it is explained is just perfect. I like the way you said "Option Elimination theory" this is the key and bottoline, else you end up wasting your time ... Thanks
Hello thanks a lot for your videos, it help a lot. You rock 🎉 I think for question 25, it's answer A & E not A & D as said. Because to deploy 2 NAT gateways across 2 AZs you need 2 public subnets sinc a subnet belongs to only one AZ
@@peaceofcode I don't think it's D. It's clearly has to be E since the question mentions high availabilty. You'll need 2 private, 1 in each AZ to make sure your application is available. That is also why you need 2 NAT, each NAT on a public subnet, so 1 AZ down mean the NAT and everything will be down with it
Thank you so much for these 150 Questions. I just cleared the exam today. Out of 65 questios 20 were same wording to wording from these 150 questions and if you understand the tricks which were told in these practice questions of 150, you will definately clear the certification. Thanks for these video.
Hello, thank you so much for great tutorials, I have a note about question 8 here. Since it saying that data must be sent to multiple traget systems I assume that "Fan out" pattern can be used here to send same data to multiple consumers via SNS SQS combo. Whereas in case of A data will be consumed by the first consumer and deleted so others will not be able to consume same data. I think D is a correct option in this case. I have exam in a few days. Please comment this. Thanks in advance.
Please keep up the good work bro, your excellent approach towards explaining complex AWS topics is unmatched. I strongly advise aspirants to go through your detailed videos, to build confidence and get a nice grip on key topics.
I think you got question 10 wrong. Presigned URLs are only used to give users temporary access to files without an AWS account. Should be upload images to S3.
For Q 34, I think the answer is C. The question states LEAST operational overhead, doesn't choosing KMS keys and managing it be overhead instead of using S3-SSE ??
Question 7. the answer is D, not B, because you cannot ingest and store data points using Lambda, we need to use Kinesis Data Analytics, it can ingest data and not only store the data points but can expose them as REST API
Kinesis is not needed as the company is using it's own Analytical Platform; So it will be an overkill. But Lambda should store and retrieve data from somewhere for the it be accessed using API is slightly ignored too :)
These are SO helpful! I spent the last few days going through your practice question series and passed my exam today! Thank you so much for making these!
For Q27: *Option A* is better because *EBS* provides *persistent high-performance storage* that ensures data is not lost during the video processing, whereas *Option D* uses *instance store, which is too risky for this use case due to its **ephemeral* nature. - In critical workloads like video processing, where data loss is not acceptable, *persistent storage* (EBS) is essential, making *Option A* the clear winner.
Question #25 should have correct answer A & E as VPC subnet can't span across AZ and we need to achieve High Availability using Two AZ so its required two subnets (one in each AZ).
Q.25 If you pick D or E then you are configuring both a public and private subnet. Option A doesn't use the public subnet and using a private one isn't required; thus I think it should be C and E. D doesn't make sense as other people already mentioned, since subnets don't span across availability zones but VPC's do
Thank you so much for these wonderful videos which helped me pass the certificate today. I really appreciate you taking the time to out this together. Please keep up the good work. If it was not for these videos and explanations i would have been a bit lost on a few questions which popped up. Thank you again very much!
Hi, looks like the answer for the Question:25 is not correct. You said its D. Can you please explain ? With one private subnet, how can you get high availability ? Because a subnet cant span across multiple availability zones. So, use of autoscaling group doesn't make any sense if you deploy your ec2 instances in only one subnet. And why do you need 2 NAT Gateways when they are managed by aws ?
Yes You are right.. By mistake I did not see the one public and private subnet in the option.... the last one will be the correct one, now regarding the usage of two NAT gateways its required to have high availability, you can refer to this article for further reading. www.packetswitch.co.uk/aws-nat-gateway-high-availability/#:~:text=NAT%20Gateway%20is%20Highly%20Available,Availability%20Zones%20lose%20Internet%20access.
for Q.38 isn't d the option with lesser operational overhead providing high availabilty ,where you just need to provide multi-Az groups for the existing infrastructure instead of migrating
You are doing excellent job bro... I really gone through your full series of questions and it was great practice with great explanation. By the way I passed the exam (AWS-SSA-C03). Keep going!!!
@@johnyart862 followed the technical explanation and practice set questions after going through tutorials for every service, which alone is not enough without knowing how to interpret the questions and then find out the correct option out of it. Horizon of AWS services is large and the way questions are asked in exam makes it tricker. Its good to solve as much as you can but with proper understanding and explanations why one option is correct and more importantly other incorrect options to discard. This video have all of them.
Question 25... I think answer should be A and E.. not D.. if you have one private subnet and one public subnet then what is the need for two NAT gateways. NAt gateway can only be attached to public subnet so that it can access the internet.
For question 7 Answer is D, Not B. For each and every minutes or 5 min data needs to be uploaded through kinesis streams. From there we can invoke Rest API to get the data.
Thank you sir Aakash! I just received my AWS Architect Associate badge just a few seconds ago!! I felt pretty confident during the exam because of this series. Thank you very much!
Question 45, I think option B should be the correct answer because the questions asks for the most COST-EFFECTIVE solution, having a second instance plus weighted routing is cheaper than auto scaling group + load balancer.
I took my exam yesterday and I passed your videos really helped, I got about 8 similar questions on the exam. Thank you so much and keep up with the good work
Question 21: The correct answer is C. Answer A is not correct, because it is not possible to encrypt an existing snapshot and the existing database instance can not be replcaed by the restored snapshot. Instead the unencrypted snapshot must be copied and encryption has to be enabled using KMS and then a new encrypted database instance can be restored from the encrypted snapshot.
I said B. To me it sounds like the question is asking us to encrypt all previous snapshots, I thought the "moving forward" only applied to encrypting the current DB. I dont get C but B makes sense to me and I think it would work.
Question 10: I think the answer should be B and then D? The question states that "during upload requests the website resizes the images to a standard size" & "must design the most operationally efficient process for image uploads". By removing the requirement for the website to resize and allowing the lambda function to do it after uploading it to the bucket ,we are essentially improving the slow upload requests by uploading the original images to amazon s3 and then using notifications to resize versus doing it all during the upload. Wouldn't C in this case NOT BE the most operationally efficient process, since it will require to configure the application to upload directly from the browser and generate a presigned URL for each user? From the question the social media company is already allowing users to upload images to its website so users being able to upload is already taken care off, its the speed which we need to take care off. Thank you appreciate your work!
Question no 45 is most cost only to meet that requirement, when you go for Aurora it's mor e cost compared with RDS. Question is just meet this requirement with cost effective not the high performance or highly available...
Thanks Aakash, your videos provided me the foundations and adequate knowledge for my to pass through the AWS Solutions Architect Associate certification. Thanks once again. Regards Ralph
Q43 answer B makes more sense, company wants to break application into smaller apps, no mention about containers, AWS Amplify can integrate with API Gateway and Lambda
Incorrect > B. AWS Amplify with Lambda and API Gateway: AWS Amplify is more focused on front-end web and mobile applications. While it can integrate with Lambda and API Gateway, it’s not as suitable for managing a backend application broken into microservices. Correct > D. Breaking into Smaller Applications: Amazon ECS allows you to run containerized applications, making it easier to break down a monolithic application into smaller microservices. Each microservice can be managed by different teams, aligning with the company’s requirement to have different teams manage different parts of the application.
@peaceofcode Q 38: if we need to consider "least operational overhead" then the answer should be D. Option B creates operational overhead to migrate to Amazon MQ and RDS. what do you think?
For question #2 I believe the answers are A and E. The AWS DataSync agent is run in the cloud " AWS DataSync agents run in AMS as an appliance." and If you don't install the EC2 in the same Availabiliy Zone as the EFS you are charged.
I would like to Thank you , I cleared my SAA-C03 exam today. I prepared well with all the concepts and this video boosted my confidence to appear for the exam . Thank you very much
All the Pros are already mentioned in the comments. Let me write some Cons (or constructive feedback) about your videos: 1. Please don't rush, I know you have to cover a lot. 2. There are instances where you have not clarified why a certain option is NOT correct. 3. This has generated user questions below stating a different choice of an answer than yours.
In question 7, doesn't location data correspond to the live location? I thought in that manner and selected D. Your thoughts? Rethink about question 25. Correct option should be A and E.
Bro this is very good thing you are doing to help people crack the exam. It is very helpful for me and others also and thank you so much. The major draw back i had is that you had to explain things a bit deep so that we would get it properly and need not worry.
Great explanation ! some doubts: Q-20: Global Acc deals with IP related issues, Like IP caching? But in Question asks for Routing Policy at end, ans make sense but then need to rephrase the Q I guess .. Q-35: Option D, in the last part it says "Private Link'' in that case: "AWS PrivateLink provides private connections between VPCs, AWS services and your on-premises networks, without exposing your traffic over the public internet." But in Question there is no indication about the "on-premises" network.
Global Accelerator does not only deal with IP related issues it can work with any kind of load, the difference is that Global Accelerator provides with 2 static ip addresses and its better for layer 4 traffic, for question 35 pls read the line "Connection must be initiated only from company's vpc". I hope it cleared your doubts!!
@@peaceofcode Also for Q 20, where is says self-managed DNS solution. I assume that means 3rd party DNS solution. So that should eliminate anything that references Route 53
Your videos and tips are fantastic. I passed the SAA-C03 yesterday. With some questions I could still hear your comments. Thank you for all the help. Please do a series on the professional exam😁
Just passed my AWS Certified Solutions Architect associate Certification. Thank you Aakash for these questions I watched all your solution architect associate videos also did with you all the 150 questions and they helped me a lot thank you🙏
@@kashifshaikh1293 it's very useful and helps you understand answering of questions but you should practice more on yourself also. I watched Aakash videos they made me understand how to approach questions and also keywords then I went to exam topics and also try answering every question that comes my way
Honestly i got help from Andrew Brown, shaping pixel nd a few others. But especially you and Andrew Brown made it possible. Wish i had money to give to you- maybe soon
Question 43 does not state anything about "Containerization" or "Containers". How were you able to decipher they were needing containers in this question?
For Q8: Option D* is the most suitable solution because it efficiently uses AWS services to ensure that when the RDS database is updated, the data is broadcast to multiple target systems through SNS, with each target system receiving the data via its respective SQS queue. This design is aligned with best practices for scalable and decoupled architectures in AWS. Option D:* - The correct answer. - This option involves subscribing to *RDS event notifications, which directly trigger an **SNS topic. The SNS topic then fans out to multiple **SQS queues*. Each SQS queue can be processed by different AWS Lambda functions or target systems. - *Why it’s best:* It directly meets the requirement by leveraging SNS's ability to broadcast messages to multiple SQS queues, ensuring that data is sent to multiple target systems. Additionally, this setup is more scalable and efficient for handling multiple targets.
Yes it could have been but its not required here first of all we are not analysing or transforming the data, the word real time will throw you off but there is no data ingestion & analytics pipeline here so kinesis data analytics is not required.
Some of the answers are so even. Like question 8. I don't know how to choose option A over option D. I also thought multiple consumers can poll from a queue but when one does the item is on visibility timeout. So the rest of the targets wouldn't be able to get the item.
@@peaceofcode I am sorry for doubting you. But could you check question 8. I don't think it is possible to send data to multiple target systems in a fanned out pattern with just SQS. I believe D is the answer. using SNS to send to multiple SQSs where each target system can get a copy of the data seems perfect. Although multiple consumers can poll the same queue. when one consumer receives the poll the rest of the consumers cannot receive that event anymore.
In Question 37 the question asks to limit the email to admins only. So choosing alternate contact list make the root user still receive the notification. Then how can B be the option ?
Question 40: A and E, We should choose E because it offers more discount up to 72% like you said, however compute plan offers a discount of only 60%, so Question 40 answers should be changed! Thanks.
Q.8 How is just SQS on its own sufficient to deliver the SAME MESSAGE to MULTIPLE TARGETS ? An SQS message is meant to be processed by just one consumer. There can be multiple consumers to a queue for sure but even in that case, they do not consume the same message; they are meant to process different messages mapped to them through message attributes so that a consumer can know which of the messages in the queue is allotted to it. There is no such thing as "SQS fan-out pattern"; use SNS instead. Q.10 How is D more performance-efficient than E ? The question states that the users are facing slow uploads already. By resizing images as soon as they are uploaded (as in option D), the performance is inhibited further in comparison to resizing all the images together later on a schedule, e.g., during off-peak hours (as in option E). Q.17 The question mentions "near -real time" as a requirement. The answer should be C. Q.34 The data needs to be encrypted and replicated across different regions. Nowhere in the question is it mentioned that the buckets should be encrypted with the SAME key. Moreover, the requirement is LEAST operational overhead. AWS-KMS keys most definitely require more setup and management effort from the customer than S3-Managed keys. Option C seems like a more appropriate choice.
Also, the answer (B) for the Question : 26 is partially true. Because, it is applicable only for the future . The question says that there is already data available in standard storage. Hence, the right answer should be to keep the data from the recent 2 years in the standard storage and the remaining data should be moved immediately to the deep glacier.
When you create a S3 lifecycle policy it gets applied to existing as well as future objects you can refer to the link here.... docs.aws.amazon.com/AmazonS3/latest/userguide/how-to-set-lifecycle-configuration-intro.html And some advice I want to give if you are appearing for the exam..., please dont assume things if not asked in the question, strictly follow the question even if it dosent make any sense, you dont wanna assume something wrong and get the answer wrong. Best of Luck!! Cheers!!
Question 25 (48:52): Option E is better than D as it meets the high availability requirement. Option D if the single subnet (AZ) fails the application becomes unavailable.
For Q7: *Option D: Amazon API Gateway with Amazon Kinesis Data Analytics* - *API Gateway* still provides the REST API layer. - *Amazon Kinesis Data Analytics* is designed specifically for real-time processing of streaming data. It can handle large volumes of data in real-time, making it ideal for scenarios where you need to continuously monitor and analyze the location of bicycles during peak hours. This option is better suited for handling the continuous data streams and real-time analytics needed in this scenario. It aligns more closely with the requirement to process and make accessible the location data during peak operating times, leveraging a multi-tier architecture. ### Summary: - *B* is suitable for event-driven compute tasks and handling API requests, but it lacks built-in capabilities for handling real-time, continuous data streams, which is critical here. - *D* provides a more robust solution for the scenario, as it directly supports real-time data processing and analysis, which is essential for tracking bicycles during peak times. Therefore, *D* is more aligned with the requirements of the architecture needed in the question.
For Q25: *A + B* is Correct: 1. *Private Subnets Only:* - Both *A* and *B* ensure that *EC2 and RDS instances are in private subnets*, meaning they are not exposed to the public internet, which satisfies the requirement for privacy and security. 2. *High Availability:* - *A* deploys the EC2 instances in private subnets and uses an RDS Multi-AZ DB instance, ensuring high availability for the database. - *B* configures two private subnets and two NAT gateways across two Availability Zones, which ensures that even if one NAT gateway fails, the other can provide internet access to the EC2 instances. This redundancy is crucial for high availability. 3. *NAT Gateway for Internet Access:* - *B* ensures that the EC2 instances have internet access through NAT gateways, allowing them to connect to third-party services (like for payment processing), without exposing them to inbound traffic from the internet. ### Why *A + D* or *A + Second D* are Incorrect: 1. *Public Subnets:* - Both versions of *D* include *public subnets, which could expose instances or components to the internet. This violates the requirement of keeping the EC2 and RDS instances **private*. 2. *Limited Subnet Configurations:* - The first *D* has only *one private subnet*, which introduces a single point of failure, reducing availability. - The second *D* includes *two public subnets*, which again raises the risk of public exposure, going against the requirement. ### Conclusion: - *A* + *B* meets all the requirements of the question: privacy, high availability, and internet access for EC2 instances. - *A + D* and *A + second D* introduce public exposure and other risks, making them less suitable for the scenario. Therefore, *A + B* is the best combination for this question.
"This is a relaxed question guys", "Every time you see this just close your eyes and pick this option guys!", "What is this answer!" Love your personality!
Thank you!!
Question 10: Shouldn't the answer be B and then D? The question states that "during upload requests the website resizes the images to a standard size" & "must design the most operationally efficient process for image uploads". By not resizing during upload, we are essentially improving the slow upload requests by uploading the original images to amazon s3 and then using notifications to resize versus doing it all during the upload. Wouldn't C in this case NOT BE the most operationally efficient process, since it will require to configure the application to upload directly from the browser and generate a presigned URL for each user? From the question the social media company is already allowing users to upload images to its website so users being able to upload is already taken care off, its the speed which we need to take care off. Thank you appreciate your work!
Agreed - B&D is the correct answer
exactly, cause the question there is nothing related to security here
agreed! B&D. you somehow decouple by uploading the image directly to S3 and then use S3 Event notification w/ Lambda to process it.
But using pre-signed will take away the load from the webserver completely. So it's much better.
B. Configure the web server to upload the original images to Amazon S3:Partially Suitable: This option could help reduce the workload on the EC2 instances, but it still keeps the web server involved in the upload process. It doesn't fully decouple the process or improve user experience significantly.
Pass the SAA exam today with 751 thanks to your videos.
This has been a blessing to me
For Question 8, the answer should be D. The option that you chose to use SQS won’t work for multiple targets. Once a message is consumed by a target it will be removed from the queue
I am also a bit confused here, because what does 'multiple target systems' mean here?
Is it one system with distributed processing? In which case SQS can work.
Or is multiple systems where each system needs to process all the messages? In which case SQS will not work and D is the correct option.
But I think here it means one system with distributed processing.
I agree with D
@peaceofcode - can you confirm?
agree with D. counterchecked with some experienced fellows.
RDS Event notification doesn't constitute an event that deletes the item from the db.
You can subscribe to an event category for the following resources:
DB instance
DB snapshot
DB parameter group
DB security group
RDS Proxy
Custom engine version
I agree - D would be the correct answer here.
Thank you so much for this 4-part series! This absolutely saved my life :D I was able to clear the AWS SAA-C03 today!
How many questions come from this
@@aaddiis45021 What else you studied ? My exam is scheduled on 6th Nov
@@rajsegodia4545updates?
This is my first UA-cam comment! Your video greatly helped me to pass the exam and kudos to your wonderful contribution to the tech community.😊Keep up the good work!!!
r u serious? just these 200 questions helped u to pass exam? :)
This teacher is just Awesome :-) These videos are Amazing and sure shot for you to pass the Exam. I just passed the exam :-)
I took test recently and almost 90% of the questions are same or very similar.
It gives opportunity to practically review your preparation and the way it is explained is just perfect.
I like the way you said "Option Elimination theory" this is the key and bottoline, else you end up wasting your time ...
Thanks
thanks for the comment!
Hello thanks a lot for your videos, it help a lot. You rock 🎉
I think for question 25, it's answer A & E not A & D as said.
Because to deploy 2 NAT gateways across 2 AZs you need 2 public subnets sinc a subnet belongs to only one AZ
Thank you, and okay ill recheck that question
@@peaceofcode I don't think it's D. It's clearly has to be E since the question mentions high availabilty. You'll need 2 private, 1 in each AZ to make sure your application is available. That is also why you need 2 NAT, each NAT on a public subnet, so 1 AZ down mean the NAT and everything will be down with it
Yes. High availability required 2 AZs and that would mean 2 pub subnets and 2 pvt subnets.
Thank you so much for these 150 Questions. I just cleared the exam today. Out of 65 questios 20 were same wording to wording from these 150 questions and if you understand the tricks which were told in these practice questions of 150, you will definately clear the certification.
Thanks for these video.
Hello, thank you so much for great tutorials,
I have a note about question 8 here. Since it saying that data must be sent to multiple traget systems I assume that "Fan out" pattern can be used here to send same data to multiple consumers via SNS SQS combo. Whereas in case of A data will be consumed by the first consumer and deleted so others will not be able to consume same data. I think D is a correct option in this case.
I have exam in a few days. Please comment this. Thanks in advance.
Hi, I also have the same opinion here
Question 13 D is answer no operational overhead. So move the exiting transformation job to ec2?
I just passed my exams today. Your explanations here gave me the deepest understanding and made all the difference. Thank you so much
Question 10: Presigned URLs are for granting time-limited access to objects which isn't relevant to the question, so B & D?
Please keep up the good work bro, your excellent approach towards explaining complex AWS topics is unmatched. I strongly advise aspirants to go through your detailed videos, to build confidence and get a nice grip on key topics.
I think you got question 10 wrong. Presigned URLs are only used to give users temporary access to files without an AWS account. Should be upload images to S3.
Agree!
Just verified, we can upload files using pre-signed URL using PUT method. But question didn't ask any secure method to upload, still debatable
💯 agreed
For Q 34, I think the answer is C. The question states LEAST operational overhead, doesn't choosing KMS keys and managing it be overhead instead of using S3-SSE ??
Because this option don't describe about multi-region like answer A using KMS-SSE. It's a trick.
You can’t replicate existing s3 bucket object in another region if you apply CRR on existing bucket
A is the correct ans for Q34
Question 7.
the answer is D, not B, because you cannot ingest and store data points using Lambda, we need to use Kinesis Data Analytics, it can ingest data and not only store the data points but can expose them as REST API
I also think so
question said the company wants to use these data points in it existing analytics platform.
@@GentoFX1 agreed this is what I also missed initially , the existing analytics platform
I am not sure if Lambda can store data
Kinesis is not needed as the company is using it's own Analytical Platform; So it will be an overkill. But Lambda should store and retrieve data from somewhere for the it be accessed using API is slightly ignored too :)
These are SO helpful! I spent the last few days going through your practice question series and passed my exam today! Thank you so much for making these!
Is really helpful for you? Am going to take exam in Jan 1st week?...concept is same?
For Q27:
*Option A* is better because *EBS* provides *persistent high-performance storage* that ensures data is not lost during the video processing, whereas *Option D* uses *instance store, which is too risky for this use case due to its **ephemeral* nature.
- In critical workloads like video processing, where data loss is not acceptable, *persistent storage* (EBS) is essential, making *Option A* the clear winner.
Question #25 should have correct answer A & E as VPC subnet can't span across AZ and we need to achieve High Availability using Two AZ so its required two subnets (one in each AZ).
it is E, he got it wrong
Q.25 If you pick D or E then you are configuring both a public and private subnet. Option A doesn't use the public subnet and using a private one isn't required; thus I think it should be C and E. D doesn't make sense as other people already mentioned, since subnets don't span across availability zones but VPC's do
Good Video! I have the exam in 3 days! Going thru these videos and hoping for the best!
Thank you so much for these wonderful videos which helped me pass the certificate today. I really appreciate you taking the time to out this together. Please keep up the good work. If it was not for these videos and explanations i would have been a bit lost on a few questions which popped up. Thank you again very much!
I have cleared the exam recently. Your videos are very helpful.
Thank you man, I was able to pass the test today and I’m grateful for this channel.. The tension was real
Hi, looks like the answer for the Question:25 is not correct. You said its D. Can you please explain ? With one private subnet, how can you get high availability ? Because a subnet cant span across multiple availability zones. So, use of autoscaling group doesn't make any sense if you deploy your ec2 instances in only one subnet. And why do you need 2 NAT Gateways when they are managed by aws ?
Yes You are right.. By mistake I did not see the one public and private subnet in the option.... the last one will be the correct one, now regarding the usage of two NAT gateways its required to have high availability, you can refer to this article for further reading.
www.packetswitch.co.uk/aws-nat-gateway-high-availability/#:~:text=NAT%20Gateway%20is%20Highly%20Available,Availability%20Zones%20lose%20Internet%20access.
for Q.38 isn't d the option with lesser operational overhead providing high availabilty ,where you just need to provide multi-Az groups for the existing infrastructure instead of migrating
#36 should be "C" and "E"....the EventBridge can be used to monitor synchronization?
You are doing excellent job bro... I really gone through your full series of questions and it was great practice with great explanation. By the way I passed the exam (AWS-SSA-C03). Keep going!!!
Did u follow this series only?
Thanks man!!
@@johnyart862 followed the technical explanation and practice set questions after going through tutorials for every service, which alone is not enough without knowing how to interpret the questions and then find out the correct option out of it. Horizon of AWS services is large and the way questions are asked in exam makes it tricker. Its good to solve as much as you can but with proper understanding and explanations why one option is correct and more importantly other incorrect options to discard. This video have all of them.
@@dipdav4434 can you provide list of resources you think are most valueable for the exam? I am about to take the exam
Passed today, 782 , thanks to you! Although we can argue about some answers, you done incredible job. Thanks!
Thank you! Your questions, explanations, and tips have been a huge help
Thanks!! Glad I was of help
Question 25... I think answer should be A and E.. not D.. if you have one private subnet and one public subnet then what is the need for two NAT gateways. NAt gateway can only be attached to public subnet so that it can access the internet.
D- is one AZ solution. So not scalable. E- is correct one
Thank you @peace of code of these exams Q&A videos, it has helped me a lot to clear my exam on 14 July.
For question 7 Answer is D, Not B. For each and every minutes or 5 min data needs to be uploaded through kinesis streams. From there we can invoke Rest API to get the data.
Thanks so much!! Because of your videos I was able to pass my exam!
For q21, isnt option C correct? As they need encrypted snapshot and database going forward
Thank you sir Aakash! I just received my AWS Architect Associate badge just a few seconds ago!! I felt pretty confident during the exam because of this series. Thank you very much!
Great job
Question 45,
I think option B should be the correct answer because the questions asks for the most COST-EFFECTIVE solution, having a second instance plus weighted routing is cheaper than auto scaling group + load balancer.
I think so
I took my exam yesterday and I passed your videos really helped, I got about 8 similar questions on the exam. Thank you so much and keep up with the good work
Great job!
only 8/ 200 ?
😊Many thanks! i have cleared the SAA-CO3 exam yesterday with 755. You are an awesome human being. God will bless you.
Question 21: The correct answer is C. Answer A is not correct, because it is not possible to encrypt an existing snapshot and the existing database instance can not be replcaed by the restored snapshot. Instead the unencrypted snapshot must be copied and encryption has to be enabled using KMS and then a new encrypted database instance can be restored from the encrypted snapshot.
I said B. To me it sounds like the question is asking us to encrypt all previous snapshots, I thought the "moving forward" only applied to encrypting the current DB. I dont get C but B makes sense to me and I think it would work.
Thank you! Please is this a new set of questions?
Yes this is a new set and it will be a series of 4 videos!!
Ok. Thank you!
Question 10: I think the answer should be B and then D? The question states that "during upload requests the website resizes the images to a standard size" & "must design the most operationally efficient process for image uploads". By removing the requirement for the website to resize and allowing the lambda function to do it after uploading it to the bucket ,we are essentially improving the slow upload requests by uploading the original images to amazon s3 and then using notifications to resize versus doing it all during the upload. Wouldn't C in this case NOT BE the most operationally efficient process, since it will require to configure the application to upload directly from the browser and generate a presigned URL for each user? From the question the social media company is already allowing users to upload images to its website so users being able to upload is already taken care off, its the speed which we need to take care off. Thank you appreciate your work!
Question no 45 is most cost only to meet that requirement, when you go for Aurora it's mor e cost compared with RDS. Question is just meet this requirement with cost effective not the high performance or highly available...
Thanks Aakash, your videos provided me the foundations and adequate knowledge for my to pass through the AWS Solutions Architect Associate certification. Thanks once again. Regards Ralph
Q43 answer B makes more sense, company wants to break application into smaller apps, no mention about containers, AWS Amplify can integrate with API Gateway and Lambda
Incorrect > B. AWS Amplify with Lambda and API Gateway: AWS Amplify is more focused on front-end web and mobile applications. While it can integrate with Lambda and API Gateway, it’s not as suitable for managing a backend application broken into microservices.
Correct > D. Breaking into Smaller Applications: Amazon ECS allows you to run containerized applications, making it easier to break down a monolithic application into smaller microservices. Each microservice can be managed by different teams, aligning with the company’s requirement to have different teams manage different parts of the application.
@peaceofcode
Q 38: if we need to consider "least operational overhead" then the answer should be D. Option B creates operational overhead to migrate to Amazon MQ and RDS. what do you think?
Thank you so much for your videos, passed the exam today with 899 score. Got so many questions from all your videos combined
For question #2 I believe the answers are A and E. The AWS DataSync agent is run in the cloud " AWS DataSync agents run in AMS as an appliance." and If you don't install the EC2 in the same Availabiliy Zone as the EFS you are charged.
it helped me to understand the exam question pattern. thank you
I would like to Thank you , I cleared my SAA-C03 exam today. I prepared well with all the concepts and this video boosted my confidence to appear for the exam . Thank you very much
Hello, can you explain in question 25, why are you not choosing option E,
I think the answer to question 7 is D as they mentioned existing analytics platform
I agree and location tracking I feel is better as a stream which is better suited for KDA.
Agree
I agree….API gateway with kinesis data analytics.usually kinesis is for real time data stream and location service is just that.
Thank you so much for your insightful videos!
🙏🙏
💐
Question 25 A and E should be the correct answer not A and D
Question 34 C is the good answer
For question 7, I think correct answer must be C, the question is storing and retrieving data, in answer B, where are you store the data
All the Pros are already mentioned in the comments. Let me write some Cons (or constructive feedback) about your videos:
1. Please don't rush, I know you have to cover a lot.
2. There are instances where you have not clarified why a certain option is NOT correct.
3. This has generated user questions below stating a different choice of an answer than yours.
In question 7, doesn't location data correspond to the live location? I thought in that manner and selected D. Your thoughts? Rethink about question 25. Correct option should be A and E.
#7 the company has its own Data Analytics software AND KDA can only ingest data from KDS & KDF
Thank u peace of code this video really helps me to clear and pass my exam it helps to understand in a easy way to crack question
Thanks to cover such a must needed topic for us.
Bro this is very good thing you are doing to help people crack the exam. It is very helpful for me and others also and thank you so much. The major draw back i had is that you had to explain things a bit deep so that we would get it properly and need not worry.
Thanks and welcome
Great explanation !
some doubts:
Q-20: Global Acc deals with IP related issues, Like IP caching? But in Question asks for Routing Policy at end, ans make sense but then need to rephrase the Q I guess ..
Q-35: Option D, in the last part it says "Private Link'' in that case: "AWS PrivateLink provides private connections between VPCs, AWS services and your on-premises networks, without exposing your traffic over the public internet." But in Question there is no indication about the "on-premises" network.
Global Accelerator does not only deal with IP related issues it can work with any kind of load, the difference is that Global Accelerator provides with 2 static ip addresses and its better for layer 4 traffic, for question 35 pls read the line "Connection must be initiated only from company's vpc". I hope it cleared your doubts!!
@@peaceofcode Also for Q 20, where is says self-managed DNS solution. I assume that means 3rd party DNS solution. So that should eliminate anything that references Route 53
Were of these questions on the SAA exam or similarly worded questions?
Your videos and tips are fantastic. I passed the SAA-C03 yesterday. With some questions I could still hear your comments. Thank you for all the help. Please do a series on the professional exam😁
Just passed my AWS Certified Solutions Architect associate Certification. Thank you Aakash for these questions I watched all your solution architect associate videos also did with you all the 150 questions and they helped me a lot thank you🙏
Congratulations, good work!!
Hi wisem is this 150 question is enough for crack the exam
@@kashifshaikh1293 it's very useful and helps you understand answering of questions but you should practice more on yourself also. I watched Aakash videos they made me understand how to approach questions and also keywords then I went to exam topics and also try answering every question that comes my way
question 16 for high scalability we should use auto scaling ?? right?
Honestly i got help from Andrew Brown, shaping pixel nd a few others. But especially you and Andrew Brown made it possible. Wish i had money to give to you- maybe soon
Passed on 2nd August, Thanks Aakash to you and these helpful videos saw many question from these videos.
Congratulations!! And thank you!!
Question 43 does not state anything about "Containerization" or "Containers". How were you able to decipher they were needing containers in this question?
For Q8:
Option D* is the most suitable solution because it efficiently uses AWS services to ensure that when the RDS database is updated, the data is broadcast to multiple target systems through SNS, with each target system receiving the data via its respective SQS queue. This design is aligned with best practices for scalable and decoupled architectures in AWS.
Option D:*
- The correct answer.
- This option involves subscribing to *RDS event notifications, which directly trigger an **SNS topic. The SNS topic then fans out to multiple **SQS queues*. Each SQS queue can be processed by different AWS Lambda functions or target systems.
- *Why it’s best:* It directly meets the requirement by leveraging SNS's ability to broadcast messages to multiple SQS queues, ensuring that data is sent to multiple target systems. Additionally, this setup is more scalable and efficient for handling multiple targets.
Thanks for you TIPS and tricks I passed the exam with 810
Question 45, it requires most cost effective solution Aurora is very expensive. RDS option A is the answer
Question no 35 I think answer should be A. Because company vpc and provider vpc can only be connected through vpc peering.
Great video.. thanks a lot..awesome explanation in a few minutes ...superb..
For question 7 shouldn't D be the correct answer as data needs to be tracked real time?
Yes it could have been but its not required here first of all we are not analysing or transforming the data, the word real time will throw you off but there is no data ingestion & analytics pipeline here so kinesis data analytics is not required.
Just passed today! Thank you!
Some of the answers are so even. Like question 8. I don't know how to choose option A over option D. I also thought multiple consumers can poll from a queue but when one does the item is on visibility timeout. So the rest of the targets wouldn't be able to get the item.
Its okay…. It will improve with practice keep practicing
@@peaceofcode I am sorry for doubting you. But could you check question 8. I don't think it is possible to send data to multiple target systems in a fanned out pattern with just SQS. I believe D is the answer. using SNS to send to multiple SQSs where each target system can get a copy of the data seems perfect. Although multiple consumers can poll the same queue. when one consumer receives the poll the rest of the consumers cannot receive that event anymore.
@@surajmsdyou're right, in order to receive the message with multiple targets you need to use sns + fane out sqs combination
Passed the exam!! Got 83/100!
@@surajmsddid you just use this resource or other resources ?
Passed the examination few days back and must say these 200 videos are helpful. great job. Will we see similar videos on AWS SAP exam soon?
In Question 37 the question asks to limit the email to admins only. So choosing alternate contact list make the root user still receive the notification.
Then how can B be the option ?
Question 40: A and E, We should choose E because it offers more discount up to 72% like you said, however compute plan offers a discount of only 60%, so Question 40 answers should be changed!
Thanks.
Let me check!!
Compute Savings Plan also includes Lambda and Fargate, so A and C are correct answers
The Best , Thanks for the amazing work =)
Thank you!!
Thank you so much for this awsome series.
#25 - correct answer is A & E. The option D with just 1 private subnet is not highly available
Please release parts 2-4. I have exams and want to study with your videos. Thank you.
As soon as possible
Q.8 How is just SQS on its own sufficient to deliver the SAME MESSAGE to MULTIPLE TARGETS ? An SQS message is meant to be processed by just one consumer. There can be multiple consumers to a queue for sure but even in that case, they do not consume the same message; they are meant to process different messages mapped to them through message attributes so that a consumer can know which of the messages in the queue is allotted to it. There is no such thing as "SQS fan-out pattern"; use SNS instead.
Q.10 How is D more performance-efficient than E ? The question states that the users are facing slow uploads already. By resizing images as soon as they are uploaded (as in option D), the performance is inhibited further in comparison to resizing all the images together later on a schedule, e.g., during off-peak hours (as in option E).
Q.17 The question mentions "near -real time" as a requirement. The answer should be C.
Q.34 The data needs to be encrypted and replicated across different regions. Nowhere in the question is it mentioned that the buckets should be encrypted with the SAME key. Moreover, the requirement is LEAST operational overhead. AWS-KMS keys most definitely require more setup and management effort from the customer than S3-Managed keys. Option C seems like a more appropriate choice.
sir what about qes no 10;: presigned url has timelimite max 7days so b and d is correct. case of social media website there is a lot of users
Muje question 7 m doubt ho raha h tracking bicycle v real time data h na.. Phir lambda q
Also, the answer (B) for the Question : 26 is partially true. Because, it is applicable only for the future . The question says that there is already data available in standard storage. Hence, the right answer should be to keep the data from the recent 2 years in the standard storage and the remaining data should be moved immediately to the deep glacier.
When you create a S3 lifecycle policy it gets applied to existing as well as future objects you can refer to the link here....
docs.aws.amazon.com/AmazonS3/latest/userguide/how-to-set-lifecycle-configuration-intro.html
And some advice I want to give if you are appearing for the exam..., please dont assume things if not asked in the question, strictly follow the question even if it dosent make any sense, you dont wanna assume something wrong and get the answer wrong.
Best of Luck!!
Cheers!!
@@peaceofcode Thank you
for those who took exam after learning from these video, how close is this to real exam? what is the source of these questions?
Hi Sir,
I want all this question & answer. Please share the link to download all Q&A document.
Question 25 (48:52): Option E is better than D as it meets the high availability requirement. Option D if the single subnet (AZ) fails the application becomes unavailable.
For Q7:
*Option D: Amazon API Gateway with Amazon Kinesis Data Analytics*
- *API Gateway* still provides the REST API layer.
- *Amazon Kinesis Data Analytics* is designed specifically for real-time processing of streaming data. It can handle large volumes of data in real-time, making it ideal for scenarios where you need to continuously monitor and analyze the location of bicycles during peak hours.
This option is better suited for handling the continuous data streams and real-time analytics needed in this scenario. It aligns more closely with the requirement to process and make accessible the location data during peak operating times, leveraging a multi-tier architecture.
### Summary:
- *B* is suitable for event-driven compute tasks and handling API requests, but it lacks built-in capabilities for handling real-time, continuous data streams, which is critical here.
- *D* provides a more robust solution for the scenario, as it directly supports real-time data processing and analysis, which is essential for tracking bicycles during peak times.
Therefore, *D* is more aligned with the requirements of the architecture needed in the question.
For Q25:
*A + B* is Correct:
1. *Private Subnets Only:*
- Both *A* and *B* ensure that *EC2 and RDS instances are in private subnets*, meaning they are not exposed to the public internet, which satisfies the requirement for privacy and security.
2. *High Availability:*
- *A* deploys the EC2 instances in private subnets and uses an RDS Multi-AZ DB instance, ensuring high availability for the database.
- *B* configures two private subnets and two NAT gateways across two Availability Zones, which ensures that even if one NAT gateway fails, the other can provide internet access to the EC2 instances. This redundancy is crucial for high availability.
3. *NAT Gateway for Internet Access:*
- *B* ensures that the EC2 instances have internet access through NAT gateways, allowing them to connect to third-party services (like for payment processing), without exposing them to inbound traffic from the internet.
### Why *A + D* or *A + Second D* are Incorrect:
1. *Public Subnets:*
- Both versions of *D* include *public subnets, which could expose instances or components to the internet. This violates the requirement of keeping the EC2 and RDS instances **private*.
2. *Limited Subnet Configurations:*
- The first *D* has only *one private subnet*, which introduces a single point of failure, reducing availability.
- The second *D* includes *two public subnets*, which again raises the risk of public exposure, going against the requirement.
### Conclusion:
- *A* + *B* meets all the requirements of the question: privacy, high availability, and internet access for EC2 instances.
- *A + D* and *A + second D* introduce public exposure and other risks, making them less suitable for the scenario.
Therefore, *A + B* is the best combination for this question.
Peace of code
Q:7 : why kinesis data analytics is not needed here ?
idk. even lambda function cant track the locations since it doesn't store any information
Since Kinesis can be used for real-time analysis, I think D is the better choice.
in the question he points out that the company has their own data analytics software
thank you so much for this! you're a lifesaver. :)
I am having exam on 11 july .will questions concepts and difficulty will be covered in this na? Like this questions only i will get in real exam?
Yes questions will be of this level
In que 7 why ans is b? Coz lambda is not a storing service and in que they ased us for that
These videos are super helpful !! When are you planning to upload the other parts? I have my exam in 2 weeks.
Will upload soon
My exam also come in 2 weeks
All is going well :) I write this comment just to support you ;)
is question 35 answer should be A? because it is 2 different vpc you need VPC peering first to have the access.
The connection should be secure and private. Peering wouldn't accomplish this.