exam questions

Exam AWS Certified Solutions Architect - Associate SAA-C02 All Questions

View all questions & answers for the AWS Certified Solutions Architect - Associate SAA-C02 exam

Exam AWS Certified Solutions Architect - Associate SAA-C02 topic 1 question 320 discussion

A company hosts more than 300 global websites and applications. The company requires a platform to analyze more than 30 TB of clickstream data each day.
What should a solutions architect do to transmit and process the clickstream data?

  • A. Design an AWS Data Pipeline to archive the data to an Amazon S3 bucket and run an Amazon EMR cluster with the data to generate analytics.
  • B. Create an Auto Scaling group of Amazon EC2 instances to process the data and send it to an Amazon S3 data lake for Amazon Redshift to use for analysis.
  • C. Cache the data to Amazon CloudFront. Store the data in an Amazon S3 bucket. When an object is added to the S3 bucket, run an AWS Lambda function to process the data for analysis.
  • D. Collect the data from Amazon Kinesis Data Streams. Use Amazon Kinesis Data Firehose to transmit the data to an Amazon S3 data lake. Load the data in Amazon Redshift for analysis.
Show Suggested Answer Hide Answer
Suggested Answer: D 🗳️

Comments

Chosen Answer:
This is a voting comment (?). It is better to Upvote an existing comment if you don't have anything to add.
Switch to a voting comment New
waqas
Highly Voted 3 years, 8 months ago
D sounds appropriate being Clicksteam.
upvoted 47 times
awsnoobster
3 years, 2 months ago
Agree with D, clickstream
upvoted 1 times
...
noahsark
3 years, 8 months ago
D - Example: Clickstream analytics https://d1.awsstatic.com/Products/product-name/diagrams/product-page-diagram_Amazon-Kinesis_Evolve-from-batch-to-real-time-Analytics.d7ed76be304a30be5720fd159469f157e7c09ede.png https://aws.amazon.com/kinesis/
upvoted 9 times
...
Ankush_sh
3 years, 6 months ago
This is just a real life use case which amazon has converted to a question. Check link below:- https://aws.amazon.com/solutions/case-studies/hearst-data-analytics/ Answer will be D. Key takeaways from the case study:- Built a clickstream analytics platform that transmits and processes more than 30 terabytes of clickstream data a day, streamed from more than 300 Hearst websites worldwide. Amazon Kinesis Firehose automatically moves buffered data from Amazon Kinesis Data Streams into persistent storage on Amazon Simple Storage Service (Amazon S3). This replaces an Amazon Elastic Compute Cloud (Amazon EC2) instance the team previously had to manage. The transformed clickstream data is pulled from a Hearst data lake and sent to Amazon Redshift for analytical queries and complex data science work. From Amazon Redshift, the data gets pushed to end users through an API to the company’s content management system.
upvoted 18 times
...
...
ericsrz
Highly Voted 3 years, 8 months ago
WHY SO MANY DIFFERENT ANSWER...
upvoted 5 times
miles3719
2 years, 9 months ago
You realise this is an exam right? What fun would it be if the question just had 1 answer option?
upvoted 1 times
...
...
axelrodb
Most Recent 1 year, 9 months ago
Selected Answer: D
https://aws.amazon.com/es/blogs/big-data/real-time-analytics-with-amazon-redshift-streaming- ingestion/
upvoted 1 times
...
lbertolini
2 years, 10 months ago
Selected Answer: D
Clickstream -> Kinesis
upvoted 1 times
...
ziiziii
2 years, 10 months ago
D tooooooo much data = firehouse
upvoted 1 times
miles3719
2 years, 9 months ago
Is firehouse an upgrade to firehose? Who wants to slide down the pole? Wait... that came out wrong
upvoted 1 times
...
...
TheWallPTA
3 years ago
Selected Answer: D
I think D
upvoted 1 times
...
25dec_
3 years, 5 months ago
D for sure
upvoted 1 times
...
prex
3 years, 5 months ago
Selected Answer: D
D - Example: Clickstream analytics
upvoted 1 times
...
prex
3 years, 5 months ago
D - Example: Clickstream analytics
upvoted 1 times
...
FF11
3 years, 6 months ago
AWS pipeline does all the administrative tasks of scheduling, execution and retry logic, track dependencies for all steps and does not execute the task until all dependencies are met. Its not real time.
upvoted 1 times
...
jnxtx
3 years, 6 months ago
Selected Answer: D
D - Its an AWS case study that has the exact same specs for the use of Kinesis
upvoted 2 times
...
gargaditya
3 years, 6 months ago
Should be D. D-kinesis makes it easy to collect,process and analyze 'streaming data' in 'real-time'. KDF has KDS as one if its inputs(optios being Clients, SDK, KPL,Kinesis Agent, KDS, Amazon Cloudwatch and AWS IoT). Further, the most important AWS destinations for KDF are S3, Redshift("copy via S3") and ElasticSearch. Redshift is used for data warehousing/analytics. (NOTE-KDS is 'realtime' and KDF is 'near-realtime' A-not very sure (EMR is used to deploy Big Data/Hadoop clusters for analytics). B-EC2/ASG may be a choice if aim is to develop the website, but here target is to handle transmission and processing of stream data. C-Cloudfront/caching will not help ,we are not doing content delivery. We are ingesting content from points across globe.
upvoted 4 times
...
Sikku1981
3 years, 7 months ago
https://aws.amazon.com/solutions/case-studies/hearst-data-analytics/ - I believe D
upvoted 2 times
...
Cotter
3 years, 7 months ago
I think this question answer D but must choose C for score to pass test.
upvoted 2 times
vvsandipvv
3 years, 7 months ago
I believe the question and answers are uploaded by some random dude from examtopics not by AWS, so discussion is required.
upvoted 1 times
...
...
JayBro2
3 years, 7 months ago
D is very tempting except for one requirement: the company hosts 300 global web sites. Amazon Kinesis Data Streams is a regional service, so D is not a complete answer. C meets all requirements. See https://aws.amazon.com/kinesis/data-streams/ and read Benefits/Durable
upvoted 4 times
Always_Wanting_Stuff
3 years, 7 months ago
https://aws.amazon.com/solutions/case-studies/hearst-data-analytics/
upvoted 8 times
craycomm
3 years, 7 months ago
exact match
upvoted 2 times
...
...
manan728
3 years, 6 months ago
C won't meet the real time requirements. Processing with lambda when some event triggers is not fast enough.
upvoted 1 times
...
...
Mircuz
3 years, 8 months ago
Head here : https://aws.amazon.com/en/datapipeline/details/ There is this exactly example on the bottom of the page. Answer involves Data Pipeline and EMR because we are talking about TB per day
upvoted 2 times
DahMac
3 years, 7 months ago
This use case, the code pipeline takes the data from S3 to redshift, not too S3 from click-stream. So I'm going with D, clickstream Kinesis data-streams.
upvoted 1 times
...
soosowon6
2 years, 2 months ago
Both Data Pipeline and EMR process data (ETL) and move it. EMR requires other tools for analysis. https://docs.aws.amazon.com/whitepapers/latest/big-data-analytics-options/amazon-emr.html
upvoted 1 times
soosowon6
2 years, 2 months ago
Answer D:
upvoted 1 times
...
...
...
NapoleonBorntoparty
3 years, 8 months ago
this will help you determine that D is correct: https://aws.amazon.com/blogs/big-data/running-amazon-payments-analytics-on-amazon-redshift-with-750tb-of-data/
upvoted 1 times
...
Community vote distribution
A (35%)
C (25%)
B (20%)
Other
Most Voted
A voting comment increases the vote count for the chosen answer by one.

Upvoting a comment with a selected answer will also increase the vote count towards that answer by one. So if you see a comment that you already agree with, you can upvote it instead of posting a new comment.

SaveCancel
Loading ...