exam questions

Exam AWS Certified Data Analytics - Specialty All Questions

View all questions & answers for the AWS Certified Data Analytics - Specialty exam

Exam AWS Certified Data Analytics - Specialty topic 1 question 120 discussion

A company needs to collect streaming data from several sources and store the data in the AWS Cloud. The dataset is heavily structured, but analysts need to perform several complex SQL queries and need consistent performance. Some of the data is queried more frequently than the rest. The company wants a solution that meets its performance requirements in a cost-effective manner.
Which solution meets these requirements?

  • A. Use Amazon Managed Streaming for Apache Kafka to ingest the data to save it to Amazon S3. Use Amazon Athena to perform SQL queries over the ingested data.
  • B. Use Amazon Managed Streaming for Apache Kafka to ingest the data to save it to Amazon Redshift. Enable Amazon Redshift workload management (WLM) to prioritize workloads.
  • C. Use Amazon Kinesis Data Firehose to ingest the data to save it to Amazon Redshift. Enable Amazon Redshift workload management (WLM) to prioritize workloads.
  • D. Use Amazon Kinesis Data Firehose to ingest the data to save it to Amazon S3. Load frequently queried data to Amazon Redshift using the COPY command. Use Amazon Redshift Spectrum for less frequently queried data.
Show Suggested Answer Hide Answer
Suggested Answer: D 🗳️

Comments

Chosen Answer:
This is a voting comment (?). It is better to Upvote an existing comment if you don't have anything to add.
Switch to a voting comment New
aws2019
Highly Voted 3 years, 6 months ago
D is right
upvoted 16 times
...
lakeswimmer
Highly Voted 3 years, 5 months ago
I think it should be D "multiple sophisticated SQL queries with consistent performance" means Redshift. WLM just prioritizes query. frequently queried can be in Redshift
upvoted 6 times
...
pk349
Most Recent 2 years ago
D: I passed the test
upvoted 2 times
...
CleverMonkey092
2 years, 2 months ago
D is the most logical answer
upvoted 2 times
...
cloudlearnerhere
2 years, 7 months ago
Correct answer is D as Kinesis Data Firehose can be used to push data directly to S3. Frequently queried data can be loaded into Redshift for querying. Less frequent data can be still stored in S3 and queried using Redshift Spectrum. Using Amazon Redshift Spectrum, you can efficiently query and retrieve structured and semistructured data from files in Amazon S3 without having to load the data into Amazon Redshift tables. Redshift Spectrum queries employ massive parallelism to run very fast against large datasets. Much of the processing occurs in the Redshift Spectrum layer, and most of the data remains in Amazon S3. Multiple clusters can concurrently query the same dataset in Amazon S3 without the need to make copies of the data for each cluster. https://docs.aws.amazon.com/redshift/latest/dg/c-using-spectrum.html Option A is wrong as Athena is not ideal for performing complex queries. Options B & C are wrong as using Redshift for all the data would not be cost-effective.
upvoted 4 times
...
rocky48
2 years, 10 months ago
Selected Answer: D
Answer is D
upvoted 3 times
...
Robot209
2 years, 11 months ago
Selected Answer: D
Answer is D
upvoted 2 times
...
Bik000
3 years ago
Selected Answer: D
My Answer is D
upvoted 2 times
...
astalavista1
3 years, 1 month ago
Selected Answer: D
D is correct
upvoted 4 times
...
youonebe
3 years, 2 months ago
Ans: C The question is focused on performance. Hence using WLM to fulfill the purpose.
upvoted 1 times
MWL
3 years, 1 month ago
WLM prioritizes workloads mostly according to the run time, like long/short run. If only the data is well patitioned, and formatted, Athena on S3 will also get result in "consistent performance". So I think the question is mainly asking about the "cost".
upvoted 2 times
...
...
Agn3001
3 years, 3 months ago
when kinesis firehost can directly write to Redshift why write to s3 and then copy command then wlm for queries ? I think answer C
upvoted 3 times
yusnardo
3 years, 2 months ago
Kinesis Data Firehose delivers your data to your S3 bucket first and then issues an Amazon Redshift COPY command to load the data into your Amazon Redshift cluster. https://docs.aws.amazon.com/firehose/latest/dev/create-destination.html#create-destination-redshift
upvoted 5 times
...
...
goutes
3 years, 6 months ago
I think it is A. Cost-effective is the requirement, so RD cannot be the one.
upvoted 3 times
sanpak
3 years, 5 months ago
But, how you are going to ingest the data from MSK to S3 ? I think this is pub-sub model, someone has to consume the Kafka Topic to put there in S3. Doest it ? it must be D or B. Where B might take more cost to put data on Redshift cluster.
upvoted 2 times
...
MWL
3 years, 1 month ago
And you will par for Athena based on data scanning. if you query frequently on some data, you will pay much for the data scanning.
upvoted 1 times
...
np2021
2 years, 2 months ago
PERFORMANCE is the requirement not COST?
upvoted 1 times
...
...
ali98
3 years, 6 months ago
Answer : D
upvoted 1 times
...
Chints01
3 years, 6 months ago
I believe the answer should be A. thoughts?
upvoted 2 times
ali98
3 years, 6 months ago
wrong - Amazon Athena can't be used for complex queries
upvoted 3 times
...
...
srinivasa
3 years, 7 months ago
Answer: D
upvoted 1 times
...
Community vote distribution
A (35%)
C (25%)
B (20%)
Other
Most Voted
A voting comment increases the vote count for the chosen answer by one.

Upvoting a comment with a selected answer will also increase the vote count towards that answer by one. So if you see a comment that you already agree with, you can upvote it instead of posting a new comment.

SaveCancel
Loading ...