exam questions

Exam AWS Certified Solutions Architect - Professional All Questions

View all questions & answers for the AWS Certified Solutions Architect - Professional exam

Exam AWS Certified Solutions Architect - Professional topic 1 question 613 discussion

A solutions architect is designing the data storage and retrieval architecture for a new application that a company will be launching soon. The application is designed to ingest millions of small records per minute from devices all around the world. Each record is less than 4 KB in size and needs to be stored in a durable location where it can be retrieved with low latency. The data is ephemeral and the company is required to store the data for 120 days only, after which the data can be deleted.
The solutions architect calculates that, during the course of a year, the storage requirements would be about 10-15 TB.
Which storage strategy is the MOST cost-effective and meets the design requirements?

  • A. Design the application to store each incoming record as a single .csv file in an Amazon S3 bucket to allow for indexed retrieval. Configure a lifecycle policy to delete data older than 120 days.
  • B. Design the application to store each incoming record in an Amazon DynamoDB table properly configured for the scale. Configure the DynamoDB Time to Live (TTL) feature to delete records older than 120 days.
  • C. Design the application to store each incoming record in a single table in an Amazon RDS MySQL database. Run a nightly cron job that executes a query to delete any records older than 120 days.
  • D. Design the application to batch incoming records before writing them to an Amazon S3 bucket. Update the metadata for the object to contain the list of records in the batch and use the Amazon S3 metadata search feature to retrieve the data. Configure a lifecycle policy to delete the data after 120 days.
Show Suggested Answer Hide Answer
Suggested Answer: B 🗳️

Comments

Chosen Answer:
This is a voting comment (?). It is better to Upvote an existing comment if you don't have anything to add.
Switch to a voting comment New
Nemer
Highly Voted 3 years, 7 months ago
B. DynamoDB with TTL, cheaper for sustained throughput of small items + suited for fast retrievals. S3 cheaper for storage only, much higher costs with writes. RDS not designed for this use case.
upvoted 32 times
oscargee
3 years, 6 months ago
DynamoDB is a Key/Value storage. And it fits big data read/write. So it cannot be used in this situation.
upvoted 2 times
sashenka
3 years, 5 months ago
Hmmm.... "The program is meant to continuously consume millions of tiny records per minute from devices located around the globe." If that doesn't say big data read/write than I don't know what is. Also, DynamoDB is perfect for this especially seeing that the 4k value is the limit size.
upvoted 4 times
...
...
...
kyo
Most Recent 3 years, 2 months ago
Selected Answer: B
Only B can do it
upvoted 1 times
...
cannottellname
3 years, 3 months ago
BBBBBBBBBBB
upvoted 1 times
...
cldy
3 years, 4 months ago
B correct.
upvoted 1 times
...
vbal
3 years, 4 months ago
Answer: D; Anyone who thinks S3 Object Metadata Search is not possible: https://aws.amazon.com/blogs/architecture/swiftly-search-metadata-with-an-amazon-s3-serverless-architecture/
upvoted 2 times
vbal
3 years, 4 months ago
I wld say building index have a cost attached which could be offset by adding more items in a batch ...
upvoted 1 times
...
...
AzureDP900
3 years, 5 months ago
B is correct!
upvoted 1 times
...
acloudguru
3 years, 5 months ago
Selected Answer: B
B. DynamoDB with TTL, cheaper for sustained throughput of small items + suited for fast retrievals. S3 cheaper for storage only, much higher costs with writes. RDS not designed for this use case.
upvoted 3 times
...
andylogan
3 years, 5 months ago
It's B
upvoted 1 times
...
tgv
3 years, 5 months ago
BBB ---
upvoted 1 times
...
DerekKey
3 years, 6 months ago
A & C - wrong B - should be correct D - I am not aware of the API that you can use to search S3 object using used-defined matadata btw. 1.000 put requests cost 0,005 and PUT request header has limitation for user-defined metadata to 2 KB
upvoted 1 times
kirrim
3 years, 5 months ago
Kendra and ElasticSearch will let you search S3 object metadata, but D sounds to me like they're saying it's a native function of S3 itself, which neither of those are. So I'm not saying D is right, just that other services can do it. Re: PUT header request with limitation for user-defined metadata to 2KB, that should be OK, you're not storing 4KB data in metadata, you'd be combining multiple 4KB data pieces into a very large flat file. The metadata would only tell you which data pieces are in that very large flat file.
upvoted 1 times
...
...
WhyIronMan
3 years, 6 months ago
I'll go with B
upvoted 2 times
...
blackgamer
3 years, 6 months ago
B is cost effective compared to C. Also low latency.
upvoted 1 times
...
Waiweng
3 years, 6 months ago
it's B
upvoted 2 times
...
Ajeeshpv
3 years, 6 months ago
B, millions of input with size less than 4 kb and low latency
upvoted 1 times
...
Kian1
3 years, 6 months ago
going with B
upvoted 1 times
01037
3 years, 6 months ago
I'll go with B, since it's an exam. But I think S3 is pretty cost-effective in this case, though I don't know what "indexed retrieval" is. As long as we give enough prefix, I think it may meet the requirement, since S3 has at least"3,500 PUT/COPY/POST/DELETE or 5,500 GET/HEAD requests per second per prefix in a bucket".
upvoted 1 times
RedKane
3 years, 6 months ago
S3 is $5 per million requests, if you have 1 million per minute it's $216,000 per month. Roughly 6x cost of DynamoDB
upvoted 4 times
tomosabc1
2 years, 6 months ago
Thanks for pointing this out.
upvoted 1 times
...
...
...
...
Ebi
3 years, 6 months ago
Answer is B Cost effective, low latency, TTL supports retention
upvoted 4 times
...
petebear55
3 years, 6 months ago
This is in the exam guys !!! answer is B Dynamo db is most suitable in these cases. 'ingest millions of small records per minute from devices all around the world.' D IS RED HERRING
upvoted 2 times
...
Community vote distribution
A (35%)
C (25%)
B (20%)
Other
Most Voted
A voting comment increases the vote count for the chosen answer by one.

Upvoting a comment with a selected answer will also increase the vote count towards that answer by one. So if you see a comment that you already agree with, you can upvote it instead of posting a new comment.

SaveCancel
Loading ...
exam
Someone Bought Contributor Access for:
SY0-701
London, 1 minute ago