exam questions

Exam AWS Certified Big Data - Specialty All Questions

View all questions & answers for the AWS Certified Big Data - Specialty exam

Exam AWS Certified Big Data - Specialty topic 1 question 38 discussion

Exam question from Amazon's AWS Certified Big Data - Specialty
Question #: 38
Topic #: 1
[All AWS Certified Big Data - Specialty Questions]

A company is building a new application in AWS. The architect needs to design a system to collect application log events. The design should be a repeatable pattern that minimizes data loss if an application instance fails, and keeps a durable copy of a log data for at least 30 days.
What is the simplest architecture that will allow the architect to analyze the logs?

  • A. Write them directly to a Kinesis Firehose. Configure Kinesis Firehose to load the events into an Amazon Redshift cluster for analysis.
  • B. Write them to a file on Amazon Simple Storage Service (S3). Write an AWS Lambda function that runs in response to the S3 event to load the events into Amazon Elasticsearch Service for analysis.
  • C. Write them to the local disk and configure the Amazon CloudWatch Logs agent to load the data into CloudWatch Logs and subsequently into Amazon Elasticsearch Service.
  • D. Write them to CloudWatch Logs and use an AWS Lambda function to load them into HDFS on an Amazon Elastic MapReduce (EMR) cluster for analysis.
Show Suggested Answer Hide Answer
Suggested Answer: B 🗳️

Comments

Chosen Answer:
This is a voting comment (?). It is better to Upvote an existing comment if you don't have anything to add.
Switch to a voting comment New
Bulti
Highly Voted 3 years, 7 months ago
Correct answer is C: Not A - because writing logs to RedShift doesn't make sense. Not B - because to write logs to S3 you will have to still configure Cloudwatch logs on the server or write code in your application to use S3 SDK to write the log data to S3 directly which is not the best or simplest solution. Not D -because EMR is not a storage service. Only option that fits is C with no additional effort except for configuring Cloudwatch log agent to ship the logs to Cloudwatch and then configure CloudWatch to send the logs directly to Elasticsearch which doesn't require Lambda to glue them together.
upvoted 11 times
...
ariane_tateishi
Most Recent 3 years, 7 months ago
For me the right option is B considering that "Write them to the local disk" will not be compliance with the requirement because if the application fail the logs will be lost.
upvoted 1 times
...
DerekKey
3 years, 7 months ago
C: You can configure a CloudWatch Logs log group to stream data it receives to your Amazon Elasticsearch Service (Amazon ES) cluster in near real-time through a CloudWatch Logs subscription.
upvoted 2 times
...
Royk2020
3 years, 7 months ago
"Simplest architecture" .... Answer is C. No code , just configuration
upvoted 2 times
...
abhineet
3 years, 7 months ago
A is perfect answer actually, kinesis firehose first writes data to s3, this meets saving logs upto 30 days with lifecycle policy on bucket, while redshift can be used for analysis
upvoted 1 times
...
kkyong
3 years, 7 months ago
B is wrong. If you are a developer ,you will know application can't write log to s3 directly . application must write log to buffer or file ,and then put to s3 so C is correct answer
upvoted 1 times
Corram
3 years, 7 months ago
good point, but i think "write them to a file on S3" leaves room for creating this file first locally and then store it to S3. no need to subsequentially write to an existing S3 object. but yeah, i still think C is correct due to Bulit's comments.
upvoted 1 times
...
...
Bulti
3 years, 7 months ago
to support answer C- Here is from Cloudwatch FAQ CloudWatch Logs uses your log data for monitoring; so, no code changes are required. Long term log retention: You can use CloudWatch Logs to store your log data indefinitely in highly durable and cost effective storage without worrying about hard drives running out of space.
upvoted 4 times
Corram
3 years, 7 months ago
to make this even more clear, Cloudwatch Log groups can also be easily streamed to ElasticSearch https://docs.aws.amazon.com/AmazonCloudWatch/latest/logs/CWL_ES_Stream.html
upvoted 1 times
...
matthew95
3 years, 7 months ago
This is C, look at this page: https://www.oreilly.com/library/view/aws-automation-cookbook/9781788394925/30fd87cf-3d67-4363-a95a-5208296d32cb.xhtml
upvoted 1 times
...
...
jiedee
3 years, 7 months ago
can anyone pls explain why not A?
upvoted 1 times
Corram
3 years, 7 months ago
to me it looks like technically this could work. however, Elasticsearch is really suited for log file analysis and both Kinesis and Redshift can be expensive.
upvoted 1 times
...
...
san2020
3 years, 7 months ago
my selection B
upvoted 3 times
Soona_Paana
3 years, 7 months ago
Itc C... But a large portion of ur answers are right :p
upvoted 1 times
...
...
ME2000
3 years, 7 months ago
The big problem with option C is how to load events into ES? So one option left is B
upvoted 2 times
iamsajal
3 years, 7 months ago
Simple way to do it through console. I guess the answer is C. https://docs.aws.amazon.com/elasticsearch-service/latest/developerguide/es-aws-integrations.html#es-aws-integrations-cloudwatch-es
upvoted 1 times
...
...
chaudh
3 years, 7 months ago
B is correct
upvoted 2 times
...
viduvivek
3 years, 7 months ago
B is Correct. Local disk or EBS durability is 5 9's, while S3 durability is 11 9's. As the requirement is to keep a durable copy of the log data, S3 is the the best option.
upvoted 2 times
...
cybe001
3 years, 8 months ago
I also pick B, S3 is simple and durable and Elasticsearch is for log analysis.
upvoted 1 times
...
Zire
3 years, 8 months ago
the "minimizes data loss if an application instance fails" phrase hints me to option C. Option B looks to simple, but you have to re-invent the services (lambda probably) to transfer the files to S3 and schedule it.
upvoted 1 times
...
M2
3 years, 8 months ago
Answer is B
upvoted 2 times
...
bigdatalearner
3 years, 8 months ago
mattyb123 are you sure ?
upvoted 2 times
...
pdach
3 years, 8 months ago
B. Question states minimizes data loss S3 should be used.
upvoted 2 times
exams
3 years, 8 months ago
B. most durable
upvoted 2 times
Nik225
3 years, 7 months ago
Slightly more durable but not nearly as simple
upvoted 1 times
...
...
...
Community vote distribution
A (35%)
C (25%)
B (20%)
Other
Most Voted
A voting comment increases the vote count for the chosen answer by one.

Upvoting a comment with a selected answer will also increase the vote count towards that answer by one. So if you see a comment that you already agree with, you can upvote it instead of posting a new comment.

SaveCancel
Loading ...