exam questions

Exam AWS Certified Data Analytics - Specialty All Questions

View all questions & answers for the AWS Certified Data Analytics - Specialty exam

Exam AWS Certified Data Analytics - Specialty topic 1 question 142 discussion

A company hosts an Apache Flink application on premises. The application processes data from several Apache Kafka clusters. The data originates from a variety of sources, such as web applications, mobile apps, and operational databases. The company has migrated some of these sources to AWS and now wants to migrate the Flink application. The company must ensure that data that resides in databases within the VPC does not traverse the internet. The application must be able to process all the data that comes from the company's AWS solution, on-premises resources, and the public internet.
Which solution will meet these requirements with the LEAST operational overhead?

  • A. Implement Flink on Amazon EC2 within the company's VPC. Create Amazon Managed Streaming for Apache Kafka (Amazon MSK) clusters in the VPC to collect data that comes from applications and databases within the VPC. Use Amazon Kinesis Data Streams to collect data that comes from the public internet. Configure Flink to have sources from Kinesis Data Streams Amazon MSK, and any on-premises Kafka clusters by using AWS Client VPN or AWS Direct Connect.
  • B. Implement Flink on Amazon EC2 within the company's VPC. Use Amazon Kinesis Data Streams to collect data that comes from applications and databases within the VPC and the public internet. Configure Flink to have sources from Kinesis Data Streams and any on-premises Kafka clusters by using AWS Client VPN or AWS Direct Connect.
  • C. Create an Amazon Kinesis Data Analytics application by uploading the compiled Flink .jar file. Use Amazon Kinesis Data Streams to collect data that comes from applications and databases within the VPC and the public internet. Configure the Kinesis Data Analytics application to have sources from Kinesis Data Streams and any on-premises Kafka clusters by using AWS Client VPN or AWS Direct Connect.
  • D. Create an Amazon Kinesis Data Analytics application by uploading the compiled Flink .jar file. Create Amazon Managed Streaming for Apache Kafka (Amazon MSK) clusters in the company's VPC to collect data that comes from applications and databases within the VPC. Use Amazon Kinesis Data Streams to collect data that comes from the public internet. Configure the Kinesis Data Analytics application to have sources from Kinesis Data Streams, Amazon MSK, and any on-premises Kafka clusters by using AWS Client VPN or AWS Direct Connect.
Show Suggested Answer Hide Answer
Suggested Answer: C 🗳️

Comments

Chosen Answer:
This is a voting comment (?). It is better to Upvote an existing comment if you don't have anything to add.
Switch to a voting comment New
alfredofmt
Highly Voted 2 years, 8 months ago
Selected Answer: C
A - WRONG, EC2 adds operational overhead. B - WRONG, EC2 adds operational overhead. C - CORRECT, You can use an interface VPC endpoint to keep traffic between your Amazon VPC and Kinesis Data Streams from leaving the Amazon network. https://docs.aws.amazon.com/streams/latest/dev/vpc.html D - WRONG, even though MSK is used in this blog post ( https://aws.amazon.com/blogs/big-data/streaming-etl-with-apache-flink-and-amazon-kinesis-data-analytics/ ), having MSK increases operational overhead compared to using KDS as in answer C.
upvoted 7 times
...
chp2022
Highly Voted 3 years ago
Selected Answer: D
D makes sense based on the link provided
upvoted 6 times
...
pk349
Most Recent 2 years ago
C: I passed the test
upvoted 3 times
...
Zast
2 years, 4 months ago
I have a doubt about C and D: "Configure the Kinesis Data Analytics application to have sources from Kinesis Data Streams and any on-premises Kafka clusters by using AWS Client VPN or AWS Direct Connect." Kinesis Data Analytics can have, as a source, only Kinesis Data Stream, Kinesis Data Firehose and S3, so, as they are worded, it seems they are incorrect. Any thoughts?
upvoted 2 times
Merrick
2 years, 3 months ago
C - KDA can have an on-premise Kafka cluster as a data source https://aws.amazon.com/ko/blogs/big-data/streaming-etl-with-apache-flink-and-amazon-kinesis-data-analytics/
upvoted 4 times
...
...
rocky48
2 years, 9 months ago
Selected Answer: C
Selected Answer: C
upvoted 2 times
rocky48
2 years, 7 months ago
(https://docs.aws.amazon.com/streams/latest/dev/vpc.html) You can use an interface VPC endpoint to keep traffic between your Amazon VPC and Kinesis Data Streams from leaving the Amazon network.
upvoted 1 times
...
...
f4bi4n
2 years, 11 months ago
Selected Answer: C
C because you can have interface endpoints for KDS (https://docs.aws.amazon.com/streams/latest/dev/vpc.html) and its less overhead
upvoted 5 times
...
CHRIS12722222
3 years ago
Looks like D is right, based on the ref link provided in the question
upvoted 3 times
CHRIS12722222
3 years ago
KDS provides public endpoint so not use for vpc connectivity
upvoted 1 times
siju13
3 years ago
Why not C?
upvoted 2 times
...
...
...
Community vote distribution
A (35%)
C (25%)
B (20%)
Other
Most Voted
A voting comment increases the vote count for the chosen answer by one.

Upvoting a comment with a selected answer will also increase the vote count towards that answer by one. So if you see a comment that you already agree with, you can upvote it instead of posting a new comment.

SaveCancel
Loading ...
exam
Someone Bought Contributor Access for:
SY0-701
London, 1 minute ago