exam questions

Exam AWS Certified Data Analytics - Specialty All Questions

View all questions & answers for the AWS Certified Data Analytics - Specialty exam

Exam AWS Certified Data Analytics - Specialty topic 1 question 95 discussion

A manufacturing company uses Amazon S3 to store its data. The company wants to use AWS Lake Formation to provide granular-level security on those data assets. The data is in Apache Parquet format. The company has set a deadline for a consultant to build a data lake.
How should the consultant create the MOST cost-effective solution that meets these requirements?

  • A. Run Lake Formation blueprints to move the data to Lake Formation. Once Lake Formation has the data, apply permissions on Lake Formation.
  • B. To create the data catalog, run an AWS Glue crawler on the existing Parquet data. Register the Amazon S3 path and then apply permissions through Lake Formation to provide granular-level security.
  • C. Install Apache Ranger on an Amazon EC2 instance and integrate with Amazon EMR. Using Ranger policies, create role-based access control for the existing data assets in Amazon S3.
  • D. Create multiple IAM roles for different users and groups. Assign IAM roles to different data assets in Amazon S3 to create table-based and column-based access controls.
Show Suggested Answer Hide Answer
Suggested Answer: B 🗳️

Comments

Chosen Answer:
This is a voting comment (?). It is better to Upvote an existing comment if you don't have anything to add.
Switch to a voting comment New
Donell
Highly Voted 3 years, 7 months ago
I will go with Option B. Here Data is already there in S3 bucket in parquet format. We just need to register the S3 bucket with Lake Formation after the catalogue creation. Blueprints are Ideal way if the Data has to be brought to S3 from other sources. From Lake Formation FAQ: How does Lake Formation organize my data in a data lake? A: You can use one of the blueprints available in Lake Formation to ingest data into your data lake. Lake Formation creates Glue workflows that crawl source tables, extract the data, and load it to S3. In S3, Lake Formation organizes the data for you, setting up partitions and data formats for optimized performance and cost. For data already in Amazon S3, you can register those buckets with Lake Formation to manage them. Lake Formation also crawls your data lake to maintain a data catalog and provides an intuitive user interface for you to search entities (by type, classification, attribute, or free-form text.)
upvoted 39 times
gunjan4392
3 years, 7 months ago
Makes sense, B for me
upvoted 3 times
...
lakeswimmer
3 years, 6 months ago
Agree Option B
upvoted 3 times
...
...
VikG12
Highly Voted 3 years, 8 months ago
Looks like 'A'. https://aws.amazon.com/blogs/big-data/building-securing-and-managing-data-lakes-with-aws-lake-formation/
upvoted 15 times
Ramshizzle
2 years, 11 months ago
I agree that it looks attractive. However, I think that the fact that the data is all already in S3 doesn't require Blueprints. The Blueprints section mentions importing data into your datalake, and your datalake will store the data in S3.
upvoted 1 times
...
allanm
2 years, 6 months ago
It can't be A because your data is already in S3. Blueprints only work if you do not have an S3 data lake in place which is not the case for this question.
upvoted 1 times
...
...
whenthan
Most Recent 1 year, 10 months ago
Selected Answer: B
https://docs.aws.amazon.com/lake-formation/latest/dg/security-permissions-example-scenario.html
upvoted 1 times
...
pk349
2 years, 1 month ago
B: I passed the test
upvoted 1 times
AWenger
2 years, 1 month ago
Hi, Are these questions still valid for the exams ? Please respond.
upvoted 1 times
gndu
2 years ago
Did he respond?
upvoted 1 times
...
...
...
koteshv
2 years, 2 months ago
It's B https://docs.aws.amazon.com/lake-formation/latest/dg/register-data-lake.html
upvoted 2 times
...
VijiTu
2 years, 5 months ago
Answer B https://docs.aws.amazon.com/lake-formation/latest/dg/register-data-lake.html
upvoted 1 times
...
cloudlearnerhere
2 years, 7 months ago
B is the right answer as per doc: You can use one of the blueprints available in Lake Formation to ingest data into your data lake. Lake Formation creates Glue workflows that crawl source tables, extract the data, and load it to S3. In S3, Lake Formation organizes the data for you, setting up partitions and data formats for optimized performance and cost. For data already in Amazon S3, you can register those buckets with Lake Formation to manage them. https://aws.amazon.com/lake-formation/faqs/ Option A is wrong as duplicating the data in AWS Lake Formation would not be the most cost-effective approach. Option C is wrong as using EMR would not be the most cost-effective approach also it would take time to set up. Option D is wrong as S3 does not provide table-based or column-based access control.
upvoted 2 times
...
klausyu999
2 years, 8 months ago
Selected Answer: B
A: move data takes more effort and cost C: doesn't need EMR D: We should use Lake Formation not IAM, and IAM can't offer column level control. So I like B
upvoted 1 times
...
klausyu999
2 years, 8 months ago
Selected Answer: B
A: move data takes more effort and cost C: doesn't need EMR D: We should use Lake Formation not IAM, and IAM can't offer column level control. So I like B
upvoted 1 times
...
rocky48
2 years, 10 months ago
Selected Answer: B
Option B.
upvoted 1 times
...
Ramshizzle
2 years, 11 months ago
Selected Answer: B
Use Lake Formation for the permissions. No need to use Blueprints because all the data is already in S3. Only thing that is required to get the data into the data lake is use a Glue Crawler.
upvoted 2 times
...
Ayaa4
2 years, 11 months ago
Answer: B
upvoted 1 times
...
it's A. B is ruled out as you don't have to run crawler etc. when you are using LakeFormation. It will do everything for you
upvoted 1 times
...
Bik000
3 years ago
Selected Answer: B
I think Answer should be B
upvoted 1 times
...
jrheen
3 years, 1 month ago
Answer : B
upvoted 1 times
...
Agn3001
3 years, 3 months ago
A - https://docs.aws.amazon.com/lake-formation/latest/dg/how-it-works.html Lake formation can move data as well as provide granular access from its blue prints - which is mainly intended for cost effective/time efficient way to implement faster data lakes. Whereas option B is building datalake on our own. build vs buy analogy wise A will suit the question from quick turnaround point of view from deployment.
upvoted 1 times
...
rb39
3 years, 3 months ago
B - the data is already in S3, no need to use a blueprint to import it
upvoted 1 times
...
Community vote distribution
A (35%)
C (25%)
B (20%)
Other
Most Voted
A voting comment increases the vote count for the chosen answer by one.

Upvoting a comment with a selected answer will also increase the vote count towards that answer by one. So if you see a comment that you already agree with, you can upvote it instead of posting a new comment.

SaveCancel
Loading ...