exam questions

Exam DP-700 All Questions

View all questions & answers for the DP-700 exam

Exam DP-700 topic 3 question 16 discussion

Actual exam question from Microsoft's DP-700
Question #: 16
Topic #: 3
[All DP-700 Questions]

You have a Fabric workspace that contains a lakehouse named Lakehouse1.

In an external data source, you have data files that are 500 GB each. A new file is added every day.

You need to ingest the data into Lakehouse1 without applying any transformations. The solution must meet the following requirements:

• Trigger the process when a new file is added.
• Provide the highest throughput.

Which type of item should you use to ingest the data?

  • A. KQL queryset
  • B. Streaming dataset
  • C. Notebook
  • D. Dataflow Gen2
Show Suggested Answer Hide Answer
Suggested Answer: D 🗳️

Comments

Chosen Answer:
This is a voting comment (?). It is better to Upvote an existing comment if you don't have anything to add.
Switch to a voting comment New
smanzana
2 weeks, 3 days ago
Selected Answer: D
Dataflow Gen 2
upvoted 1 times
...
vernillen
3 weeks, 2 days ago
Selected Answer: D
The Dataflow Gen2 is the best, and basically because of "Trigger automatically when a new file arrives". A DFGEN2 supports event-based triggers, where a notebook needs to be added to a pipeline, this pipeline needs to be configured with an event-based trigger, ... So, notebook DOES provide high throughput, but requires too much extra configuration. If you're looking at the stand-alone items, it's 100% Dataflow Gen2
upvoted 3 times
...
kaushikcbs
3 weeks, 5 days ago
Selected Answer: D
The best choice is D. Dataflow Gen2. Why Dataflow Gen2? - Optimized for large-scale ingestion: Handles high-throughput data ingestion efficiently. - Supports automatic triggers: Can detect new files and start ingestion immediately. - No transformations required: Dataflow Gen2 allows direct ingestion into Lakehouse1 without modification. Why Not Other Options? - A. KQL queryset → Used for querying data, not for ingestion. - B. Streaming dataset → Designed for real-time reporting, not batch ingestion. - C. Notebook → Requires manual execution or scheduling, making it less efficient for automated ingestion. For more details, check out Microsoft's documentation on Dataflow Gen2 ingestion in Fabric. 🚀 Let me know if you need further clarification!
upvoted 4 times
...
DarioReymago
3 weeks, 6 days ago
Selected Answer: C
por el tamaño del archivo
upvoted 3 times
...
Seshathri
4 weeks ago
Selected Answer: C
Offers full control, supports large-scale data ingestion, integrates well with triggers, and maximizes throughput
upvoted 2 times
...
Community vote distribution
A (35%)
C (25%)
B (20%)
Other
Most Voted
A voting comment increases the vote count for the chosen answer by one.

Upvoting a comment with a selected answer will also increase the vote count towards that answer by one. So if you see a comment that you already agree with, you can upvote it instead of posting a new comment.

SaveCancel
Loading ...