exam questions

Exam DP-201 All Questions

View all questions & answers for the DP-201 exam

Exam DP-201 topic 25 question 1 discussion

Actual exam question from Microsoft's DP-201
Question #: 1
Topic #: 25
[All DP-201 Questions]

What should you recommend as a batch processing solution for Health Interface?

  • A. Azure CycleCloud
  • B. Azure Stream Analytics
  • C. Azure Data Factory
  • D. Azure Databricks
Show Suggested Answer Hide Answer
Suggested Answer: B 🗳️
Scenario: ADatum identifies the following requirements for the Health Interface application:
Support a more scalable batch processing solution in Azure.
Reduce the amount of time it takes to add data from new hospitals to Health Interface.
Data Factory integrates with the Azure Cosmos DB bulk executor library to provide the best performance when you write to Azure Cosmos DB.
Reference:
https://docs.microsoft.com/en-us/azure/data-factory/connector-azure-cosmos-db
Design data processing solutions

Comments

Chosen Answer:
This is a voting comment (?). It is better to Upvote an existing comment if you don't have anything to add.
Switch to a voting comment New
Amitkhanna
Highly Voted 5 years, 2 months ago
How come batch processing using Azure Stream analytics why not Azure data bricks? seems like wrong answer this should be D.
upvoted 64 times
maciejt
4 years, 1 month ago
it's actually ADF as per their explanation, they marked it wrong. Bricks would also do I guess, there's little that ADF can do that databricks can't, if anything.
upvoted 4 times
maciejt
4 years, 1 month ago
ok, ADF can use copy data from on-premise source, spark, which is used by ADF data fows and data bricks can't do that
upvoted 1 times
...
...
...
Luke97
Highly Voted 5 years, 1 month ago
Technology choices for batch processing are 1. Azure Synapse Analytics 2. Azure HDInsight 3. Azure Data Lake Analytics 4. Azure Databricks https://docs.microsoft.com/en-us/azure/architecture/data-guide/technology-choices/batch-processing
upvoted 34 times
rajneesharora
4 years, 3 months ago
ADF has Data Flows, why is ADF not listed as part of Batch Processing? Secondly, changing the Units, will scale ADF as well... Sending data from On-Premise cant be done via DataBricks, DataBricks can act on it once data is in Azure, ADF seems to be the option
upvoted 2 times
...
...
massnonn
Most Recent 3 years, 6 months ago
for batch processing is databricks
upvoted 1 times
...
PowerBIRangerGuru
4 years ago
Why is it that no body is choosing Azure stream analytics as the input of the processing solution is messages generated by the website.
upvoted 2 times
...
Qrm_1972
4 years ago
Correct Answer: B Explanation/Reference: Explanation: Scenario: ADatum identifies the following requirements for the Health Interface application: Support a more scalable batch processing solution in Azure. Reduce the amount of time it takes to add data from new hospitals to Health Interface. Data Factory integrates with the Azure Cosmos DB bulk executor library to provide the best performance when you write to Azure Cosmos DB. Reference: https://docs.microsoft.com/en-us/azure/data-factory/connector-azure-cosmos-db
upvoted 2 times
...
dbdev
4 years ago
I would choose ADF. https://devblogs.microsoft.com/cosmosdb/migrating-relational-data-into-cosmos-db-using-azure-data-factory-and-azure-databricks/
upvoted 1 times
...
davita8
4 years, 1 month ago
D. Azure Databricks
upvoted 1 times
...
maciejt
4 years, 1 month ago
Not sure if databricks can access on prem data source. If yes, then no question D. If not, then you have to use ADF copy data activity to opy from on prem to staging. But as different hospitals have different data formats then you have to transform it to common format. ADF can use mappng data flow or call databricks notebook to do that (but only from staged data already in Azure). dataflow unfortunately is not auto scalable, you have to redefine how many cores you want to use, so I would call databricks notebook from ADF after copy data in ADF. Cosest anwer seems C - ADF.
upvoted 1 times
...
AlexD332
4 years, 2 months ago
https://docs.microsoft.com/en-us/azure/architecture/data-guide/technology-choices/batch-processing it seems Databricks
upvoted 1 times
...
AmolRajmane
4 years, 3 months ago
Don't go by word "batch". read this: Health Interface - ADatum has a critical application named Health Interface that receives hospital messages related to patient care and status updates. So stream analytics seems to be correct.
upvoted 2 times
...
mohowzeh
4 years, 4 months ago
The more reactions I read, the more confused I get. My 2 cents: in this case, the hospitals send the data in batch. This means not message-by-message, but a file containing several messages or records. Most of the discussion here looks at "batch processing", which is another story to do with analysing big data stored in files. To me, batch processing is not the correct context of this case. What we need is to ingest files coming from the hospital from time to time. Azure Data Factory seems right to me. The answer's comment also seems to point to this solution, so the answer itself might be a typo.
upvoted 3 times
...
Johnnien
4 years, 5 months ago
Can I use ADF only for solution of both Health Insights and Health Interface?
upvoted 1 times
...
Johnnien
4 years, 5 months ago
Which product would provide the best performance?
upvoted 2 times
...
BungyTex
4 years, 6 months ago
It has B showing as the answer, but then the description underneath implies C where it talks about data Factory and Cosmos DB. Data Factory is scalable.
upvoted 1 times
...
syu31svc
4 years, 6 months ago
"Minimize the number of services required to perform data processing, development, scheduling, monitoring, and the operationalizing of pipelines." I would pick Data Factory as the answer
upvoted 1 times
syu31svc
4 years, 5 months ago
Disregard this; Databricks for batch processing
upvoted 1 times
...
...
sandGrain
4 years, 7 months ago
The answer should be D: Databricks. Purely because of Scalability factor. ADF can be used but Databricks is better when it comes to scaling.
upvoted 2 times
maciejt
4 years, 1 month ago
ADF can call databricks notebook in its pipeline
upvoted 1 times
...
...
Shrikant_Kulkarni
4 years, 7 months ago
They mentioned, health interface application received data in batches (group of messages as batch from existing c# application). If ADF is answer how solution is expecting to receive data (http source / json files on blob store?) with varying schema and perform bulk insert into cosmodb? It has to be ADB receiving messages / batches on stream and ingesting them into cosmodb.
upvoted 2 times
...
Community vote distribution
A (35%)
C (25%)
B (20%)
Other
Most Voted
A voting comment increases the vote count for the chosen answer by one.

Upvoting a comment with a selected answer will also increase the vote count towards that answer by one. So if you see a comment that you already agree with, you can upvote it instead of posting a new comment.

SaveCancel
Loading ...