exam questions

Exam DP-100 All Questions

View all questions & answers for the DP-100 exam

Exam DP-100 topic 4 question 9 discussion

Actual exam question from Microsoft's DP-100
Question #: 9
Topic #: 4
[All DP-100 Questions]

You register a model that you plan to use in a batch inference pipeline.
The batch inference pipeline must use a ParallelRunStep step to process files in a file dataset. The script has the ParallelRunStep step runs must process six input files each time the inferencing function is called.
You need to configure the pipeline.
Which configuration setting should you specify in the ParallelRunConfig object for the PrallelRunStep step?

  • A. process_count_per_node= "6"
  • B. node_count= "6"
  • C. mini_batch_size= "6"
  • D. error_threshold= "6"
Show Suggested Answer Hide Answer
Suggested Answer: C 🗳️

Comments

Chosen Answer:
This is a voting comment (?). It is better to Upvote an existing comment if you don't have anything to add.
Switch to a voting comment New
Jolin130
Highly Voted 4 years ago
The answer should be C. For FileDataset input, this field is the number of files user script can process in one run() call.
upvoted 40 times
...
shivaborusu
Highly Voted 4 years ago
Answer is C, https://docs.microsoft.com/en-us/python/api/azureml-pipeline-steps/azureml.pipeline.steps.parallel_run_config.parallelrunconfig?view=azure-ml-py
upvoted 19 times
...
MonayiYC
Most Recent 11 months, 3 weeks ago
mini_batch_size: Number of files passed per scoring script run. https://learn.microsoft.com/en-us/training/modules/deploy-model-batch-endpoint/3-deploy-your-mlflow-model-batch-endpoint
upvoted 1 times
...
james2033
1 year, 1 month ago
Selected Answer: C
ParallelRunStep means run parallel in single one node. Question keyword 'six input files each time the inferencing function is called'. Answer keyword: mini_batch_size="6" .
upvoted 1 times
...
phdykd
1 year, 9 months ago
The ParallelRunStep step runs must process six input files each time the inferencing function is called, so the correct configuration setting to specify in the ParallelRunConfig object for the ParallelRunStep step is the mini_batch_size, which should be set to 6. Therefore, the correct answer is: C. mini_batch_size= "6"
upvoted 2 times
...
RamundiGR
1 year, 10 months ago
as stated in Microsoft Documentation this should be mini_batch_size: https://docs.microsoft.com/en-us/python/api/azureml-contrib-pipeline-steps/azureml.contrib.pipeline.steps.parallelrunconfig?view=azure-ml-py. @ExamTopic Can we please fix the answer.
upvoted 1 times
...
RamundiGR
1 year, 10 months ago
I believe the answer is C can we please fix the mistake? We are referring on the number of files here so why should be A?
upvoted 1 times
...
michaelmorar
1 year, 12 months ago
Selected Answer: C
Node count is about number of compute instances, servers etc.,
upvoted 3 times
...
ning
2 years, 6 months ago
Selected Answer: C
I am sure C is the answer, number of nodes = how many computer inst, number of processes = how many parallel running processes on a node, min_batch_size = min number of files processed
upvoted 2 times
...
pancman
2 years, 7 months ago
Selected Answer: C
Correct answer is C. Question notes that the type of dateset to be used is a file dataset. Microsoft documentation says mini_batch_size: For FileDataset input, this field is the number of files a user script can process in one run() call https://docs.microsoft.com/en-us/python/api/azureml-pipeline-steps/azureml.pipeline.steps.parallel_run_config.parallelrunconfig?view=azure-ml-py
upvoted 2 times
...
dija123
2 years, 11 months ago
Selected Answer: C
Answer is C
upvoted 5 times
...
hargur
3 years, 1 month ago
on 19Oct2021
upvoted 2 times
...
1q2w3e4r5t
3 years, 3 months ago
the answer should be mini_batch_size mini_batch_size Union[str, int] For FileDataset input, this field is the number of files a user script can process in one run() call. For TabularDataset input, this field is the approximate size of data the user script can process in one run() call. Example values are 1024, 1024KB, 10MB, and 1GB. (optional, default value is 10 files for FileDataset and 1MB for TabularDataset.)
upvoted 1 times
...
ljljljlj
3 years, 4 months ago
On exam 2021/7/10
upvoted 4 times
...
anwar1
3 years, 6 months ago
This Q&A model set has quite a few wrong answers... #Examtopics, kindly fix the wrong answers.
upvoted 4 times
...
rishi_ram
3 years, 6 months ago
Answer is C: Refer below code: def run(mini_batch): # This runs for each batch resultList = [] # process each file in the batch for f in mini_batch: # Read comma-delimited data into an array data = np.genfromtxt(f, delimiter=',') # Reshape into a 2-dimensional array for model input prediction = model.predict(data.reshape(1, -1)) # Append prediction to results resultList.append("{}: {}".format(os.path.basename(f), prediction[0])) return resultList
upvoted 1 times
...
BilJon
3 years, 8 months ago
mini_batch_size is correct
upvoted 5 times
...
Community vote distribution
A (35%)
C (25%)
B (20%)
Other
Most Voted
A voting comment increases the vote count for the chosen answer by one.

Upvoting a comment with a selected answer will also increase the vote count towards that answer by one. So if you see a comment that you already agree with, you can upvote it instead of posting a new comment.

SaveCancel
Loading ...