exam questions

Exam DP-203 All Questions

View all questions & answers for the DP-203 exam

Exam DP-203 topic 4 question 39 discussion

Actual exam question from Microsoft's DP-203
Question #: 39
Topic #: 4
[All DP-203 Questions]

You have an Azure Data Lake Storage Gen2 account that contains two folders named Folder1 and Folder2.

You use Azure Data Factory to copy multiple files from Folder1 to Folder2.

You receive the following error.

Operation on target Copy_sks failed: Failure happened on 'Sink' side.
ErrorCode=DelimitedTextMoreColumnsThanDefined,
'Type=Microsoft.DataTransfer.Common.Snared.HybridDeliveryException,
Message=Error found when processing 'Csv/Tsv Format Text' source
'0_2020_11_09_11_43_32.avro' with row number 53: found more columns than expected column count 27., Source=Microsoft.DataTransfer.Comnon,'

What should you do to resolve the error?

  • A. Change the Copy activity setting to Binary Copy.
  • B. Lower the degree of copy parallelism.
  • C. Add an explicit mapping.
  • D. Enable fault tolerance to skip incompatible rows.
Show Suggested Answer Hide Answer
Suggested Answer: A 🗳️

Comments

Chosen Answer:
This is a voting comment (?). It is better to Upvote an existing comment if you don't have anything to add.
Switch to a voting comment New
Yemeral
Highly Voted 2 years, 1 month ago
Selected Answer: A
Correct answer is A. We are just copying files between folders. Selecting binary copy, ADF will not check schema. With D we would discard data With C we would change file contents
upvoted 15 times
...
chryckie
Highly Voted 2 years, 1 month ago
Selected Answer: A
It's tricky. Not D, because you don't just throw away data. Likely not C, because it doesn't solve for future schema variability. (Avro formats are usually chosen in situations where the schema may evolve over time, because they store both the data and schema in the file itself.) A makes most sense, since you're just trying to move files over. Binary preserves everything as-is, and you can read/interpret them as ASCII/UTF-8/whatever later.
upvoted 10 times
chryckie
2 years, 1 month ago
Oh! Also, the message says it's trying to process the Avro file as a Csv/Tsv Format Text. That's likely the issue.
upvoted 1 times
...
...
renan_ineu
Most Recent 8 months, 3 weeks ago
Selected Answer: A
The goal is to copy from source to destination. Period. Binary is not analysed (images, videos, etc). Treat text as binary and you'll ve fine. https://learn.microsoft.com/en-us/azure/data-factory/pipeline-trigger-troubleshoot-guide#you-see-a-delimitedtextmorecolumnsthandefined-error-when-copying-a-pipeline
upvoted 1 times
...
tadenet
1 year ago
Selected Answer: C
chatgpt: The best solution to resolve the error is: Add an explicit mapping. Adding an explicit mapping will ensure that the data from the source file is correctly mapped to the destination columns, thus resolving the error related to the column count mismatch. Changing the copy activity setting to binary copy may not directly address the root cause of the error, which is a column count mismatch. Binary copy may still encounter the same issue if the source data does not match the expected column count in the destination. Therefore, while changing the copy activity setting to binary copy might be beneficial in some scenarios, it may not effectively resolve the specific error mentioned.
upvoted 3 times
...
MBRSDG
1 year, 2 months ago
Selected Answer: A
Error explicitly points error to a AVRO file, not CSV. I don't know if it could be possible to have schema mismatch on a AVRO file (I don't think it resembles a CSV at all), but setting up a binary copy activity colves the error if the requirement is just to copy files from A to B. Obviously, it is not the best solution, but it depends on the variety of file formats inside the folder. In case all the files are AVRO formatted, there's a beautiful dataset enabling to properly handle such a format --> https://learn.microsoft.com/en-us/azure/data-factory/format-avro
upvoted 2 times
...
jppdks
1 year, 2 months ago
Selected Answer: A
Resolution https://learn.microsoft.com/en-us/azure/data-factory/pipeline-trigger-troubleshoot-guide#you-see-a-delimitedtextmorecolumnsthandefined-error-when-copying-a-pipeline Select the Binary Copy option while creating the Copy activity. This way, for bulk copies or migrating your data from one data lake to another, Data Factory won't open the files to read the schema. Instead, Data Factory will treat each file as binary and copy it to the other location.
upvoted 3 times
...
j888
1 year, 3 months ago
Adding an explicit mapping allows you to define the exact structure of the data being copied, including the number and names of columns. This ensures that the Copy activity can handle any inconsistencies in the source data and prevent this error from occurring.
upvoted 2 times
...
Bill_Walker
1 year, 4 months ago
Correct answer is A https://learn.microsoft.com/en-us/azure/data-factory/pipeline-trigger-troubleshoot-guide#you-see-a-delimitedtextmorecolumnsthandefined-error-when-copying-a-pipeline
upvoted 4 times
...
Azure_2023
1 year, 4 months ago
Selected Answer: A
Switch the Copy activity to Binary Copy Binary Copy can help to resolve this error by copying the data from the source file without any data conversion. This means that the data will be copied as-is, even if it contains more columns than the sink table expects. However, it does not support data transformation. To use Binary Copy for this task, you will need to configure the source and sink connections to point to the respective folders. For the source connection, you can use the Delimited Text connector in ADF. For the sink connection, you can also use the Delimited Text connector or another connector that supports the file format in the target folder. Use an explicit mapping If you want to copy the data from the source file and transform it to match the schema of the sink table, you can use an explicit mapping. This will allow you to map the source columns to the corresponding sink columns.
upvoted 2 times
...
SATHTECH
1 year, 6 months ago
C. Add an explicit mapping. Explicit mapping involves specifying the mapping between source and destination columns explicitly. By doing this, you can ensure that each column in the source file is correctly mapped to its corresponding column in the destination file, which helps to address issues related to column count mismatches. While other options may have their use cases, such as changing the copy activity setting to Binary Copy or enabling fault tolerance to skip incompatible rows, adding an explicit mapping (Option C) is specifically designed to handle issues where the source and destination structures do not match in terms of column count or order. Therefore, in the context of resolving a "DelimitedTextMoreColumnsThanDefined" error, adding an explicit mapping is the most appropriate action.
upvoted 7 times
...
matiandal
1 year, 7 months ago
Vote for C we have a schema mismatch -) Also • Option A: Binary Copy is used for copying non-parseable files like images or videos, not for structured data like CSV.
upvoted 4 times
...
pperf
1 year, 8 months ago
Selected Answer: A
https://sqlwithmanoj.com/2020/07/29/azure-data-factory-adf-pipeline-failure-found-more-columns-than-expected-column-count-delimitedtextmorecolumnsthandefined/
upvoted 3 times
...
EliteAllen
1 year, 8 months ago
Selected Answer: A
A. Change the Copy activity setting to Binary Copy: This would bypass the error by copying the files as-is without interpreting the contents. This method might be suitable if the files are not strictly delimited text files or if you plan to handle the data inconsistency at a later stage or in a different part of the pipeline.
upvoted 2 times
...
kkk5566
1 year, 9 months ago
Selected Answer: A
A is correct
upvoted 2 times
...
Tightbot
1 year, 9 months ago
Selected Answer: C
I would go with Option C- Add an explicit mapping. Laying out possible derivations from the question 1. the actual error says - column mismatch . 2. Even though the filename is "filename.avro" , it could just be a filename, the source file type is CSV/TSV. Possible answers 1. Add an explicit mapping 2. Enabling Fault tolerance to skip incompatible rows I think both would be a possible solution, but to me, skipping incompatible rows is more of a temporary solution and explicit mapping would be more permanent for this error. I'm also excluding future schema issues that arise after this as there is no information about it.
upvoted 2 times
...
[Removed]
1 year, 9 months ago
Selected Answer: A
It says CSV/tsv source but file is avro so A is the answer
upvoted 2 times
...
andjurovicela
1 year, 11 months ago
Selected Answer: D
I was pondering a bit about this one, and decided to go with D. Reasoning behind this is because the question was "how to resolve this error?" and 100% preservation of source data hasn't been a condition, hence D is the most straightforward.
upvoted 1 times
...
Community vote distribution
A (35%)
C (25%)
B (20%)
Other
Most Voted
A voting comment increases the vote count for the chosen answer by one.

Upvoting a comment with a selected answer will also increase the vote count towards that answer by one. So if you see a comment that you already agree with, you can upvote it instead of posting a new comment.

SaveCancel
Loading ...