exam questions

Exam AWS Certified Machine Learning - Specialty All Questions

View all questions & answers for the AWS Certified Machine Learning - Specialty exam

Exam AWS Certified Machine Learning - Specialty topic 1 question 124 discussion

A data scientist has developed a machine learning translation model for English to Japanese by using Amazon SageMaker's built-in seq2seq algorithm with
500,000 aligned sentence pairs. While testing with sample sentences, the data scientist finds that the translation quality is reasonable for an example as short as five words. However, the quality becomes unacceptable if the sentence is 100 words long.
Which action will resolve the problem?

  • A. Change preprocessing to use n-grams.
  • B. Add more nodes to the recurrent neural network (RNN) than the largest sentence's word count.
  • C. Adjust hyperparameters related to the attention mechanism.
  • D. Choose a different weight initialization type.
Show Suggested Answer Hide Answer
Suggested Answer: C 🗳️

Comments

Chosen Answer:
This is a voting comment (?). It is better to Upvote an existing comment if you don't have anything to add.
Switch to a voting comment New
cnethers
Highly Voted 3 years, 1 month ago
I agree with an answer of C Attention mechanism. The disadvantage of an encoder-decoder framework is that model performance decreases as and when the length of the source sequence increases because of the limit of how much information the fixed-length encoded feature vector can contain. To tackle this problem, in 2015, Bahdanau et al. proposed the attention mechanism. In an attention mechanism, the decoder tries to find the location in the encoder sequence where the most important information could be located and uses that information and previously decoded words to predict the next token in the sequence.
upvoted 27 times
...
AIWave
Most Recent 8 months, 3 weeks ago
Selected Answer: C
C. By tuning attention-related hyperparameters (such as attention type, attention layer size, and dropout), the model can focus on relevant parts of the input sequence during translation.
upvoted 1 times
...
loict
1 year, 1 month ago
Selected Answer: C
A. NO - n-grams are more the opposite, it is to capture local information B. NO - it could help, but not best C. YES - best practice (https://docs.aws.amazon.com/sagemaker/latest/dg/seq-2-seq-hyperparameters.html) D. NO - weight are for vectorized words, they do not relate to sequences
upvoted 1 times
...
Mickey321
1 year, 2 months ago
Selected Answer: C
Action C: Adjusting hyperparameters related to the attention mechanism can help improve the translation quality for long sentences, because the attention mechanism allows the decoder to focus on the most relevant parts of the source sentence at each time step. According to the Amazon SageMaker documentation, the seq2seq algorithm supports several types of attention mechanisms, such as dot, general, concat, and location. The data scientist can experiment with different values of the hyperparameters attention_type, attention_coverage_type, and attention_num_hidden to find the optimal configuration for the translation task.
upvoted 1 times
...
AjoseO
1 year, 8 months ago
Selected Answer: C
C. Adjust hyperparameters related to the attention mechanism. The seq2seq algorithm uses an attention mechanism to dynamically focus on relevant parts of the input sequence for each output sequence element. Increasing the attention mechanism's ability to learn dependencies between long input and output sequences might help improve the translation quality for long sentences. The data scientist could try adjusting relevant hyperparameters such as attention depth or attention scale, or try a different attention mechanism such as scaled dot-product attention, to see if that improves the translation quality for long sentences.
upvoted 4 times
...
peterfish
2 years, 3 months ago
Selected Answer: C
i go with C
upvoted 4 times
...
SriAkula
2 years, 8 months ago
Ans: C Explanation: https://docs.aws.amazon.com/sagemaker/latest/dg/seq-2-seq- howitworks.html
upvoted 2 times
...
AddiWei
2 years, 8 months ago
This is such a niche question for a niche market. Geared towards someone who specializes in NLP.
upvoted 3 times
...
Juka3lj
3 years ago
c is correct https://docs.aws.amazon.com/sagemaker/latest/dg/seq-2-seq-howitworks.html
upvoted 3 times
...
rajesriv
3 years, 1 month ago
I believe the answer is C https://docs.aws.amazon.com/sagemaker/latest/dg/seq-2-seq-howitworks.html
upvoted 3 times
...
Community vote distribution
A (35%)
C (25%)
B (20%)
Other
Most Voted
A voting comment increases the vote count for the chosen answer by one.

Upvoting a comment with a selected answer will also increase the vote count towards that answer by one. So if you see a comment that you already agree with, you can upvote it instead of posting a new comment.

SaveCancel
Loading ...
exam
Someone Bought Contributor Access for:
SY0-701
London, 1 minute ago