exam questions

Exam Professional Machine Learning Engineer All Questions

View all questions & answers for the Professional Machine Learning Engineer exam

Exam Professional Machine Learning Engineer topic 1 question 273 discussion

Actual exam question from Google's Professional Machine Learning Engineer
Question #: 273
Topic #: 1
[All Professional Machine Learning Engineer Questions]

You work for a rapidly growing social media company. Your team builds TensorFlow recommender models in an on-premises CPU cluster. The data contains billions of historical user events and 100,000 categorical features. You notice that as the data increases, the model training time increases. You plan to move the models to Google Cloud. You want to use the most scalable approach that also minimizes training time. What should you do?

  • A. Deploy the training jobs by using TPU VMs with TPUv3 Pod slices, and use the TPUEmbeading API
  • B. Deploy the training jobs in an autoscaling Google Kubernetes Engine cluster with CPUs
  • C. Deploy a matrix factorization model training job by using BigQuery ML
  • D. Deploy the training jobs by using Compute Engine instances with A100 GPUs, and use the tf.nn.embedding_lookup API
Show Suggested Answer Hide Answer
Suggested Answer: A 🗳️

Comments

Chosen Answer:
This is a voting comment (?). It is better to Upvote an existing comment if you don't have anything to add.
Switch to a voting comment New
daidai75
Highly Voted 9 months, 3 weeks ago
Selected Answer: A
TPU (Tensor Processing Units) VMs are specialized hardware accelerators designed by Google specifically for machine learning tasks. TPUv3 Pod slices offer high scalability and are excellent for distributed training tasks. The TPUEmbedding API is optimized for handling large volumes of categorical features, which fits your scenario with 100,000 categorical features. This option is likely to offer the fastest training times due to specialized hardware and optimized APIs for large-scale machine learning tasks.
upvoted 8 times
...
omermahgoub
Most Recent 6 months, 2 weeks ago
Selected Answer: A
Addressing Bottleneck: As data size increases, CPU-based training becomes increasingly slow. TPUs are specifically designed to address this challenge, significantly accelerating training. Large Categorical Features: TPUEmbedding API efficiently handles embedding lookups for a vast number of categorical features, a common characteristic of recommender system data.
upvoted 4 times
...
JG123
7 months, 3 weeks ago
Option C
upvoted 1 times
...
guilhermebutzke
8 months, 2 weeks ago
Selected Answer: A
My Answer: A: most scalable approach that also minimizes training time: TPU using TPUEmbeading API https://www.tensorflow.org/api_docs/python/tf/tpu/experimental/embedding/TPUEmbedding
upvoted 2 times
...
Community vote distribution
A (35%)
C (25%)
B (20%)
Other
Most Voted
A voting comment increases the vote count for the chosen answer by one.

Upvoting a comment with a selected answer will also increase the vote count towards that answer by one. So if you see a comment that you already agree with, you can upvote it instead of posting a new comment.

SaveCancel
Loading ...
exam
Someone Bought Contributor Access for:
SY0-701
London, 1 minute ago