You have deployed multiple versions of an image classification model on AI Platform. You want to monitor the performance of the model versions over time. How should you perform this comparison?
A.
Compare the loss performance for each model on a held-out dataset.
B.
Compare the loss performance for each model on the validation data.
C.
Compare the receiver operating characteristic (ROC) curve for each model using the What-If Tool.
D.
Compare the mean average precision across the models using the Continuous Evaluation feature.
The answer is A. I am not sure why people choose B vs A as you may overfit your validation set. And you are using your held-out set really rare == no option to overfit.
In the official study guide, this was the explanation given for answer B :
"The image classification model is a deep learning model. You minimize the loss of deep learning models to get the best model. So comparing loss performance for each model on validation data is the correct answer."
I choose by myself D. But as I read the post here https://www.v7labs.com/blog/mean-average-precision, I was not sure about D.
It wrote mAP is commonly used for object detection or instance segmentation tasks.
Validation Dataset in GCP context: not trained dataset and not seen dataset
D. Compare the mean average precision across the models using the Continuous Evaluation feature
https://cloud.google.com/vertex-ai/docs/evaluation/introduction
Vertex AI provides model evaluation metrics, such as precision and recall, to help you determine the performance of your models...
Vertex AI supports evaluation of the following model types:
AuPRC: The area under the precision-recall (PR) curve, also referred to as average precision. This value ranges from zero to one, where a higher value indicates a higher-quality model.
o monitor the performance of the model versions over time, you should compare the loss performance for each model on the validation data. Therefore, option B is the correct answer.
Please, How? B is not monitoring. It is a validation. The definition of monitoring states:
"observe and check the progress or quality of (something) over a period of time"
So it is a continuous process. Each option A,B,C are just one time check, not monitoring.
The best option to monitor the performance of multiple versions of an image classification model on AI Platform over time is to compare the loss performance for each model on the validation data.
Option B is the best approach because comparing the loss performance of each model on the validation data is a common method to monitor machine learning model performance over time. The validation data is a subset of the data that is not used for model training, but is used to evaluate its performance during training and to compare different versions of the model. By comparing the loss performance of each model on the same validation data, you can determine which version of the model has better performance.
If you have multiple model versions in a single model and have created an evaluation job for each one, you can view a chart comparing the mean average precision of the model versions over time
A voting comment increases the vote count for the chosen answer by one.
Upvoting a comment with a selected answer will also increase the vote count towards that answer by one.
So if you see a comment that you already agree with, you can upvote it instead of posting a new comment.
chohan
Highly Voted 4 years agoDanny2021
Highly Voted 3 years, 9 months agoTienH
Most Recent 1 month agojkkim_jt
8 months agobludw
11 months, 3 weeks agoRyanTan
3 months, 1 week agoWookjae
1 year agoGoosemoose
1 year agoGoosemoose
1 year agosaadci
1 year agojoqu
6 months, 1 week agoSum_Sum
1 year, 7 months agoclaude2046
1 year, 8 months agoLiting
1 year, 11 months agoSamuelTsch
1 year, 11 months agoVoyager2
2 years agoM25
2 years, 1 month agolucaluca1982
2 years, 2 months agoprakashkumar1234
2 years, 2 months agoJarek7
2 years, 1 month agoFatiy
2 years, 3 months agoenghabeth
2 years, 4 months ago