Deal of The Day! Hurry Up, Grab the Special Discount - Save 25% - Ends In 00:00:00 Coupon code: SAVE25
Welcome to Pass4Success

- Free Preparation Discussions

Google Exam Professional Machine Learning Engineer Topic 3 Question 99 Discussion

Actual exam question for Google's Professional Machine Learning Engineer exam
Question #: 99
Topic #: 3
[All Professional Machine Learning Engineer Questions]

You are developing an image recognition model using PyTorch based on ResNet50 architecture. Your code is working fine on your local laptop on a small subsample. Your full dataset has 200k labeled images You want to quickly scale your training workload while minimizing cost. You plan to use 4 V100 GPUs. What should you do? (Choose Correct Answer and Give Reference and Explanation)

Show Suggested Answer Hide Answer
Suggested Answer: C

Traffic splitting is a feature of Vertex AI that allows you to distribute the prediction requests among multiple models or model versions within the same endpoint. You can specify the percentage of traffic that each model or model version receives, and change it at any time. Traffic splitting can help you test the new model in production without creating a new endpoint or a separate service. You can deploy the new model to the existing Vertex AI endpoint, and use traffic splitting to send 5% of production traffic to the new model. You can monitor the end-user metrics, such as listening time, to compare the performance of the new model and the previous model. If the end-user metrics improve between models over time, you can gradually increase the percentage of production traffic sent to the new model. This solution can help you test the new model in production while minimizing complexity and cost.Reference:

Traffic splitting | Vertex AI

Deploying models to endpoints | Vertex AI


Contribute your Thoughts:

Ernestine
6 days ago
I'm leaning towards Option C. Creating a Vertex AI Workbench instance with the required GPUs sounds like a great way to get my model training up and running quickly. Plus, I don't have to worry about managing the infrastructure myself.
upvoted 0 times
...
Lauran
14 days ago
Option A seems like the easiest way to scale my training workload. I can simply configure a Compute Engine VM with the necessary dependencies and use Vertex AI to train my model. Definitely the fastest and most cost-effective solution.
upvoted 0 times
...
Jesusita
14 days ago
I'm not sure, but I think option D could also be a valid choice. Creating a GKE cluster with a node pool that has 4 V100 GPUs might be a good solution too.
upvoted 0 times
...
Paul
19 days ago
I agree with Robt. Option A seems like the most efficient way to scale the training workload while minimizing cost.
upvoted 0 times
...
Robt
22 days ago
I think the correct answer is A. It involves configuring a Compute Engine VM with the necessary dependencies and using Vertex AI with the required GPUs.
upvoted 0 times
...

Save Cancel
az-700  pass4success  az-104  200-301  200-201  cissp  350-401  350-201  350-501  350-601  350-801  350-901  az-720  az-305  pl-300  

Warning: Cannot modify header information - headers already sent by (output started at /pass.php:70) in /pass.php on line 77