You recently deployed a model lo a Vertex Al endpoint and set up online serving in Vertex Al Feature Store. You have configured a daily batch ingestion job to update your featurestore During the batch ingestion jobs you discover that CPU utilization is high in your featurestores online serving nodes and that feature retrieval latency is high. You need to improve online serving performance during the daily batch ingestion. What should you do?
Vertex AI Feature Store provides two options for online serving: Bigtable and optimized online serving. Both options support autoscaling, which means that the number of online serving nodes can automatically adjust to the traffic demand. By enabling autoscaling, you can improve the online serving performance and reduce the feature retrieval latency during the daily batch ingestion. Autoscaling also helps you optimize the cost and resource utilization of your featurestore.Reference:
Online serving | Vertex AI | Google Cloud
New Vertex AI Feature Store: BigQuery-Powered, GenAI-Ready | Google Cloud Blog
Sanjuana
11 months agoSlyvia
11 months agoIsadora
12 months agoLarue
12 months agoJina
1 years agoWalton
1 years agoVi
1 years agoPercy
1 years agoShelba
1 years agoOdette
1 years agoMitzie
1 years agoMaia
1 years agoSherly
1 years agoCorrinne
1 years ago