A company has developed an ML model for image classification. The company wants to deploy the model to production so that a web application can use the model.
The company needs to implement a solution to host the model and serve predictions without managing any of the underlying infrastructure.
Which solution will meet these requirements?
Amazon SageMaker Serverless Inference is the correct solution for deploying an ML model to production in a way that allows a web application to use the model without the need to manage the underlying infrastructure.
Amazon SageMaker Serverless Inference provides a fully managed environment for deploying machine learning models. It automatically provisions, scales, and manages the infrastructure required to host the model, removing the need for the company to manage servers or other underlying infrastructure.
Why Option A is Correct:
No Infrastructure Management: SageMaker Serverless Inference handles the infrastructure management for deploying and serving ML models. The company can simply provide the model and specify the required compute capacity, and SageMaker will handle the rest.
Cost-Effectiveness: The serverless inference option is ideal for applications with intermittent or unpredictable traffic, as the company only pays for the compute time consumed while handling requests.
Integration with Web Applications: This solution allows the model to be easily accessed by web applications via RESTful APIs, making it an ideal choice for hosting the model and serving predictions.
Why Other Options are Incorrect:
B . Use Amazon CloudFront to deploy the model: CloudFront is a content delivery network (CDN) service for distributing content, not for deploying ML models or serving predictions.
C . Use Amazon API Gateway to host the model and serve predictions: API Gateway is used for creating, deploying, and managing APIs, but it does not provide the infrastructure or the required environment to host and run ML models.
D . Use AWS Batch to host the model and serve predictions: AWS Batch is designed for running batch computing workloads and is not optimized for real-time inference or hosting machine learning models.
Thus, A is the correct answer, as it aligns with the requirement of deploying an ML model without managing any underlying infrastructure.
An AI company periodically evaluates its systems and processes with the help of independent software vendors (ISVs). The company needs to receive email message notifications when an ISV's compliance reports become available.
Which AWS service can the company use to meet this requirement?
AWS Data Exchange is a service that allows companies to securely exchange data with third parties, such as independent software vendors (ISVs). AWS Data Exchange can be configured to provide notifications, including email notifications, when new datasets or compliance reports become available.
Option D (Correct): 'AWS Data Exchange': This is the correct answer because it enables the company to receive notifications, including email messages, when ISVs' compliance reports are available.
Option A: 'AWS Audit Manager' is incorrect because it focuses on assessing an organization's own compliance, not receiving third-party compliance reports.
Option B: 'AWS Artifact' is incorrect as it provides access to AWS's compliance reports, not ISVs'.
Option C: 'AWS Trusted Advisor' is incorrect as it offers optimization and best practices guidance, not compliance report notifications.
AWS AI Practitioner Reference:
AWS Data Exchange Documentation: AWS explains how Data Exchange allows organizations to subscribe to third-party data and receive notifications when updates are available.
A company deployed an AI/ML solution to help customer service agents respond to frequently asked questions. The questions can change over time. The company wants to give customer service agents the ability to ask questions and receive automatically generated answers to common customer questions. Which strategy will meet these requirements MOST cost-effectively?
RAG combines large pre-trained models with retrieval mechanisms to fetch relevant context from a knowledge base. This approach is cost-effective as it eliminates the need for frequent model retraining while ensuring responses are contextually accurate and up to date. Reference: AWS RAG Techniques.
A company is building a large language model (LLM) question answering chatbot. The company wants to decrease the number of actions call center employees need to take to respond to customer questions.
Which business objective should the company use to evaluate the effect of the LLM chatbot?
The business objective to evaluate the effect of an LLM chatbot aimed at reducing the actions required by call center employees should be average call duration.
Average Call Duration:
This metric measures the time taken to handle a customer call or query. A successful LLM chatbot should reduce the call duration by efficiently providing answers, minimizing the need for human intervention.
By decreasing the average call duration, the company can improve call center efficiency, reduce costs, and enhance the user experience.
Why Option B is Correct:
Direct Impact: The objective aligns directly with the goal of reducing the number of actions call center employees must take.
Operational Efficiency: Reducing call duration is a clear indicator of the chatbot's effectiveness in assisting customers without human help.
Why Other Options are Incorrect:
A . Website engagement rate: Is unrelated to call center operations.
C . Corporate social responsibility: Does not relate to call center efficiency.
D . Regulatory compliance: Is important but does not measure the effectiveness of a chatbot in reducing employee actions.
Which AWS feature records details about ML instance data for governance and reporting?
Amazon SageMaker Model Cards provide a centralized and standardized repository for documenting machine learning models. They capture key details such as the model's intended use, training and evaluation datasets, performance metrics, ethical considerations, and other relevant information. This documentation facilitates governance and reporting by ensuring that all stakeholders have access to consistent and comprehensive information about each model. While Amazon SageMaker Debugger is used for real-time debugging and monitoring during training, and Amazon SageMaker Model Monitor tracks deployed models for data and prediction quality, neither offers the comprehensive documentation capabilities of Model Cards. Amazon SageMaker JumpStart provides pre-built models and solutions but does not focus on governance documentation.
Harley
5 days agoHassie
8 days agoClay
19 days agoAvery
1 months agoMerri
1 months agoShelia
2 months agoGene
2 months agoCletus
2 months agoAmmie
3 months agoEthan
3 months agoJosefa
3 months agoTiera
4 months agoGarry
4 months agoEdmond
4 months agoBernardo
4 months agoAlyce
4 months agoNatalie
5 months agoLenna
5 months agoLuisa
5 months agoAgustin
5 months agoVerlene
5 months agoGlory
6 months agoThaddeus
6 months agoBrendan
6 months agoIluminada
6 months agoGlory
6 months agoSharee
7 months agoAlonzo
7 months agoDoug
7 months agoFletcher
7 months agoRolland
7 months agoFanny
8 months agoSalome
8 months agoVannessa
8 months ago