Black Friday Special Limited Time Flat 70% Discount offer - Ends in 0d 00h 00m 00s - Coupon code: 70spcl

Amazon Web Services AIF-C01 AWS Certified AI Practitioner Exam(AI1-C01) Exam Practice Test

Page: 1 / 9
Total 87 questions

AWS Certified AI Practitioner Exam(AI1-C01) Questions and Answers

Question 1

A company is using an Amazon Bedrock base model to summarize documents for an internal use case. The company trained a custom model to improve the summarization quality.

Which action must the company take to use the custom model through Amazon Bedrock?

Options:

A.

Purchase Provisioned Throughput for the custom model.

B.

Deploy the custom model in an Amazon SageMaker endpoint for real-time inference.

C.

Register the model with the Amazon SageMaker Model Registry.

D.

Grant access to the custom model in Amazon Bedrock.

Question 2

A company has petabytes of unlabeled customer data to use for an advertisement campaign. The company wants to classify its customers into tiers to advertise and promote the company's products.

Which methodology should the company use to meet these requirements?

Options:

A.

Supervised learning

B.

Unsupervised learning

C.

Reinforcement learning

D.

Reinforcement learning from human feedback (RLHF)

Question 3

An AI company periodically evaluates its systems and processes with the help of independent software vendors (ISVs). The company needs to receive email message notifications when an ISV's compliance reports become available.

Which AWS service can the company use to meet this requirement?

Options:

A.

AWS Audit Manager

B.

AWS Artifact

C.

AWS Trusted Advisor

D.

AWS Data Exchange

Question 4

An AI practitioner is using an Amazon Bedrock base model to summarize session chats from the customer service department. The AI practitioner wants to store invocation logs to monitor model input and output data.

Which strategy should the AI practitioner use?

Options:

A.

Configure AWS CloudTrail as the logs destination for the model.

B.

Enable invocation logging in Amazon Bedrock.

C.

Configure AWS Audit Manager as the logs destination for the model.

D.

Configure model invocation logging in Amazon EventBridge.

Question 5

A large retailer receives thousands of customer support inquiries about products every day. The customer support inquiries need to be processed and responded to quickly. The company wants to implement Agents for Amazon Bedrock.

What are the key benefits of using Amazon Bedrock agents that could help this retailer?

Options:

A.

Generation of custom foundation models (FMs) to predict customer needs

B.

Automation of repetitive tasks and orchestration of complex workflows

C.

Automatically calling multiple foundation models (FMs) and consolidating the results

D.

Selecting the foundation model (FM) based on predefined criteria and metrics

Question 6

A company has installed a security camera. The company uses an ML model to evaluate the security camera footage for potential thefts. The company has discovered that the model disproportionately flags people who are members of a specific ethnic group.

Which type of bias is affecting the model output?

Options:

A.

Measurement bias

B.

Sampling bias

C.

Observer bias

D.

Confirmation bias

Question 7

A medical company is customizing a foundation model (FM) for diagnostic purposes. The company needs the model to be transparent and explainable to meet regulatory requirements.

Which solution will meet these requirements?

Options:

A.

Configure the security and compliance by using Amazon Inspector.

B.

Generate simple metrics, reports, and examples by using Amazon SageMaker Clarify.

C.

Encrypt and secure training data by using Amazon Macie.

D.

Gather more data. Use Amazon Rekognition to add custom labels to the data.

Question 8

A company makes forecasts each quarter to decide how to optimize operations to meet expected demand. The company uses ML models to make these forecasts.

An AI practitioner is writing a report about the trained ML models to provide transparency and explainability to company stakeholders.

What should the AI practitioner include in the report to meet the transparency and explainability requirements?

Options:

A.

Code for model training

B.

Partial dependence plots (PDPs)

C.

Sample data for training

D.

Model convergence tables

Question 9

A company has a foundation model (FM) that was customized by using Amazon Bedrock to answer customer queries about products. The company wants to validate the model's responses to new types of queries. The company needs to upload a new dataset that Amazon Bedrock can use for validation.

Which AWS service meets these requirements?

Options:

A.

Amazon S3

B.

Amazon Elastic Block Store (Amazon EBS)

C.

Amazon Elastic File System (Amazon EFS)

D.

AWS Snowcone

Question 10

A company is building a contact center application and wants to gain insights from customer conversations. The company wants to analyze and extract key information from the audio of the customer calls.

Which solution meets these requirements?

Options:

A.

Build a conversational chatbot by using Amazon Lex.

B.

Transcribe call recordings by using Amazon Transcribe.

C.

Extract information from call recordings by using Amazon SageMaker Model Monitor.

D.

Create classification labels by using Amazon Comprehend.

Question 11

A company has thousands of customer support interactions per day and wants to analyze these interactions to identify frequently asked questions and develop insights.

Which AWS service can the company use to meet this requirement?

Options:

A.

Amazon Lex

B.

Amazon Comprehend

C.

Amazon Transcribe

D.

Amazon Translate

Question 12

A company wants to build an interactive application for children that generates new stories based on classic stories. The company wants to use Amazon Bedrock and needs to ensure that the results and topics are appropriate for children.

Which AWS service or feature will meet these requirements?

Options:

A.

Amazon Rekognition

B.

Amazon Bedrock playgrounds

C.

Guardrails for Amazon Bedrock

D.

Agents for Amazon Bedrock

Question 13

A research company implemented a chatbot by using a foundation model (FM) from Amazon Bedrock. The chatbot searches for answers to questions from a large database of research papers.

After multiple prompt engineering attempts, the company notices that the FM is performing poorly because of the complex scientific terms in the research papers.

How can the company improve the performance of the chatbot?

Options:

A.

Use few-shot prompting to define how the FM can answer the questions.

B.

Use domain adaptation fine-tuning to adapt the FM to complex scientific terms.

C.

Change the FM inference parameters.

D.

Clean the research paper data to remove complex scientific terms.

Question 14

An e-commerce company wants to build a solution to determine customer sentiments based on written customer reviews of products.

Which AWS services meet these requirements? (Select TWO.)

Options:

A.

Amazon Lex

B.

Amazon Comprehend

C.

Amazon Polly

D.

Amazon Bedrock

E.

Amazon Rekognition

Question 15

What are tokens in the context of generative AI models?

Options:

A.

Tokens are the basic units of input and output that a generative AI model operates on, representing words, subwords, or other linguistic units.

B.

Tokens are the mathematical representations of words or concepts used in generative AI models.

C.

Tokens are the pre-trained weights of a generative AI model that are fine-tuned for specific tasks.

D.

Tokens are the specific prompts or instructions given to a generative AI model to generate output.

Question 16

A company wants to create a chatbot by using a foundation model (FM) on Amazon Bedrock. The FM needs to access encrypted data that is stored in an Amazon S3 bucket.

The data is encrypted with Amazon S3 managed keys (SSE-S3).

The FM encounters a failure when attempting to access the S3 bucket data.

Which solution will meet these requirements?

Options:

A.

Ensure that the role that Amazon Bedrock assumes has permission to decrypt data with the correct encryption key.

B.

Set the access permissions for the S3 buckets to allow public access to enable access over the internet.

C.

Use prompt engineering techniques to tell the model to look for information in Amazon S3.

D.

Ensure that the S3 data does not contain sensitive information.

Question 17

A company has developed an ML model for image classification. The company wants to deploy the model to production so that a web application can use the model.

The company needs to implement a solution to host the model and serve predictions without managing any of the underlying infrastructure.

Which solution will meet these requirements?

Options:

A.

Use Amazon SageMaker Serverless Inference to deploy the model.

B.

Use Amazon CloudFront to deploy the model.

C.

Use Amazon API Gateway to host the model and serve predictions.

D.

Use AWS Batch to host the model and serve predictions.

Question 18

An AI practitioner is using a large language model (LLM) to create content for marketing campaigns. The generated content sounds plausible and factual but is incorrect.

Which problem is the LLM having?

Options:

A.

Data leakage

B.

Hallucination

C.

Overfitting

D.

Underfitting

Question 19

A company wants to use a large language model (LLM) on Amazon Bedrock for sentiment analysis. The company wants to know how much information can fit into one prompt.

Which consideration will inform the company's decision?

Options:

A.

Temperature

B.

Context window

C.

Batch size

D.

Model size

Question 20

A company wants to use a large language model (LLM) to develop a conversational agent. The company needs to prevent the LLM from being manipulated with common prompt engineering techniques to perform undesirable actions or expose sensitive information.

Which action will reduce these risks?

Options:

A.

Create a prompt template that teaches the LLM to detect attack patterns.

B.

Increase the temperature parameter on invocation requests to the LLM.

C.

Avoid using LLMs that are not listed in Amazon SageMaker.

D.

Decrease the number of input tokens on invocations of the LLM.

Question 21

A company is building an ML model to analyze archived data. The company must perform inference on large datasets that are multiple GBs in size. The company does not need to access the model predictions immediately.

Which Amazon SageMaker inference option will meet these requirements?

Options:

A.

Batch transform

B.

Real-time inference

C.

Serverless inference

D.

Asynchronous inference

Question 22

An education provider is building a question and answer application that uses a generative AI model to explain complex concepts. The education provider wants to automatically change the style of the model response depending on who is asking the question. The education provider will give the model the age range of the user who has asked the question.

Which solution meets these requirements with the LEAST implementation effort?

Options:

A.

Fine-tune the model by using additional training data that is representative of the various age ranges that the application will support.

B.

Add a role description to the prompt context that instructs the model of the age range that the response should target.

C.

Use chain-of-thought reasoning to deduce the correct style and complexity for a response suitable for that user.

D.

Summarize the response text depending on the age of the user so that younger users receive shorter responses.

Question 23

A company is using few-shot prompting on a base model that is hosted on Amazon Bedrock. The model currently uses 10 examples in the prompt. The model is invoked once daily and is performing well. The company wants to lower the monthly cost.

Which solution will meet these requirements?

Options:

A.

Customize the model by using fine-tuning.

B.

Decrease the number of tokens in the prompt.

C.

Increase the number of tokens in the prompt.

D.

Use Provisioned Throughput.

Question 24

Which option is a use case for generative AI models?

Options:

A.

Improving network security by using intrusion detection systems

B.

Creating photorealistic images from text descriptions for digital marketing

C.

Enhancing database performance by using optimized indexing

D.

Analyzing financial data to forecast stock market trends

Question 25

A company is building a chatbot to improve user experience. The company is using a large language model (LLM) from Amazon Bedrock for intent detection. The company wants to use few-shot learning to improve intent detection accuracy.

Which additional data does the company need to meet these requirements?

Options:

A.

Pairs of chatbot responses and correct user intents

B.

Pairs of user messages and correct chatbot responses

C.

Pairs of user messages and correct user intents

D.

Pairs of user intents and correct chatbot responses

Question 26

A company wants to deploy a conversational chatbot to answer customer questions. The chatbot is based on a fine-tuned Amazon SageMaker JumpStart model. The application must comply with multiple regulatory frameworks.

Which capabilities can the company show compliance for? (Select TWO.)

Options:

A.

Auto scaling inference endpoints

B.

Threat detection

C.

Data protection

D.

Cost optimization

E.

Loosely coupled microservices

Page: 1 / 9
Total 87 questions