Pre-Summer Special Limited Time 70% Discount Offer - Ends in 0d 00h 00m 00s - Coupon code: validbest

Pass the Amazon Web Services AWS Certified AI Practitioner AIF-C01 Questions and answers with ValidTests

Exam AIF-C01 All Questions
Exam AIF-C01 Premium Access

View all detail and faqs for the AIF-C01 exam

Viewing page 4 out of 11 pages
Viewing questions 31-40 out of questions
Questions # 31:

A company wants to build an ML application.

Select and order the correct steps from the following list to develop a well-architected ML workload. Each step should be selected one time. (Select and order FOUR.)

• Deploy model

• Develop model

• Monitor model

• Define business goal and frame ML problem

Question # 31

Options:

Expert Solution
Questions # 32:

A company has a generative AI application that uses a pre-trained foundation model (FM) on Amazon Bedrock. The company wants the FM to include more context by using company information.

Which solution meets these requirements MOST cost-effectively?

Options:

A.

Use Amazon Bedrock Knowledge Bases.

B.

Choose a different FM on Amazon Bedrock.

C.

Use Amazon Bedrock Agents.

D.

Deploy a custom model on Amazon Bedrock.

Expert Solution
Questions # 33:

An e-commerce company wants to build a solution to determine customer sentiments based on written customer reviews of products.

Which AWS services meet these requirements? (Select TWO.)

Options:

A.

Amazon Lex

B.

Amazon Comprehend

C.

Amazon Polly

D.

Amazon Bedrock

E.

Amazon Rekognition

Expert Solution
Questions # 34:

A company is using a pre-trained large language model (LLM) to build a chatbot for product recommendations. The company needs the LLM outputs to be short and written in a specific language.

Which solution will align the LLM response quality with the company's expectations?

Options:

A.

Adjust the prompt.

B.

Choose an LLM of a different size.

C.

Increase the temperature.

D.

Increase the Top K value.

Expert Solution
Questions # 35:

A company built an AI-powered resume screening system. The company used a large dataset to train the model. The dataset contained resumes that were not representative of all demographics. Which core dimension of responsible AI does this scenario present?

Options:

A.

Fairness.

B.

Explainability.

C.

Privacy and security.

D.

Transparency.

Expert Solution
Questions # 36:

A company uses Amazon SageMaker for its ML pipeline in a production environment. The company has large input data sizes up to 1 GB and processing times up to 1 hour. The company needs near real-time latency.

Which SageMaker inference option meets these requirements?

Options:

A.

Real-time inference

B.

Serverless inference

C.

Asynchronous inference

D.

Batch transform

Expert Solution
Questions # 37:

A company is developing an ML model to make loan approvals. The company must implement a solution to detect bias in the model. The company must also be able to explain the model's predictions.

Which solution will meet these requirements?

Options:

A.

Amazon SageMaker Clarify

B.

Amazon SageMaker Data Wrangler

C.

Amazon SageMaker Model Cards

D.

AWS AI Service Cards

Expert Solution
Questions # 38:

A company is developing a new model to predict the prices of specific items. The model performed well on the training dataset. When the company deployed the model to production, the model's performance decreased significantly.

What should the company do to mitigate this problem?

Options:

A.

Reduce the volume of data that is used in training.

B.

Add hyperparameters to the model.

C.

Increase the volume of data that is used in training.

D.

Increase the model training time.

Expert Solution
Questions # 39:

A company wants to use language models to create an application for inference on edge devices. The inference must have the lowest latency possible.

Which solution will meet these requirements?

Options:

A.

Deploy optimized small language models (SLMs) on edge devices.

B.

Deploy optimized large language models (LLMs) on edge devices.

C.

Incorporate a centralized small language model (SLM) API for asynchronous communication with edge devices.

D.

Incorporate a centralized large language model (LLM) API for asynchronous communication with edge devices.

Expert Solution
Questions # 40:

A company is building an ML model. The company collected new data and analyzed the data by creating a correlation matrix, calculating statistics, and visualizing the data.

Which stage of the ML pipeline is the company currently in?

Options:

A.

Data pre-processing

B.

Feature engineering

C.

Exploratory data analysis

D.

Hyperparameter tuning

Expert Solution
Viewing page 4 out of 11 pages
Viewing questions 31-40 out of questions