Winter Sale Special Limited Time 65% Discount Offer - Ends in 0d 00h 00m 00s - Coupon code: v4s65

AIF-C01 Exam Dumps - AWS Certified AI Practitioner Exam

Searching for workable clues to ace the Amazon Web Services AIF-C01 Exam? You’re on the right place! ExamCert has realistic, trusted and authentic exam prep tools to help you achieve your desired credential. ExamCert’s AIF-C01 PDF Study Guide, Testing Engine and Exam Dumps follow a reliable exam preparation strategy, providing you the most relevant and updated study material that is crafted in an easy to learn format of questions and answers. ExamCert’s study tools aim at simplifying all complex and confusing concepts of the exam and introduce you to the real exam scenario and practice it with the help of its testing engine and real exam dumps

Go to page:
Question # 9

An AI practitioner is using an Amazon Bedrock base model to summarize session chats from the customer service department. The AI practitioner wants to store invocation logs to monitor model input and output data.

A.

Configure AWS CloudTrail as the logs destination for the model.

B.

Enable model invocation logging in Amazon Bedrock.

C.

Configure AWS Audit Manager as the logs destination for the model.

D.

Configure model invocation logging in Amazon EventBridge.

Full Access
Question # 10

A large retail bank wants to develop an ML system to help the risk management team decide on loan allocations for different demographics.

What must the bank do to develop an unbiased ML model?

A.

Reduce the size of the training dataset.

B.

Ensure that the ML model predictions are consistent with historical results.

C.

Create a different ML model for each demographic group.

D.

Measure class imbalance on the training dataset. Adapt the training process accordingly.

Full Access
Question # 11

A company wants to make a chatbot to help customers. The chatbot will help solve technical problems without human intervention. The company chose a foundation model (FM) for the chatbot. The chatbot needs to produce responses that adhere to company tone.

Which solution meets these requirements?

A.

Set a low limit on the number of tokens the FM can produce.

B.

Use batch inferencing to process detailed responses.

C.

Experiment and refine the prompt until the FM produces the desired responses.

D.

Define a higher number for the temperature parameter.

Full Access
Question # 12

A retail company wants to build an ML model to recommend products to customers. The company wants to build the model based on responsible practices. Which practice should the company apply when collecting data to decrease model bias?

A.

Use data from only customers who match the demography of the company's overall customer base.

B.

Collect data from customers who have a past purchase history.

C.

Ensure that the data is balanced and collected from a diverse group.

D.

Ensure that the data is from a publicly available dataset.

Full Access
Question # 13

A company wants to use AI to protect its application from threats. The AI solution needs to check if an IP address is from a suspicious source.

A.

Build a speech recognition system.

B.

Create a natural language processing (NLP) named entity recognition system.

C.

Develop an anomaly detection system.

D.

Create a fraud forecasting system.

Full Access
Question # 14

A research company implemented a chatbot by using a foundation model (FM) from Amazon Bedrock. The chatbot searches for answers to questions from a large database of research papers.

After multiple prompt engineering attempts, the company notices that the FM is performing poorly because of the complex scientific terms in the research papers.

How can the company improve the performance of the chatbot?

A.

Use few-shot prompting to define how the FM can answer the questions.

B.

Use domain adaptation fine-tuning to adapt the FM to complex scientific terms.

C.

Change the FM inference parameters.

D.

Clean the research paper data to remove complex scientific terms.

Full Access
Question # 15

A company has developed a large language model (LLM) and wants to make the LLM available to multiple internal teams. The company needs to select the appropriate inference mode for each team.

Select the correct inference mode from the following list for each use case. Each inference mode should be selected one or more times. (Select THREE.)

* Batch transform

* Real-time inference

Full Access
Question # 16

A bank is fine-tuning a large language model (LLM) on Amazon Bedrock to assist customers with questions about their loans. The bank wants to ensure that the model does not reveal any private customer data.

Which solution meets these requirements?

A.

Use Amazon Bedrock Guardrails.

B.

Remove personally identifiable information (PII) from the customer data before fine-tuning the LLM.

C.

Increase the Top-K parameter of the LLM.

D.

Store customer data in Amazon S3. Encrypt the data before fine-tuning the LLM.

Full Access
Go to page: