AWS AI Practitioner Certification Exam Dumps and Braindumps

AWS AI Practitioner Exam Simulator

Despite the title of this article, this isn’t an AWS AI Practitioner exam braindump in the traditional sense.

I don’t believe in cheating.

A true braindump is when someone takes the actual exam and tries to rewrite every question they remember, essentially dumping the test content online. That’s unethical and a clear violation of AWS’s exam policies. There’s no integrity or value in that approach.

This set of AWS AI Practitioner exam questions is nothing like that.

Much better than a certification exam dump!

All of the questions here come from my AWS AI Practitioner Udemy course and from my AWS certification site: certificationexams.pro. It hosts hundreds of original practice questions designed around AWS certification objectives and AWS AI Practitioner exam topics, so be sure to check it out if you want to be thoroughly prepared for the exam.

But as I was saying, this is not a true ‘exam dump.’ Each AI Practitioner certification question here has been thoughtfully written to align with the official exam guide, testing your understanding of AI and ML fundamentals, AWS AI services, responsible AI concepts, and practical business use cases, without ever copying or disclosing real AWS exam content.

The goal is to help you learn ethically, build real knowledge, and feel confident working with AWS AI tools like Amazon SageMaker, AWS Bedrock, Comprehend, Rekognition, and Lex.

If you can answer these questions with confidence, and understand why each option is right or wrong, you won’t just pass the AWS AI Practitioner exam. You’ll gain a true understanding of how certified AI solutions are designed, deployed, and managed on AWS.

So, call it a braindump if you like, but it’s really a smart, honest study companion created to help you think like a cloud AI professional.

These AWS AI Practitioner exam questions are designed to challenge you, but each one includes clear explanations, practical insights, and exam tips to help you succeed on test day.

Learn deeply, study smart, and good luck on your AWS AI Practitioner certification journey.

Git, GitHub & GitHub Copilot Certification Made Easy

Want to get certified on the most popular AI, ML & DevOps technologies of the day? These five resources will help you get GitHub certified in a hurry.

Get certified in the latest AI, ML and DevOps technologies. Advance your career today.

AWS AI Practitioner Exam Questions

A data scientist at HarborLink Wireless is developing a churn detection capability for a subscription platform and is partnering with an applied research group to choose the most effective model. The solution must identify customers who are likely to cancel within the next nine months. What should the data scientist ask the research group to do to best ensure the most appropriate model is selected?

  • ❏ A. List and validate potential data sources for the project

  • ❏ B. Precisely scope the churn prediction objective and define success criteria

  • ❏ C. Define the application’s target audience as broadly as possible

  • ❏ D. Establish strict budget limits for model training and tuning

Which SageMaker capability provides a single visual interface to import data from Amazon S3, prepare, explore, and visualize it for ML?

  • ❏ A. Amazon SageMaker Clarify

  • ❏ B. Amazon SageMaker Studio

  • ❏ C. SageMaker Data Wrangler

  • ❏ D. Amazon SageMaker Feature Store

A digital lending startup is migrating its credit risk models to AWS and wants to adopt responsible AI practices. The team needs a managed capability that can examine training data and model predictions for fairness and provide clear explanations for individual outcomes. Which SageMaker use case is the most appropriate for this requirement?

  • ❏ A. Automate hyperparameter tuning to improve model performance

  • ❏ B. Amazon SageMaker Model Monitor

  • ❏ C. Detect and quantify bias in datasets and model outputs and generate explainability reports

  • ❏ D. Build and deploy models using an entirely no-code interface

What is a key benefit of running generative AI workloads on AWS?

  • ❏ A. Automatic full interpretability for every AI output

  • ❏ B. Global scale with robust security and compliance for generative AI

  • ❏ C. Amazon Bedrock removes the need for fine-tuning in all cases

  • ❏ D. AWS guarantees all generative AI outputs are factually correct

An engineering team at BrightWave Commerce is building a production AI solution that pulls clickstream data, catalog metadata, and product images. They need a secure and scalable way to define, run, and track a repeatable five-stage process that prepares data, trains and evaluates three models, and then promotes the best version to staging and production across accounts. Which AWS service should they use to orchestrate this end-to-end machine learning workflow?

  • ❏ A. AWS Step Functions

  • ❏ B. Amazon SageMaker Clarify

  • ❏ C. Amazon SageMaker Pipelines

  • ❏ D. Amazon SageMaker Autopilot

What is the core difference between supervised and unsupervised learning?

  • ❏ A. Supervised is only clustering; unsupervised only regression

  • ❏ B. Supervised needs labels; unsupervised trains without any data

  • ❏ C. Supervised learns from labeled examples to predict outputs; unsupervised finds patterns in unlabeled data

  • ❏ D. Semi-supervised learning uses both labeled and unlabeled data

A data analytics startup named Northstar Insights runs workloads in three AWS Regions to satisfy local data residency obligations. The team needs to select services whose resources and APIs are limited to an AWS Region to ensure isolation and compliance. Which services satisfy this requirement? (Choose 2)

  • ❏ A. AWS WAF

  • ❏ B. Amazon Rekognition

  • ❏ C. AWS Identity and Access Management (AWS IAM)

  • ❏ D. AWS Lambda

  • ❏ E. Amazon CloudFront

Which AWS managed service best supports fast full-text indexing and vector similarity search for RAG over 50 million documents?

  • ❏ A. Amazon RDS for PostgreSQL

  • ❏ B. Amazon DynamoDB

  • ❏ C. Amazon OpenSearch Service

  • ❏ D. Amazon Neptune

A creative marketing agency named LumenWave Studio is building generative AI tools to draft product descriptions and automatically reply to customer messages. To track outcomes during a 90-day pilot and deploy models reliably, the team needs a clear distinction between model inference and model evaluation. Model inference uses a trained model to turn new inputs into outputs, while model evaluation measures how well the model performs against selected metrics. Which statements best distinguish model inference from model evaluation? (Choose 2)

  • ❏ A. Model evaluation benchmarks performance with metrics on held-out data, whereas model inference generates real-time or batch outputs from new inputs

  • ❏ B. Model inference and model evaluation describe the same activity of creating AI content

  • ❏ C. Model inference applies a trained model to produce predictions or responses, while model evaluation measures accuracy and quality using predefined metrics

  • ❏ D. Model evaluation only happens during training phases, and model inference is used exclusively after deployment

  • ❏ E. Model inference is the stage that trains new models, and model evaluation is what creates responses from the trained model

Which tasks are LLMs naturally best suited for?

  • ❏ A. Sub-millisecond OLTP in Amazon Aurora

  • ❏ B. Natural language tasks: generate text, Q&A, summarize docs, write code

  • ❏ C. Real-time robotics control loops

  • ❏ D. Amazon Rekognition

A data science team at Nebula Analytics is planning a 14-day evaluation to control training spend in Amazon SageMaker while still meeting deadlines. Which sequence orders these SageMaker training approaches from the lowest typical cost to the highest?

  • ❏ A. Multi-GPU Distributed Training, On-Demand Training, Spot Training

  • ❏ B. Spot Training, On-Demand Training, Multi-GPU Distributed Training

  • ❏ C. On-Demand Training, Spot Training, Multi-GPU Distributed Training

  • ❏ D. Spot Training, Multi-GPU Distributed Training, On-Demand Training

Which scenario best illustrates algorithmic bias in an ML system?

  • ❏ A. A SageMaker endpoint shows higher latency at peak traffic

  • ❏ B. A lending model consistently scores applicants from certain ZIP codes lower despite similar credit

  • ❏ C. A weather model sometimes misses storms due to natural variability

  • ❏ D. A human reviewer prioritizes cases using personal judgment

A fintech analytics startup called BlueLedger needs to uphold its responsible AI policy by automatically tracking production machine learning models for prediction accuracy, fairness, and data drift with scheduled evaluations and alerts. Which AWS service provides continuous monitoring for deployed models?

  • ❏ A. Amazon SageMaker Clarify

  • ❏ B. Amazon SageMaker Model Cards

  • ❏ C. Amazon CloudWatch

  • ❏ D. Amazon SageMaker Model Monitor

Which SageMaker inference option is best for per-request latency under 90 ms at about 15 million predictions per day?

  • ❏ A. SageMaker Batch Transform

  • ❏ B. SageMaker Real-Time Endpoints

  • ❏ C. SageMaker Serverless Inference

  • ❏ D. SageMaker Asynchronous Inference

The corporate legal team at Sierra Financial Group plans to build an AI solution powered by large language models to process contracts and litigation filings. Each document averages about 150 pages, and the team reviews roughly 25,000 pages per month. They want the system to condense the content and surface the main points so attorneys can work faster. Which approach would best meet this need?

  • ❏ A. Build a multilingual translation workflow

  • ❏ B. Amazon Comprehend

  • ❏ C. Implement an LLM-based document summarization assistant

  • ❏ D. Create a recommendation engine for related cases

What is the key difference between Amazon Mechanical Turk and Amazon SageMaker Ground Truth for ML data labeling and workforce management?

  • ❏ A. SageMaker Ground Truth can only use MTurk workers and cannot use private or vendor workforces

  • ❏ B. Mechanical Turk is a general on-demand crowd marketplace, while SageMaker Ground Truth is ML-focused labeling with automation and managed workforces

  • ❏ C. Amazon Augmented AI replaces Ground Truth and removes the need for human workforces

  • ❏ D. Mechanical Turk specializes in ML labeling with auto-labeling, and Ground Truth is a general human-task marketplace

A digital media analytics company is training a large foundation model on AWS using a mix of instance families and pricing options. How can the team keep the training running with minimal disruption while controlling compute spend?

  • ❏ A. Use only On-Demand Instances to ensure capacity

  • ❏ B. Run exclusively on Spot Instances and restart training whenever there is an interruption

  • ❏ C. Configure an automated policy that prefers Spot but falls back to On-Demand when Spot is unavailable, and enable periodic checkpointing

  • ❏ D. Purchase Compute Savings Plans for all training workloads to avoid interruptions

Which AWS service provides pre-trained image labeling to categorize 50,000 product photos per day without training a model?

  • ❏ A. Amazon Textract

  • ❏ B. Amazon Bedrock

  • ❏ C. Amazon Rekognition

  • ❏ D. Amazon SageMaker

A compliance software vendor is adopting Amazon Bedrock to summarize lengthy policies and auto respond to customer tickets within 90 seconds. The architects are evaluating Retrieval Augmented Generation and Bedrock Agents to decide which approach will deliver accurate, context aware answers and task automation. What best describes how RAG differs from an Agent in Amazon Bedrock?

  • ❏ A. Both RAG and Agents are primarily about fetching context from knowledge stores to improve a prompt response

  • ❏ B. Agent means retrieving facts from a datastore to enrich output, while RAG is a workflow that plans steps and invokes tools

  • ❏ C. RAG focuses on retrieving and injecting relevant external context into the model, while an Agent coordinates multi step reasoning and tool or API calls using a foundation model

  • ❏ D. Both RAG and Agents are planning oriented systems that iteratively interpret inputs and act without external retrieval

Which capability provides per-output rationales explaining why a generative summary omitted specific text?

  • ❏ A. Amazon SageMaker Clarify

  • ❏ B. Output-level explainability

  • ❏ C. Amazon Bedrock Guardrails

  • ❏ D. Amazon SageMaker Model Cards

A startup in online education named BrightCourse plans to migrate growing workloads to AWS and is evaluating service models to understand how much control they retain. The team plans to use Amazon EC2 for scalable compute and wants clarity on the category this service fits into so they can plan how to manage operating systems, networking, and storage. Which cloud service model does Amazon EC2 represent?

  • ❏ A. Software as a Service (SaaS)

  • ❏ B. Platform as a Service (PaaS)

  • ❏ C. Infrastructure as a Service (IaaS)

  • ❏ D. Network as a Service (NaaS)

How can Amazon Bedrock adapt a foundation model to understand domain-specific jargon and acronyms? (Choose 2)

  • ❏ A. Train a new foundation model from scratch using raw documents

  • ❏ B. Managed fine-tuning on labeled domain data with AWS handling infrastructure

  • ❏ C. Guardrails for Amazon Bedrock change model weights to enforce safety

  • ❏ D. Prompt engineering with few-shot examples without altering weights

  • ❏ E. Knowledge Bases for Amazon Bedrock retrain the model when documents are ingested

EcoVision Labs manages a library of over 60,000 trail-camera images and wants to automatically find and label which animals appear in each photo without human review. Which approach should they use?

  • ❏ A. Named entity recognition

  • ❏ B. Image inpainting

  • ❏ C. Object detection

  • ❏ D. Anomaly detection

Which scenarios are best suited for Retrieval-Augmented Generation with Amazon Bedrock? (Choose 2)

  • ❏ A. Amazon Personalize

  • ❏ B. Domain-specific Q&A assistant

  • ❏ C. Fine-tuning a foundation model

  • ❏ D. Support Q&A assistant

  • ❏ E. Image generation from prompts

A research publisher uses a generative AI service to create 120-word abstracts from 30-page whitepapers, and editors compare these outputs to human-written reference summaries to ensure key information is retained. Which evaluation metric best measures the quality of these summaries against the references?

  • ❏ A. BLEU score

  • ❏ B. ROUGE score

  • ❏ C. F1 score

  • ❏ D. ROC AUC

Which AWS service efficiently cleans, normalizes, and applies repeatable transformations to a 30 TB dataset in Amazon S3 for ML training?

  • ❏ A. Amazon SageMaker Model Monitor

  • ❏ B. Amazon SageMaker Data Wrangler

  • ❏ C. AWS Glue DataBrew

  • ❏ D. Amazon SageMaker Ground Truth

A regional fintech startup called LumaPay has rolled out a generative AI assistant to answer questions about balances, loan offerings, and recent transactions. Users have begun attempting prompt injection by asking the assistant to disregard its internal rules and to produce unreliable financial guidance. The engineering team wants the most effective control to ensure the assistant adheres to system instructions and delivers safe, trustworthy outputs. What should the team implement?

  • ❏ A. Use a larger foundation model with more parameters

  • ❏ B. Enable conversation memory to retain the full chat context

  • ❏ C. Apply strict validation and sanitization to user prompts before the model processes them

  • ❏ D. AWS WAF

Which AWS service provides managed human review workflows to add about 30 reviewers to ML predictions without building custom applications?

  • ❏ A. Amazon Forecast

  • ❏ B. Amazon Mechanical Turk

  • ❏ C. Amazon Augmented AI (A2I)

  • ❏ D. Amazon SageMaker Ground Truth

A small game studio using Amazon Bedrock wants to create concept art with generative AI. They heard diffusion models work well but are unclear about the fundamental mechanism. Which statement best describes how diffusion models generate images?

  • ❏ A. Diffusion models search the training corpus for the closest visual matches using vector similarity and return them

  • ❏ B. Diffusion models refine pictures by applying supervised filters that accentuate boundaries between pixels

  • ❏ C. Diffusion models synthesize images by starting from pure noise and repeatedly denoising with a learned reverse process

  • ❏ D. Diffusion models build images primarily through an encoder-decoder with attention layers

Which approach enforces least privilege so only approved users can invoke selected model endpoints and read specific Amazon S3 datasets?

  • ❏ A. Only resource-based policies

  • ❏ B. IAM roles with least-privilege, resource-scoped permissions to specific endpoints and S3 data

  • ❏ C. One shared IAM policy for all users and resources

  • ❏ D. Use AWS Organizations SCPs as the only control

An edtech startup, LumenLearn, is mapping Amazon SageMaker tools to common ML tasks for a new project. Match each service to the task it best fits: A) SageMaker Data Wrangler B) SageMaker Canvas C) SageMaker Ground Truth 1) Aggregates human input for labeling, review, and evaluation to enhance model relevance across the ML lifecycle 2) Provides hundreds of built-in transformations to prepare data sets for machine learning 3) A no-code, point-and-click interface for creating predictions from data. Which mapping is correct?

  • ❏ A. A-3, B-2, C-1

  • ❏ B. A-2, B-1, C-3

  • ❏ C. A-2, B-3, C-1

  • ❏ D. A-1, B-3, C-2

What is the term for converting words or concepts into dense numeric vectors that encode semantic meaning?

  • ❏ A. Tokens

  • ❏ B. Amazon Bedrock

  • ❏ C. Text embeddings

  • ❏ D. Positional encodings

RiverNorth Mobile plans to release a virtual support assistant that resolves device and connectivity issues without handing off to human agents. The team has selected a foundation model on AWS and must ensure every reply reflects the company’s brand voice and writing style. What is the most effective way to accomplish this?

  • ❏ A. Increase the temperature parameter to produce more varied wording

  • ❏ B. Configure a low maximum token output to constrain responses

  • ❏ C. Use prompt engineering with clear style guidelines and iterative refinement

  • ❏ D. Perform batch inference to generate comprehensive responses

For evaluating LLM-generated summaries to ensure key information coverage while keeping them brief, which metric should be prioritized?

  • ❏ A. BERTScore

  • ❏ B. BLEU

  • ❏ C. ROUGE-L (recall-focused)

  • ❏ D. Compression ratio

A regional wealth management firm, Orion Ledger, uses a generative AI system to produce client portfolio briefings every 30 days. Clients have recently reported that some write-ups include incorrect return percentages and misleading totals. The firm wants to raise confidence in the AI-generated reports. Which responsible AI focus best aligns with this goal?

  • ❏ A. Promotes fairness by reducing bias in data and outcomes

  • ❏ B. AWS Key Management Service

  • ❏ C. Ensures generated content is factually correct and grounded in trustworthy sources

  • ❏ D. Improves interpretability so model reasoning is easier to understand

AWS AI Practitioner Practice Test Answers

A data scientist at HarborLink Wireless is developing a churn detection capability for a subscription platform and is partnering with an applied research group to choose the most effective model. The solution must identify customers who are likely to cancel within the next nine months. What should the data scientist ask the research group to do to best ensure the most appropriate model is selected?

  • ✓ B. Precisely scope the churn prediction objective and define success criteria

Precisely scope the churn prediction objective and define success criteria is the correct choice because it forces the team to agree on the nine month horizon, the customer segments to consider, and the measurable metrics that will determine which model performs best.

Defining the objective and success criteria lets the research group design experiments that target the business question and avoid common pitfalls. By fixing the time window and the positive class definition you can choose appropriate evaluation measures such as recall for churners or business-weighted cost metrics. Clear criteria also guide feature selection, cross validation strategies that respect temporal ordering, and decisions about class imbalance so that model comparisons are fair and directly tied to outcomes the business cares about.

List and validate potential data sources for the project is useful early because good inputs are necessary for any model but it does not by itself guarantee that the most appropriate model will be chosen for the stated nine month objective.

Define the application’s target audience as broadly as possible is counterproductive because a broad audience increases variance in behavior and makes evaluation and optimization less precise which hinders selecting a model that meets specific business goals.

Establish strict budget limits for model training and tuning is a practical constraint that can limit experimentation but it does not determine which model is most appropriate for the problem and the agreed success metrics.

Before testing algorithms narrow the prediction scope and lock the success metrics so that model comparisons are meaningful and experiments move faster.

Which SageMaker capability provides a single visual interface to import data from Amazon S3, prepare, explore, and visualize it for ML?

  • ✓ C. SageMaker Data Wrangler

SageMaker Data Wrangler is the correct option because it provides a single visual interface to import data from Amazon S3, prepare the data, explore and visualize it, and then export prepared datasets and features for training or pipelines.

SageMaker Data Wrangler is a native SageMaker tool that offers point and click transforms, built in visualizations, and connectors to common data sources so you can perform cleaning and feature engineering without heavy coding and then export results to training jobs or SageMaker Pipelines.

Amazon SageMaker Clarify is incorrect because it focuses on bias detection and model explainability rather than general interactive data preparation and visualization.

Amazon SageMaker Studio is incorrect because Studio is the overall integrated development environment that hosts tools like Data Wrangler and it is not itself the dedicated visual data preparation tool.

Amazon SageMaker Feature Store is incorrect because the Feature Store is for managing and serving features at runtime and it does not provide an interactive visual interface for cleaning and exploring raw data.

Focus on key phrases such as single visual interface and clean/explore/visualize to map the requirement to Data Wrangler.

A digital lending startup is migrating its credit risk models to AWS and wants to adopt responsible AI practices. The team needs a managed capability that can examine training data and model predictions for fairness and provide clear explanations for individual outcomes. Which SageMaker use case is the most appropriate for this requirement?

  • ✓ C. Detect and quantify bias in datasets and model outputs and generate explainability reports

The most appropriate choice is Detect and quantify bias in datasets and model outputs and generate explainability reports. This capability maps to SageMaker Clarify and it is intended to examine training data and model predictions for fairness and to provide clear explanations for individual outcomes.

SageMaker Clarify computes quantitative bias metrics on datasets and on model predictions and it produces feature attributions and explainability reports that help interpret why a model made a particular prediction. It supports analysis before training and after training and it can be integrated into model pipelines to help maintain fairness and transparency across the ML lifecycle.

Automate hyperparameter tuning to improve model performance refers to SageMaker Automatic Model Tuning and it focuses on searching hyperparameter space to optimize model metrics rather than measuring bias or producing per prediction explanations.

Amazon SageMaker Model Monitor is intended to detect data and model quality drift in production and it can notify you about distribution changes. It does not provide the detailed bias metrics and per outcome explainability that Clarify offers.

Build and deploy models using an entirely no-code interface describes services such as SageMaker Canvas or Autopilot and they simplify model creation and deployment but they are not the managed capability for bias detection and model explainability.

When an exam scenario mentions fairness metrics or interpreting individual predictions think SageMaker Clarify. If the scenario mentions production drift think Model Monitor and if it mentions hyperparameter search think Automatic Model Tuning.

What is a key benefit of running generative AI workloads on AWS?

  • ✓ B. Global scale with robust security and compliance for generative AI

The correct choice is Global scale with robust security and compliance for generative AI. AWS provides a worldwide infrastructure footprint for low latency access and it offers mature identity controls encryption options network controls and a broad set of compliance programs that make it a strong platform for enterprise generative AI.

The Global scale with robust security and compliance for generative AI option is right because global regions and edge locations reduce latency and enable deployment near users. AWS security services such as IAM encryption and networking controls protect data and workloads and the compliance certifications help meet regulatory requirements. Together these capabilities support scalable secure governance for production generative AI workloads.

The option Automatic full interpretability for every AI output is incorrect because explainability depends on the model and tooling and it is not automatically available for every output. Some models provide better explainability features but full automatic interpretability across all models and outputs is not guaranteed.

The option Amazon Bedrock removes the need for fine-tuning in all cases is incorrect because many workloads still benefit from fine tuning prompt engineering or retrieval augmentation. Amazon Bedrock simplifies access to and management of foundation models but it does not remove the need to adapt models for specific tasks.

The option AWS guarantees all generative AI outputs are factually correct is incorrect because no vendor can guarantee perfect factual accuracy of model outputs. Users must validate results and apply testing guardrails and augmentation techniques to reduce hallucinations.

Focus on choices that call out AWS strengths such as global infrastructure security and compliance and avoid answers that use absolute words like always or guaranteed.

An engineering team at BrightWave Commerce is building a production AI solution that pulls clickstream data, catalog metadata, and product images. They need a secure and scalable way to define, run, and track a repeatable five-stage process that prepares data, trains and evaluates three models, and then promotes the best version to staging and production across accounts. Which AWS service should they use to orchestrate this end-to-end machine learning workflow?

  • ✓ C. Amazon SageMaker Pipelines

The correct choice is Amazon SageMaker Pipelines. This service is specifically designed to define run and track repeatable multi stage machine learning workflows and to promote models across environments.

Amazon SageMaker Pipelines provides native capabilities for composing the five stage process in the scenario including data preparation training and evaluation. It integrates with the SageMaker model registry to record experiment metadata capture lineage and manage model versions so teams can securely promote the best model to staging and production across accounts.

Amazon SageMaker Pipelines also scales with SageMaker processing and training jobs and supports IAM controls and cross account model sharing which makes it suitable for production deployments that require repeatability and governance.

AWS Step Functions can orchestrate steps across many AWS services but it does not offer ML native features like built in experiment tracking lineage or a model registry which the scenario requires.

Amazon SageMaker Clarify focuses on bias detection and explainability for datasets and trained models and it does not orchestrate full multi step ML pipelines or handle model promotion.

Amazon SageMaker Autopilot automates model building for a single dataset and it does not provide the workflow orchestration experiment management and cross account model promotion capabilities needed for a multi model multi stage production pipeline.

Look for phrases like end-to-end ML workflow and model registry in the question and choose the service that natively supports experiment tracking lineage and model promotion.

What is the core difference between supervised and unsupervised learning?

  • ✓ C. Supervised learns from labeled examples to predict outputs; unsupervised finds patterns in unlabeled data

Supervised learns from labeled examples to predict outputs; unsupervised finds patterns in unlabeled data is correct because it describes the fundamental difference between having target labels and not having them during training.

In supervised learning models are trained on input and output pairs so they learn to map features to known targets and they are used for tasks such as classification and regression where the objective is to predict labels or continuous values.

In unsupervised learning there are no target labels so algorithms search for structure within the data and typical outcomes include clusters associations and lower dimensional representations that support tasks like clustering anomaly detection and feature learning.

Supervised is only clustering; unsupervised only regression is incorrect because clustering is usually an unsupervised task and regression is a supervised task and neither paradigm is limited to a single task type.

Supervised needs labels; unsupervised trains without any data is incorrect because unsupervised methods do use data they simply do not require labeled targets.

Semi-supervised learning uses both labeled and unlabeled data is a true statement but it does not answer the question because it does not define the core distinction between supervised and unsupervised learning.

Look for the words labeled versus unlabeled. If the goal is to predict known targets choose supervised and if the goal is to discover structure choose unsupervised.

A data analytics startup named Northstar Insights runs workloads in three AWS Regions to satisfy local data residency obligations. The team needs to select services whose resources and APIs are limited to an AWS Region to ensure isolation and compliance. Which services satisfy this requirement? (Choose 2)

  • ✓ B. Amazon Rekognition

  • ✓ D. AWS Lambda

The correct choices are Amazon Rekognition and AWS Lambda. Both services are scoped to a single AWS Region and you select the Region for resources and call Region specific endpoints to support data residency and isolation requirements.

AWS Lambda functions are created and executed in the Region you choose and their resources such as function code and environment data remain in that Region unless you move them intentionally. Amazon Rekognition processes images and video through Region endpoints and stores any service data in that Region so it aligns with regional compliance needs.

AWS WAF is not confined to a single Region when it is associated with Amazon CloudFront. In that scenario the web ACL is managed at the CloudFront distribution level and applies globally across edge locations rather than being Region limited.

AWS Identity and Access Management (AWS IAM) is a global service and its users roles and policies operate across all Regions in an account so it does not by itself provide Region isolation.

Amazon CloudFront is a global content delivery network that uses edge locations around the world to serve content and it is not tied to a single Region.

Remember that global services like AWS IAM and Amazon CloudFront span Regions while most compute and AI services such as AWS Lambda and Amazon Rekognition are Region scoped.

Which AWS managed service best supports fast full-text indexing and vector similarity search for RAG over 50 million documents?

  • ✓ C. Amazon OpenSearch Service

The correct option is Amazon OpenSearch Service. It natively combines full-text indexing, relevance ranking, and vector similarity search and it scales horizontally to handle tens of millions of documents for retrieval augmented generation workloads.

Amazon OpenSearch Service provides inverted indexes and BM25 style relevance ranking for keyword search and it offers k-NN and serverless vector capabilities for embedding based semantic retrieval. The service is built for distributed search with sharding and replicas and it enables efficient retrieval and ranking by semantic relevance across large corpora, which is why it is the best fit for fast full-text indexing and vector similarity search at this scale.

Amazon RDS for PostgreSQL can host extensions such as pgvector to store embeddings but it lacks search engine features like optimized inverted indexes, built-in relevance ranking, and distributed search optimizations, so it is less suitable for tens of millions of passages.

Amazon DynamoDB is a key-value and document store and it does not provide native full-text indexing or vector similarity search, so it cannot deliver ranked semantic retrieval for RAG workloads at this scale.

Amazon Neptune focuses on graph queries and relationship traversals rather than large-scale text indexing or vector similarity search, so it is not well matched to RAG use cases that require fast full-text and embedding-based search.

Focus on keywords such as full-text indexing and vector similarity and prefer managed search services when the question specifies large scale and relevance ranking.

A creative marketing agency named LumenWave Studio is building generative AI tools to draft product descriptions and automatically reply to customer messages. To track outcomes during a 90-day pilot and deploy models reliably, the team needs a clear distinction between model inference and model evaluation. Model inference uses a trained model to turn new inputs into outputs, while model evaluation measures how well the model performs against selected metrics. Which statements best distinguish model inference from model evaluation? (Choose 2)

  • ✓ A. Model evaluation benchmarks performance with metrics on held-out data, whereas model inference generates real-time or batch outputs from new inputs

  • ✓ C. Model inference applies a trained model to produce predictions or responses, while model evaluation measures accuracy and quality using predefined metrics

The correct choices are Model evaluation benchmarks performance with metrics on held-out data, whereas model inference generates real-time or batch outputs from new inputs and Model inference applies a trained model to produce predictions or responses, while model evaluation measures accuracy and quality using predefined metrics.

These answers are correct because model inference describes applying a trained model to new inputs to produce predictions or responses in real time or in batches and model evaluation describes measuring model behavior against held out datasets and predefined metrics so teams can quantify accuracy and quality and compare model variants.

Keeping inference and evaluation separate lets LumenWave Studio run production traffic and collect outputs while also running offline and online evaluations and monitoring so the team can track outcomes reliably during the 90 day pilot and validate deployment changes.

The statement Model inference and model evaluation describe the same activity of creating AI content is incorrect because the two activities serve different goals and workflows and they are not the same.

The statement Model evaluation only happens during training phases, and model inference is used exclusively after deployment is incorrect because model evaluation can occur before and after deployment and inference is used during testing as well as in production.

The statement Model inference is the stage that trains new models, and model evaluation is what creates responses from the trained model is incorrect because training is a separate phase and evaluation measures model quality rather than generating responses.

Focus on the operational roles so remember that inference is runtime prediction and evaluation is metric based assessment for comparing models and monitoring production.

Which tasks are LLMs naturally best suited for?

  • ✓ B. Natural language tasks: generate text, Q&A, summarize docs, write code

Natural language tasks: generate text, Q&A, summarize docs, write code is correct because large language models are trained on vast collections of text and are optimized for understanding and producing human language across many formats.

These models learn patterns of language and context so they can produce coherent text, extract answers from documents, condense long content, and suggest or generate code. This aligns with managed services and ecosystems that host and integrate LLMs such as Amazon Bedrock and developer tools for code assistance.

Sub-millisecond OLTP in Amazon Aurora is incorrect because transactional database workloads require deterministic low latency and ACID guarantees that database engines provide and LLM inference does not.

Real-time robotics control loops is incorrect because closed loop control requires microsecond or millisecond deterministic timing and predictable behavior and LLM inference is comparatively variable and not designed for control tasks.

Amazon Rekognition is incorrect because that service targets computer vision tasks such as image and video analysis and it is not intended for natural language generation or understanding.

Important map keywords like generate text, Q&A, summarize, and code generation to LLMs and services such as Amazon Bedrock when you choose answers on the exam.

A data science team at Nebula Analytics is planning a 14-day evaluation to control training spend in Amazon SageMaker while still meeting deadlines. Which sequence orders these SageMaker training approaches from the lowest typical cost to the highest?

  • ✓ B. Spot Training, On-Demand Training, Multi-GPU Distributed Training

The correct option is Spot Training, On-Demand Training, Multi-GPU Distributed Training because this sequence orders approaches from the lowest typical cost to the highest for a short evaluation window.

Spot Training is generally the least expensive since it uses discounted spare EC2 capacity which reduces hourly charges. On-Demand Training is in the middle because it runs stable instances at standard pricing and avoids interruptions for predictable cost. Multi-GPU Distributed Training is usually the most expensive since it runs multiple high performance GPU instances and adds networking overhead which raises total training spend.

Multi-GPU Distributed Training, On-Demand Training, Spot Training is wrong because it places the most costly approach first and the cheapest last which does not match typical pricing behavior.

On-Demand Training, Spot Training, Multi-GPU Distributed Training is incorrect because it ranks Spot higher cost than On-Demand while Spot is normally cheaper due to discounted spare capacity.

Spot Training, Multi-GPU Distributed Training, On-Demand Training is incorrect because it puts distributed multi GPU ahead of On-Demand even though multi GPU distributed setups usually increase cost relative to single on demand instances.

For cost ordering questions remember that Spot is usually the cheapest, On-Demand sits in the middle, and multi GPU distributed setups tend to be the most expensive. Read wording carefully when the exam asks about minimizing cost within a fixed time window.

Which scenario best illustrates algorithmic bias in an ML system?

  • ✓ B. A lending model consistently scores applicants from certain ZIP codes lower despite similar credit

The correct option is A lending model consistently scores applicants from certain ZIP codes lower despite similar credit. This choice describes algorithmic bias because it shows a systematic pattern where the model produces disparate outcomes for specific groups even when relevant qualifications are comparable.

This situation indicates the model may be using ZIP code as a proxy for protected characteristics and producing unfair results across groups. Such patterns are what practitioners look for when diagnosing algorithmic bias and they call for fairness metrics and explainability to identify which features drive the disparity. On AWS you can use Amazon SageMaker Clarify to detect pre-training and post-training bias metrics and to generate explainability reports that reveal proxy features driving unfair outcomes.

A SageMaker endpoint shows higher latency at peak traffic is incorrect because it describes a performance or scalability issue rather than a fairness or bias problem in model outputs.

A weather model sometimes misses storms due to natural variability is incorrect because occasional mispredictions reflect stochastic error and uncertainty rather than systematic group-level disparity.

A human reviewer prioritizes cases using personal judgment is incorrect because it describes potential human bias and decision making rather than bias produced by an automated algorithm.

Look for systematic differences in outcomes across groups when qualifications are similar as that pattern usually signals algorithmic bias.

A fintech analytics startup called BlueLedger needs to uphold its responsible AI policy by automatically tracking production machine learning models for prediction accuracy, fairness, and data drift with scheduled evaluations and alerts. Which AWS service provides continuous monitoring for deployed models?

  • ✓ D. Amazon SageMaker Model Monitor

Amazon SageMaker Model Monitor is the correct option because it provides continuous monitoring of deployed models for prediction accuracy fairness and data drift and it supports scheduled evaluations with alerting.

Amazon SageMaker Model Monitor is purpose built to detect data quality and model quality issues by comparing production traffic to baselines and it can run configurable monitoring jobs on endpoints and emit metrics for alerting so teams can surface problems quickly.

Amazon SageMaker Clarify helps with bias analysis and explainability during training and evaluation but it does not provide continuous scheduled monitoring on live endpoints in the same operational way as the Model Monitor.

Amazon SageMaker Model Cards provides governance and documentation about model attributes and intended usage and it is focused on transparency rather than running ongoing operational checks on live predictions.

Amazon CloudWatch collects infrastructure and application telemetry and can host alarms and dashboards but it does not include native ML workflows for data drift bias or automated model quality monitoring that compare live traffic to baselines.

Watch for phrases like continuous monitoring drift and data quality when the question asks about automated checks on deployed models.

Which SageMaker inference option is best for per-request latency under 90 ms at about 15 million predictions per day?

  • ✓ B. SageMaker Real-Time Endpoints

SageMaker Real-Time Endpoints is the best choice for per-request latency under 90 ms at about 15 million predictions per day because it provides persistent, auto-scalable endpoints that deliver consistent, low latency for synchronous inference.

The real-time approach runs long lived model servers and supports autoscaling and provisioned concurrency so it can sustain high requests per second and avoid cold starts that add latency. Real-time endpoints are purpose built for synchronous, per-request inference and offer predictable tail latency when the instance types and scaling policies are sized to the workload.

SageMaker Batch Transform is intended for offline batch jobs that process large datasets and does not serve synchronous per-request traffic, so it cannot meet strict per-request latency requirements.

SageMaker Serverless Inference removes server management but can experience cold-starts and variable latency that risk missing tight SLAs at high throughput, so it is not the best choice for guaranteed sub-90 ms per-request latency.

SageMaker Asynchronous Inference uses a queued, job style model and is built for long running or bursty requests that can tolerate higher latency, which makes it unsuitable for strict, synchronous per-request timing needs.

For strict per-request latency and steady high TPS choose SageMaker Real-Time Endpoints and validate instance types and autoscaling settings to reliably meet the sub-90 ms requirement.

The corporate legal team at Sierra Financial Group plans to build an AI solution powered by large language models to process contracts and litigation filings. Each document averages about 150 pages, and the team reviews roughly 25,000 pages per month. They want the system to condense the content and surface the main points so attorneys can work faster. Which approach would best meet this need?

  • ✓ C. Implement an LLM-based document summarization assistant

The best fit is Implement an LLM-based document summarization assistant. This approach directly meets the requirement to condense long legal documents and surface the main points so attorneys can work faster.

Implement an LLM-based document summarization assistant can produce concise extractive or abstractive summaries and highlight action items and key facts. It works with long documents by chunking text and using retrieval augmented generation with embeddings so the assistant can reference the full corpus without exceeding model context. This approach scales to tens of thousands of pages per month and supports quality controls such as human review and iterative prompting to ensure legal accuracy.

Build a multilingual translation workflow focuses on converting text between languages and does not reduce or condense content or extract the most relevant points for legal review.

Create a recommendation engine for related cases is intended to rank or suggest related items and it does not extract or summarize the contents of a single long document.

Amazon Comprehend can identify entities and key phrases but it does not produce comprehensive, coherent summaries of lengthy filings and it is not optimized for abstractive summarization the way an LLM assistant is.

When the goal is to condense long text and surface key points choose an LLM summarization approach. Use chunking and retrieval augmented generation to handle very long documents and add human review for legal accuracy.

What is the key difference between Amazon Mechanical Turk and Amazon SageMaker Ground Truth for ML data labeling and workforce management?

  • ✓ B. Mechanical Turk is a general on-demand crowd marketplace, while SageMaker Ground Truth is ML-focused labeling with automation and managed workforces

Mechanical Turk is a general on-demand crowd marketplace, while SageMaker Ground Truth is ML-focused labeling with automation and managed workforces is correct.

The statement is correct because Mechanical Turk provides a flexible public marketplace for many types of human tasks and it scales to a broad crowd. The service is general purpose and it is not specialized for machine learning workflows. By contrast SageMaker Ground Truth is purpose built for creating labeled training datasets and it adds features such as automated pre-labeling, managed labeling workflows, quality controls, and direct support for private, vendor, or MTurk workforces.

SageMaker Ground Truth can only use MTurk workers and cannot use private or vendor workforces is wrong because Ground Truth supports multiple workforce types including private teams and vendors from the AWS Marketplace in addition to MTurk. The statement misrepresents the service capability.

Amazon Augmented AI replaces Ground Truth and removes the need for human workforces is incorrect because Amazon Augmented AI is designed to add human review into inference or prediction workflows and it complements labeling and review tasks rather than replacing Ground Truth or eliminating human workforces.

Mechanical Turk specializes in ML labeling with auto-labeling, and Ground Truth is a general human-task marketplace is incorrect because it reverses the actual roles. Ground Truth is the ML-focused labeling service with automation features and Mechanical Turk is the general crowd marketplace.

Look for keywords like auto-labeling and labeling workflows to pick Ground Truth and watch for phrases like general human task marketplace to identify Mechanical Turk.

A digital media analytics company is training a large foundation model on AWS using a mix of instance families and pricing options. How can the team keep the training running with minimal disruption while controlling compute spend?

  • ✓ C. Configure an automated policy that prefers Spot but falls back to On-Demand when Spot is unavailable, and enable periodic checkpointing

Configure an automated policy that prefers Spot but falls back to On-Demand when Spot is unavailable, and enable periodic checkpointing is the correct choice because it combines the cost benefits of Spot with the reliability of On Demand and it preserves training progress through checkpoints.

This automated policy uses Spot instances when capacity and price make sense and it automatically shifts to On Demand when Spot capacity is reclaimed. Checkpointing saves model state periodically so training can resume from the last checkpoint rather than restarting from scratch. Together this approach minimizes both spend and disruption for long running foundation model training.

Use only On-Demand Instances to ensure capacity is not ideal because it does give stable capacity but it prevents you from taking advantage of lower cost Spot capacity which raises total training cost for large workloads.

Run exclusively on Spot Instances and restart training whenever there is an interruption is not ideal because Spot alone can produce frequent interruptions and lost progress unless you add automated failover and checkpointing which this option does not include.

Purchase Compute Savings Plans for all training workloads to avoid interruptions is misleading because Savings Plans affect pricing and reduce cost for consistent usage but they do not change Spot interruption behavior or provide capacity guarantees.

Favor a mixed strategy that uses Spot with checkpointing and automatic fallback to On Demand so you reduce cost and keep training running with minimal disruption.

Which AWS service provides pre-trained image labeling to categorize 50,000 product photos per day without training a model?

  • ✓ C. Amazon Rekognition

Amazon Rekognition is correct because it provides pre-trained image analysis and label detection that can categorize large volumes of product photos without requiring you to train a model.

Amazon Rekognition offers managed APIs for detecting objects, scenes, and labels and it supports batch and asynchronous processing so it can handle tens of thousands of images per day for automated product photo categorization.

Amazon Textract is incorrect because it is designed to extract text and structured data from documents and scanned pages rather than to perform general image labeling of objects and scenes.

Amazon Bedrock is incorrect because it provides access to foundation models and generative AI capabilities and does not offer a native pre-trained image labeling API for bulk photo categorization.

Amazon SageMaker is incorrect because it is a platform for building, training, and deploying custom machine learning models and it therefore requires training a model which the question explicitly excludes.

When a question emphasizes pre-trained, no model training, and image labeling choose Rekognition. If it mentions document OCR think Textract and if it asks for custom model training think SageMaker.

A compliance software vendor is adopting Amazon Bedrock to summarize lengthy policies and auto respond to customer tickets within 90 seconds. The architects are evaluating Retrieval Augmented Generation and Bedrock Agents to decide which approach will deliver accurate, context aware answers and task automation. What best describes how RAG differs from an Agent in Amazon Bedrock?

  • ✓ C. RAG focuses on retrieving and injecting relevant external context into the model, while an Agent coordinates multi step reasoning and tool or API calls using a foundation model

RAG focuses on retrieving and injecting relevant external context into the model, while an Agent coordinates multi step reasoning and tool or API calls using a foundation model is the correct option. This choice correctly separates the roles so RAG grounds model outputs with retrieved data and Agents orchestrate planning and tool invocation to complete tasks.

RAG is about retrieving relevant documents or embeddings from knowledge stores and then injecting that context into prompts so the foundation model produces accurate, context aware answers. This makes RAG the right approach when the priority is grounding responses in proprietary or up to date data for compliance summaries and ticket responses.

Agent refers to a system that uses a foundation model to plan multi step reasoning and to call external tools or APIs as part of a workflow. Agents are the better fit when the scenario requires orchestration such as invoking databases, ticketing systems, or other services to automate tasks within a time bound SLA.

Both RAG and Agents are primarily about fetching context from knowledge stores to improve a prompt response is incorrect because it treats Agents as only retrieval mechanisms. Agents do more than fetch context and they perform planning and tool use to complete actions.

Agent means retrieving facts from a datastore to enrich output, while RAG is a workflow that plans steps and invokes tools is incorrect because it reverses the definitions of the two approaches. RAG is the retrieval and grounding pattern and Agents are the orchestration and tool calling pattern.

Both RAG and Agents are planning oriented systems that iteratively interpret inputs and act without external retrieval is incorrect because RAG explicitly relies on external retrieval to ground answers and Agents often incorporate retrieval but are defined by their ability to plan and use tools rather than by excluding retrieval.

For exam scenarios pick RAG when the requirement emphasizes grounding answers in your documents and pick Agent when the requirement emphasizes multi step workflows or calling external tools.

Which capability provides per-output rationales explaining why a generative summary omitted specific text?

  • ✓ B. Output-level explainability

Output-level explainability is correct because it provides per-output, human readable rationales that explain why a generative summary included or omitted specific text in a given response.

This capability attaches explanations directly to each generated response so auditors and reviewers can inspect the rationale behind individual outputs. It surfaces reasons tied to the model decision for a single summary and it helps trace which source content influenced inclusion or omission on a per-response basis.

Amazon SageMaker Clarify is not correct because it focuses on dataset and model level bias detection and feature attributions for structured prediction tasks and it does not produce per-output rationales for generative summaries.

Amazon Bedrock Guardrails is not correct because it enforces safety and policy constraints and it steers or filters outputs rather than explaining why a particular response left out specific content.

Amazon SageMaker Model Cards is not correct because it documents model metadata risk and intended use and it does not generate explanations tied to individual outputs.

When a question asks why a model included or omitted content look for words like rationale and per-output or explainability and pick features that provide explanations for each response rather than documentation or safety controls.

A startup in online education named BrightCourse plans to migrate growing workloads to AWS and is evaluating service models to understand how much control they retain. The team plans to use Amazon EC2 for scalable compute and wants clarity on the category this service fits into so they can plan how to manage operating systems, networking, and storage. Which cloud service model does Amazon EC2 represent?

  • ✓ C. Infrastructure as a Service (IaaS)

The correct choice is Infrastructure as a Service (IaaS). Amazon EC2 provides virtual servers where BrightCourse controls the operating system patching networking configuration attached storage and application stack which aligns with the IaaS model.

With EC2 you provision instances and take responsibility for operating system maintenance security updates and network settings while AWS manages the physical hardware and virtualization layer. This arrangement lets the startup customize OS images install required software and attach block storage for scalable compute needs.

Software as a Service (SaaS) is incorrect because it delivers fully managed applications and does not provide server or OS level control that EC2 offers.

Platform as a Service (PaaS) is incorrect because it abstracts the operating system and runtime so you only deploy code without managing instances and EC2 requires instance management.

Network as a Service (NaaS) is incorrect because it refers to managed networking capabilities and does not describe virtual machine provisioning which is the function of EC2.

When you manage the operating system and virtual machines think IaaS and when you only deploy code without server management think PaaS.

How can Amazon Bedrock adapt a foundation model to understand domain-specific jargon and acronyms? (Choose 2)

  • ✓ B. Managed fine-tuning on labeled domain data with AWS handling infrastructure

  • ✓ D. Prompt engineering with few-shot examples without altering weights

Managed fine-tuning on labeled domain data with AWS handling infrastructure and Prompt engineering with few-shot examples without altering weights are the correct options for adapting a foundation model in Amazon Bedrock to domain specific jargon and acronyms.

Managed fine-tuning on labeled domain data with AWS handling infrastructure updates model parameters using labeled examples so the model internalizes abbreviations and specialized terminology and applies them consistently. Managed fine tuning is the right choice when you can provide curated labeled data and you need the model behavior to change at the parameter level.

Prompt engineering with few-shot examples without altering weights steers model outputs by showing examples or structured prompts at inference time and does not change the underlying weights. Prompting is appropriate when you need quick iteration or when labeled fine tuning data is not available.

Train a new foundation model from scratch using raw documents is incorrect because Amazon Bedrock exposes pretrained foundation models and it does not provide a workflow to train a new foundation model from scratch within the service.

Guardrails for Amazon Bedrock change model weights to enforce safety is incorrect because guardrails operate at inference time to constrain or filter outputs and they do not modify model parameters.

Knowledge Bases for Amazon Bedrock retrain the model when documents are ingested is incorrect because Knowledge Bases provide retrieval augmented context at query time and they do not retrain the foundation model when new documents are added.

Focus on whether an option mentions updating weights or steering outputs without changing parameters when you decide between fine tuning and prompt engineering.

EcoVision Labs manages a library of over 60,000 trail-camera images and wants to automatically find and label which animals appear in each photo without human review. Which approach should they use?

  • ✓ C. Object detection

The correct choice is Object detection. It both finds where animals appear in each photo and assigns labels to each instance so EcoVision Labs can automatically tag and organize the large trail-camera image library without human review.

Object detection models analyze image regions and return category labels along with bounding box coordinates for every detected instance which supports counting animals and creating searchable metadata for the dataset.

Named entity recognition works on text to extract people and place names and other entities so it cannot locate or label animals in images.

Image inpainting restores or fills missing or corrupted pixels so it is used for image repair rather than detecting or labeling objects.

Anomaly detection highlights unusual or outlier samples which can flag rare images but it does not provide per-object labels or locations needed to identify animals in photos.

Match the data modality to the technique and choose object detection for images when you need per-instance labels and locations.

Which scenarios are best suited for Retrieval-Augmented Generation with Amazon Bedrock? (Choose 2)

  • ✓ B. Domain-specific Q&A assistant

  • ✓ D. Support Q&A assistant

The correct options are Domain-specific Q&A assistant and Support Q&A assistant. These two choices best match Retrieval-Augmented Generation because RAG returns relevant documents and uses them to ground model responses so answers remain accurate and traceable.

RAG workflows with Amazon Bedrock typically use a knowledge base and vector search to find enterprise documents and then supply that context to a foundation model during prompting. This grounding reduces hallucinations and helps the model cite or mirror authoritative content which fits both domain specific Q and A assistants and support Q and A assistants.

Amazon Personalize is a managed recommendations service for personalization and it does not provide document retrieval or grounding for text generation so it is not a RAG use case.

Fine-tuning a foundation model involves changing model weights for customization and it does not by itself provide live retrieval of evolving documents or the on demand grounding that RAG provides.

Image generation from prompts focuses on creating visual content and typically does not rely on retrieving text documents to ground responses so it is not suited to RAG.

Look for scenarios that require grounded answers from company documents and mention of vector search or knowledge base to identify RAG use cases.

A research publisher uses a generative AI service to create 120-word abstracts from 30-page whitepapers, and editors compare these outputs to human-written reference summaries to ensure key information is retained. Which evaluation metric best measures the quality of these summaries against the references?

  • ✓ B. ROUGE score

ROUGE score is the correct metric for this task because it measures overlap between the generated 120-word abstracts and human reference summaries and it emphasizes recall so it captures how much key content from the 30-page whitepapers is preserved.

ROUGE score compares n-grams sequences and longest common subsequences between the candidate and reference summaries and it offers recall-oriented variants that align with evaluating coverage and relevance in summarization tasks. Using ROUGE lets editors quantify how much of the reference content is retained rather than focusing on exact phrasing alone.

BLEU score focuses on precision of n-gram matches and it was designed for machine translation, which makes it less suitable when the priority is ensuring coverage of reference content rather than exact wording.

F1 score applies to classification tasks with discrete labels and it does not measure textual overlap or content recall, so it cannot assess the quality of generative summaries in terms of preserved information.

ROC AUC evaluates a classifier’s separability across thresholds and it is concerned with ranking predictions for binary or multi-class tasks, which makes it inappropriate for comparing free text summaries for content coverage.

For summarization evaluations prioritize recall and content coverage and choose ROUGE to measure reference overlap. For translation tasks focus on precision and consider BLEU.

Which AWS service efficiently cleans, normalizes, and applies repeatable transformations to a 30 TB dataset in Amazon S3 for ML training?

  • ✓ B. Amazon SageMaker Data Wrangler

Amazon SageMaker Data Wrangler is the correct option for efficiently cleaning normalizing and applying repeatable transformations to a 30 TB dataset stored in Amazon S3 for machine learning training.

Data Wrangler provides visual and programmatic transforms and built in feature normalization so you can build consistent preprocessing steps. It lets you export transformed workflows to processing jobs and SageMaker pipelines so the same cleaning and feature engineering can run at scale directly against S3 data. These capabilities make it suitable for large datasets and repeatable ML pipelines.

Amazon SageMaker Model Monitor is incorrect because it focuses on monitoring model and data quality after deployment rather than preparing training data.

AWS Glue DataBrew is incorrect because it targets general analytics data preparation and does not provide the same tight SageMaker training and MLOps integration for exporting preprocessing into training pipelines.

Amazon SageMaker Ground Truth is incorrect because it is a data labeling service used to create annotated datasets and not a tool for cleaning or transforming features.

Look for keywords like clean normalize and repeatable transformations to choose the data preparation tool and look for monitor or drift to point to a monitoring service.

A regional fintech startup called LumaPay has rolled out a generative AI assistant to answer questions about balances, loan offerings, and recent transactions. Users have begun attempting prompt injection by asking the assistant to disregard its internal rules and to produce unreliable financial guidance. The engineering team wants the most effective control to ensure the assistant adheres to system instructions and delivers safe, trustworthy outputs. What should the team implement?

  • ✓ C. Apply strict validation and sanitization to user prompts before the model processes them

The correct choice is Apply strict validation and sanitization to user prompts before the model processes them. This control most directly prevents user instructions from overriding system prompts and reduces the chance the assistant will produce unsafe or unreliable financial guidance.

Validating and sanitizing inputs removes manipulative patterns and normalizes prompts into an allowed format. It also enables enforcement of policy constraints and templated inputs at the model boundary so the model only receives approved instructions. Implementing these checks is an operational security control that can be audited and updated as attack patterns evolve.

Use a larger foundation model with more parameters can improve fluency and capability but it does not provide a security control and larger models remain vulnerable to carefully crafted injection prompts.

Enable conversation memory to retain the full chat context may increase coherence but it can preserve malicious instructions across turns and amplify prompt injection risks rather than neutralize them.

AWS WAF protects web traffic and common HTTP layer exploits but it does not parse or sanitize natural language prompts inside chat messages so it is not a specific defense against prompt injection.

On prompt injection questions look for answers that constrain model inputs. Favor input validation and templating at the model boundary rather than larger models or more context.

Which AWS service provides managed human review workflows to add about 30 reviewers to ML predictions without building custom applications?

  • ✓ C. Amazon Augmented AI (A2I)

The correct option is Amazon Augmented AI (A2I).

Amazon Augmented AI (A2I) provides a fully managed human-in-the-loop workflow that routes model outputs to reviewers and supports private or public workforces. It lets you onboard internal and external reviewers quickly without building custom applications and it integrates with services such as Amazon SageMaker, Amazon Textract, and Amazon Comprehend while providing auditing and quality controls for reviews.

Amazon Forecast is incorrect because it focuses on time series forecasting and does not offer managed human review workflows for model outputs.

Amazon Mechanical Turk is incorrect because it is a crowdsourcing marketplace that requires custom orchestration and does not provide the turnkey human review pipelines that Amazon Augmented AI (A2I) delivers.

Amazon SageMaker Ground Truth is incorrect because it is intended for labeling training data and not for operational human review of live predictions.

When a question asks about human review of model predictions look for Amazon Augmented AI as the managed service and remember that Ground Truth is focused on labeling training data.

A small game studio using Amazon Bedrock wants to create concept art with generative AI. They heard diffusion models work well but are unclear about the fundamental mechanism. Which statement best describes how diffusion models generate images?

  • ✓ C. Diffusion models synthesize images by starting from pure noise and repeatedly denoising with a learned reverse process

The correct answer is Diffusion models synthesize images by starting from pure noise and repeatedly denoising with a learned reverse process.

Diffusion models learn a forward process that gradually corrupts training images with noise and a learned reverse process that removes noise step by step so new images can be sampled by starting from random noise and iteratively denoising. The reverse model is trained to predict either the less noisy image or the noise itself and that training objective lets the model generate diverse and high fidelity concept art from scratch.

Diffusion models search the training corpus for the closest visual matches using vector similarity and return them is incorrect because that describes retrieval where examples are fetched instead of being synthesized from a learned probability distribution.

Diffusion models refine pictures by applying supervised filters that accentuate boundaries between pixels is incorrect because that describes enhancement of existing images rather than the generative denoising process that creates new images from noise.

Diffusion models build images primarily through an encoder-decoder with attention layers is misleading because some diffusion architectures incorporate encoder-decoder blocks and attention but the defining mechanism is the iterative denoising reverse process rather than any single encoder-decoder design.

On the exam look for phrases like start from noise or iterative denoising and treat options about retrieval or simple sharpening as distractors.

Which approach enforces least privilege so only approved users can invoke selected model endpoints and read specific Amazon S3 datasets?

  • ✓ B. IAM roles with least-privilege, resource-scoped permissions to specific endpoints and S3 data

IAM roles with least-privilege, resource-scoped permissions to specific endpoints and S3 data is correct because it limits who can invoke particular model endpoints and who can read targeted S3 objects so only approved users have access.

IAM roles let you grant precise actions such as sagemaker:InvokeEndpoint on specific endpoint ARNs and s3:GetObject on selected buckets or prefixes. They attach permissions to identities and enforce least privilege while supporting separation of duties, and you can combine IAM roles with resource policies where the service supports them to provide layered controls.

Only resource-based policies is incorrect because not all services support resource policies and relying only on them omits the identity scoping that IAM roles provide. Resource policies can complement identity policies but they often cannot stand alone for fine grained access control.

One shared IAM policy for all users and resources is incorrect because a shared policy centralizes broad permissions and fails to restrict access to specific endpoints or datasets. Such policies usually violate least privilege and increase the blast radius of any compromised identity.

Use AWS Organizations SCPs as the only control is incorrect because service control policies only set account level guardrails and they do not grant permissions. SCPs should be used alongside identity and resource policies and not as the sole mechanism for granting or enforcing least privilege.

On the exam look for phrases like least privilege and resource scoped. Prefer identity based IAM roles scoped to exact ARNs and actions when a question requires only approved users to invoke endpoints or read S3 data.

An edtech startup, LumenLearn, is mapping Amazon SageMaker tools to common ML tasks for a new project. Match each service to the task it best fits: A) SageMaker Data Wrangler B) SageMaker Canvas C) SageMaker Ground Truth 1) Aggregates human input for labeling, review, and evaluation to enhance model relevance across the ML lifecycle 2) Provides hundreds of built-in transformations to prepare data sets for machine learning 3) A no-code, point-and-click interface for creating predictions from data. Which mapping is correct?

  • ✓ C. A-2, B-3, C-1

A-2, B-3, C-1 is correct. A-2, B-3, C-1 assigns SageMaker Data Wrangler to data preparation and built-in transforms, SageMaker Canvas to a no-code point-and-click interface for producing predictions, and SageMaker Ground Truth to human-in-the-loop labeling and review that improves model quality across the ML lifecycle.

SageMaker Data Wrangler provides a large catalog of prebuilt transformations and a visual flow for cleaning and preparing datasets so you can export features for training and analysis. Using Data Wrangler reduces manual preprocessing work and helps maintain consistent transforms across experiments.

SageMaker Canvas gives business users and analysts a graphical way to build datasets and generate predictions without writing code. Canvas connects to data sources and models so nondevelopers can obtain insights quickly and iterate on scenarios.

SageMaker Ground Truth manages labeling workflows, aggregates human input, and supports review and auditing to produce high quality labeled data for supervised learning and for model evaluation.

A-3, B-2, C-1 is incorrect because it swaps the roles of Data Wrangler and Canvas and assigns data preparation to Canvas instead of to Data Wrangler.

A-2, B-1, C-3 is incorrect because it assigns human-in-the-loop responsibilities to Canvas and treats Ground Truth as a no-code prediction interface, which does not reflect their intended functions.

A-1, B-3, C-2 is incorrect because it maps Data Wrangler to human labeling and Ground Truth to data transforms, and both mappings contradict the services core purposes.

Remember that Data Wrangler pairs with data transforms, Canvas pairs with no-code predictions, and Ground Truth pairs with human labeling for model training and evaluation.

What is the term for converting words or concepts into dense numeric vectors that encode semantic meaning?

  • ✓ C. Text embeddings

Text embeddings is the correct option because embeddings are dense numeric vectors that encode the semantic meaning and relationships between words and concepts.

Text embeddings capture context beyond surface level tokens and enable tasks such as similarity search, clustering, and retrieval augmented generation. They map words and concepts into vector space so systems can measure meaning and relatedness by distance and direction in that space.

Tokens is incorrect because tokens are the basic units of text that models process and they do not by themselves represent semantic information as dense numeric vectors.

Amazon Bedrock is incorrect because it is a managed service for building generative AI applications and not the name of the vector representation that encodes semantics.

Positional encodings is incorrect because positional encodings supply order information to models and they do not represent semantic meaning as dense vectors in the same way embeddings do.

When a question mentions dense numeric vectors or similarity search choose embeddings and avoid options that name services or basic tokenization concepts.

RiverNorth Mobile plans to release a virtual support assistant that resolves device and connectivity issues without handing off to human agents. The team has selected a foundation model on AWS and must ensure every reply reflects the company’s brand voice and writing style. What is the most effective way to accomplish this?

  • ✓ C. Use prompt engineering with clear style guidelines and iterative refinement

Use prompt engineering with clear style guidelines and iterative refinement is the correct option because carefully crafted system instructions and example responses can steer a foundation model to produce consistent replies that reflect RiverNorth Mobile’s brand voice and writing style without retraining the model.

This approach works by embedding explicit tone rules and representative few shot exemplars into prompts and by iterating on those prompts based on sample outputs and edge cases. It lets the team refine phrasing, required terminology, and response structure so the virtual assistant reliably follows brand guidelines while preserving the underlying foundation model.

Increase the temperature parameter to produce more varied wording is incorrect because raising temperature increases randomness and variability, which makes it harder to guarantee consistent brand voice.

Configure a low maximum token output to constrain responses is incorrect since limiting tokens only restricts length and does not enforce tone or stylistic rules.

Perform batch inference to generate comprehensive responses is incorrect because batch inference affects throughput and cost rather than controlling style or wording of individual replies.

When questions focus on tone or brand consistency prefer prompt engineering with clear examples and iterative testing. Validate with sample prompts and remember that temperature and max tokens control variability and length and not adherence to style.

For evaluating LLM-generated summaries to ensure key information coverage while keeping them brief, which metric should be prioritized?

  • ✓ C. ROUGE-L (recall-focused)

The correct choice is ROUGE-L (recall-focused). This metric should be prioritized when evaluating LLM generated summaries to ensure key information from the source is covered while keeping the summary brief.

ROUGE-L quantifies overlap with trusted reference summaries by focusing on longest common subsequences and recall. Emphasizing recall aligns with the objective of coverage because it rewards summaries that capture essential points from the source even when wording differs.

BERTScore uses embeddings to measure semantic similarity and it can show relatedness and paraphrase quality. It does not directly target recall oriented coverage so it is less suitable when the priority is verifying that key facts are present.

BLEU is precision oriented and it was developed for machine translation. It rewards exact n gram matches and it can penalize correct paraphrases so it is not ideal for measuring coverage in summarization.

Compression ratio captures brevity by measuring how much shorter the summary is relative to the source. It does not indicate whether essential information was retained so it should not be prioritized when coverage is the evaluation objective.

When a question emphasizes coverage or key information pick a recall oriented metric such as ROUGE L. Use brevity metrics only as a separate check for length.

A regional wealth management firm, Orion Ledger, uses a generative AI system to produce client portfolio briefings every 30 days. Clients have recently reported that some write-ups include incorrect return percentages and misleading totals. The firm wants to raise confidence in the AI-generated reports. Which responsible AI focus best aligns with this goal?

  • ✓ C. Ensures generated content is factually correct and grounded in trustworthy sources

Ensures generated content is factually correct and grounded in trustworthy sources is the correct choice because the clients reported incorrect return percentages and misleading totals and the firm needs factual accuracy to restore confidence in the AI reports.

This focus requires grounding model outputs to authoritative data sources and adding verification checks and human review. Ensures generated content is factually correct and grounded in trustworthy sources helps prevent hallucinations and numeric errors by tying generated statements to validated records and implementing automated reconciliation with source data.

Promotes fairness by reducing bias in data and outcomes is important for equitable treatment and bias mitigation but it does not address the immediate problem of incorrect numbers or misleading totals.

Improves interpretability so model reasoning is easier to understand increases transparency and helps users follow how conclusions were reached but greater interpretability alone will not correct factual inaccuracies in generated figures.

AWS Key Management Service is a cloud security service for encryption key management and it is not a responsible AI focus for verifying the factual correctness of model outputs.

When the scenario highlights incorrect facts or numbers choose the option mentioning accuracy or grounding because those terms target factual integrity and verification.

Darcy Declute

Darcy DeClute is a Certified Cloud Practitioner and author of the Scrum Master Certification Guide. Popular both on Udemy and social media, Darcy’s @Scrumtuous account has well over 250K followers on Twitter/X.


Git, GitHub & GitHub Copilot Certification Made Easy

Want to get certified on the most popular AI, ML & DevOps technologies of the day? These five resources will help you get GitHub certified in a hurry.

Get certified in the latest AI, ML and DevOps technologies. Advance your career today.