AWS AI Practitioner Exam Questions

AWS AI Practitioner Practice Exam Questions

If you’re looking for AWS AI Practitioner practice exams and real AWS AI Practitioner exam questions, you’ve come to the right place.

Well, when I say real, I don’t mean real. These aren’t AI Practitioner exam dumps or AWS braindumps. These are expertly sourced AWS exam questions (I’m the expert) that come from my AI Practitioner Udemy course and my certificationexams.pro website which has helped thousands of people pass tough AWS certification exams like this one.

Study hard, do lots of AWS practice tests and you’ll be sure to get your AWS AI Practitioner certification on your first try.

A digital media startup, BrightWave Studios, is evaluating Amazon Q Developer to speed up coding, automate repetitive work, and streamline development workflows. The team wants to know where they can use Amazon Q Developer during daily development so they can decide whether it fits their toolchain. What should you tell them about its availability across developer tools and AWS interfaces?

  • ❏ A. Amazon Q Developer is limited to desktop IDEs only

  • ❏ B. Amazon Q Developer is available in IDEs and in the AWS Management Console

  • ❏ C. Amazon Q Developer is offered solely through AWS Chatbot integrations like Slack and Amazon Chime

  • ❏ D. Amazon Q Developer can be used only inside the AWS Management Console

An online marketplace named NovaMart has collected about 120 TB of unlabeled clickstream and purchase history from its website and mobile app. The marketing team wants to group shoppers into meaningful tiers to target promotions and loyalty rewards. Which approach should the team choose?

  • ❏ A. Reinforcement learning

  • ❏ B. Unsupervised learning

  • ❏ C. Supervised learning

  • ❏ D. Amazon SageMaker Ground Truth

PulseWave, a music streaming startup, wants to refine its playlist ranking using reinforcement learning from human feedback so results better match subjective listener tastes. What should the team do first to prepare the reward model before the reinforcement training begins?

  • ❏ A. Analyze historical clickstream and rating logs with statistical methods

  • ❏ B. Amazon Personalize

  • ❏ C. Collect side-by-side human rankings of alternative model outputs to fit a reward model

  • ❏ D. Generate synthetic reward scores for unlabeled examples using a pre-trained model

A language learning app uses a large language model to craft short hint messages during quizzes. The team wants the hints to be concise yet engaging so players do not get overwhelmed. Which parameter change would best enforce shorter responses?

  • ❏ A. Decrease the temperature setting

  • ❏ B. Choose a smaller model size

  • ❏ C. Set a lower max tokens limit

  • ❏ D. Expand the context window

A digital learning startup called NovaScholars uses a foundation model to power an AI study assistant that answers questions in algebra, biology, and world history. Leadership wants to ensure the assistant avoids offensive or biased language and prevents clearly incorrect claims so student trust and organizational ethics are protected. What should the team implement to enforce safe and responsible outputs?

  • ❏ A. Use prompt engineering to encourage safer replies

  • ❏ B. Amazon SageMaker Clarify

  • ❏ C. Implement guardrails and content filtering for moderated generation

  • ❏ D. Rely on output logging and retrospective analysis after release

An insurance startup needs to audit how its Amazon VPC security group rules have changed over time across multiple AWS accounts and Regions. Which AWS service provides a detailed, searchable history of these configuration changes?

  • ❏ A. AWS Security Hub

  • ❏ B. AWS Config

  • ❏ C. AWS CloudTrail

  • ❏ D. AWS Audit Manager

BlueNova Analytics plans to train a proprietary large language model using only its internal datasets and wants to reduce the carbon footprint of the training process. Which Amazon EC2 instance family should the team choose?

  • ❏ A. Amazon EC2 G series

  • ❏ B. Amazon EC2 Trn series

  • ❏ C. Amazon EC2 C series

  • ❏ D. Amazon EC2 P series

A drone logistics startup is developing onboard vision for quadcopters to identify landing zones, recognize delivery markers, and avoid hazards. The engineering group must distinguish computer vision from image processing so they apply the right methods to each part of the perception pipeline. Which statement best captures how computer vision differs from image processing?

  • ❏ A. Image processing by itself lets the drone interpret delivery markers without any AI recognition models

  • ❏ B. Computer vision only tweaks pixels and does not use machine learning

  • ❏ C. Computer vision targets semantic understanding such as object or scene recognition, while image processing emphasizes pixel-level operations like filtering or edge detection

  • ❏ D. Computer vision and image processing are equivalent and interchangeable in drone scenarios

A streaming service plans to use Amazon Bedrock to craft tailored welcome messages that reflect each member’s recent behavior, such as the last two shows watched or a recent plan change. Which prompt engineering approach will most directly improve the personalization of these messages?

  • ❏ A. Expanding the response token limit

  • ❏ B. Supplying user context and recent actions directly in the prompt

  • ❏ C. Switching to a compact model variant

  • ❏ D. Fine-tuning the base model

Caldera Insurance needs an AI-driven enterprise search that can index internal knowledge bases, third-party wikis, document repositories, and employee FAQs so staff can quickly retrieve precise answers across sources such as SharePoint Online, Confluence, and Amazon S3. Which AWS service should they choose to deliver this centralized intelligent search experience?

  • ❏ A. Amazon Comprehend

  • ❏ B. Amazon SageMaker Data Wrangler

  • ❏ C. Amazon Kendra

  • ❏ D. Amazon Textract

BrightByte Devices, a mid-sized electronics retailer, needs to forecast weekly demand for its gaming accessories across 18 locations. The team must blend internal sales history with third-party market data but has minimal experience with ML and coding. They want a simple, no-code way to build and compare predictive models. Which solution should they use?

  • ❏ A. Store data in Amazon S3 and train a forecasting model with Amazon SageMaker built-in algorithms

  • ❏ B. Prepare datasets in Amazon SageMaker Data Wrangler and build models using SageMaker built-in algorithms

  • ❏ C. Import the data into Amazon SageMaker Canvas and create a demand prediction model using its point-and-click workflow

  • ❏ D. Import the data into Amazon SageMaker Data Wrangler and try to generate demand predictions with the Amazon Personalize Trending-Now recipe

An engineering group at DeltaNova Labs is building a text summarization feature with foundation models on Amazon Bedrock and is reviewing the required preprocessing steps. In this context, what does tokenization mainly do?

  • ❏ A. Tokenization calculates probabilities to pick the next token during text generation

  • ❏ B. Tokenization segments input text into manageable units such as words or subwords so models can learn and infer

  • ❏ C. Tokenization masks sensitive attributes by replacing them with nonreversible stand-ins

  • ❏ D. AWS Key Management Service

An online learning platform needs its chatbot to answer organization-specific questions by adapting a general foundation model using proprietary support tickets and FAQ articles. What is the correct sequence of steps to complete this fine-tuning workflow from start to finish?

  • ❏ A. Clean and prepare the training dataset Choose a base foundation model Fine-tune the model on the curated data Validate the tuned model’s performance Release the tuned model to production

  • ❏ B. Choose a base foundation model Fine-tune the model on the curated data Validate the tuned model’s performance Clean and prepare the training dataset Release the tuned model to production

  • ❏ C. Choose a base foundation model Clean and prepare the training dataset Fine-tune the model on the curated data Validate the tuned model’s performance Release the tuned model to production

  • ❏ D. Choose a base foundation model Clean and prepare the training dataset Validate the tuned model’s performance Fine-tune the model on the curated data Release the tuned model to production

An online payments provider is building a fraud-detection ML pipeline on AWS that handles highly sensitive customer records. The security team must restrict who can use the data and also ensure the information is not tampered with as it passes through feature engineering, training, and inference. To select the right controls, how should they differentiate between data access control and data integrity?

  • ❏ A. Data access control guarantees data accuracy and consistency, and data integrity decides who can read or change the data

  • ❏ B. Data access control means enabling encryption with AWS KMS, while data integrity is primarily achieved through AWS CloudTrail logging

  • ❏ C. Data access control governs authentication and authorization for identities, while data integrity ensures information remains accurate, consistent, and untampered

  • ❏ D. Both data access control and data integrity are primarily about encrypting data in transit and at rest

A fintech analytics company is moving regulated payment records into AWS and needs a clear understanding of the AWS shared responsibility model to satisfy audit requirements. Which statements correctly describe how security responsibilities are divided between AWS and the customer? (Choose 2)

  • ❏ A. AWS operates, secures, and maintains the underlying facilities, hardware, and global network

  • ❏ B. AWS defines and enforces the customer’s internal compliance policies and control procedures

  • ❏ C. The customer configures encryption, identity and access management, and application-layer protections

  • ❏ D. AWS automatically performs application-level backups of each customer’s data without customer configuration

  • ❏ E. The customer must provide physical security for AWS data center buildings and server rooms

Git, GitHub & GitHub Copilot Certification Made Easy

Want to get certified on the most popular AI, ML & DevOps technologies of the day? These five resources will help you get GitHub certified in a hurry.

Get certified in the latest AI, ML and DevOps technologies. Advance your career today.

Skyline Outfitters is rolling out an LLM-powered virtual assistant for its customer support center to cut down the number of steps agents perform when answering inquiries, targeting a 25% reduction in agent effort. Which key performance indicator should the company track to best assess the assistant’s impact on agent efficiency?

  • ❏ A. Website session depth

  • ❏ B. Corporate sustainability goals

  • ❏ C. Average handle time (AHT)

  • ❏ D. Regulatory compliance posture

A digital publishing company is building a personalization feature on AWS and wants to move faster this quarter. The product managers need a precise explanation of how a machine learning algorithm differs from a machine learning model so they can assign work for an 8 week sprint. How should you describe the difference?

  • ❏ A. An ML algorithm is responsible for securing the ML pipeline, and an ML model performs feature engineering

  • ❏ B. An ML algorithm is a general method that learns from data to fit a function, and an ML model is the learned function with parameters after training

  • ❏ C. Amazon SageMaker

  • ❏ D. An ML algorithm is a prebuilt neural network, and an ML model is the raw dataset used for training

A regional insurance provider plans to build a generative AI solution to derive insights from policy and claims interactions. The team wants to maximize efficiency by spending less time provisioning, patching, and scaling infrastructure. How can AWS generative AI services help the team meet this objective?

  • ❏ A. By focusing the service on manual data labeling rather than model training and deployment

  • ❏ B. By using AWS Outposts to run all model workloads on-premises under company management

  • ❏ C. By offloading provisioning, patching, and auto scaling of the underlying infrastructure to managed generative AI services

  • ❏ D. By requiring the team to operate and patch its own GPU fleet for training and inference

An online education company is rolling out an AI semantic search feature using Knowledge Bases for Amazon Bedrock to support retrieval-augmented answers across its course library. The team wants a vector store with out-of-the-box integration so they can persist and query text embeddings without building custom connectors. Which vector database should they use?

  • ❏ A. Amazon Redshift

  • ❏ B. Amazon OpenSearch

  • ❏ C. Amazon DynamoDB

  • ❏ D. Amazon RDS

NovaLearn is building an AI tutor that processes sensitive student records and chat messages and retains activity logs for 90 days. Which practice would most effectively support responsible and socially acceptable behavior when handling this data?

  • ❏ A. Amazon SageMaker Model Monitor

  • ❏ B. Enforce comprehensive data privacy policies and perform recurring audits of data access and use

  • ❏ C. AWS Key Management Service

  • ❏ D. Use proprietary datasets without obtaining user consent

A digital retailer, NovaMart, is building an AI solution to automatically label items in photos that customers upload to product listings. The ML engineers are reviewing different neural network families to achieve high-accuracy image categorization at scale. Which neural network architecture is the most appropriate for this image classification use case?

  • ❏ A. Amazon Rekognition

  • ❏ B. Generative adversarial networks (GANs)

  • ❏ C. Convolutional neural networks (CNNs)

  • ❏ D. Recurrent neural networks (RNNs)

NorthBridge Ledger, a fintech firm, uses AI to analyze highly sensitive payment transactions. The company must demonstrate conformance with global security frameworks such as ISO/IEC 27001 and SOC 2 during quarterly audits. Which AWS service provides direct access to AWS compliance reports and certifications to support these reviews?

  • ❏ A. AWS Config

  • ❏ B. AWS Artifact

  • ❏ C. AWS Trusted Advisor

  • ❏ D. AWS CloudTrail

A logistics firm is developing an internal assistant to condense driver shift notes and translate safety manuals into eight languages. The engineers plan to evaluate Transformer-based models to process lengthy, unstructured text while preserving context across sentences. Which choice best characterizes the Transformer approach for this scenario?

  • ❏ A. A deterministic rules engine targeted at classifying structured records

  • ❏ B. A technique restricted to numeric time-series forecasting tasks only

  • ❏ C. Amazon Translate

  • ❏ D. A deep learning architecture that employs self-attention to model relationships among tokens in a sequence

A product team at Vega Retail has launched a text-generation model on Amazon Bedrock with safety guardrails enabled. During internal red-team exercises, testers craft carefully worded inputs that bypass the guardrails and cause the model to return responses that violate the content policy even though restrictions are in place. Which security risk best describes this behavior?

  • ❏ A. Indirect prompt injection through chained context

  • ❏ B. Model jailbreak using adversarial prompting

  • ❏ C. Training-time backdoor that exposes parameters without authorization

  • ❏ D. Inference-time leakage of memorized sensitive data

A retail startup, Aurora Threads, is deploying a customer support chatbot on Amazon Bedrock and wants every reply to match their brand’s friendly and confident voice. Which technique should the team use to reliably guide the model to use the desired tone without retraining the model?

  • ❏ A. Few-shot prompting

  • ❏ B. Temperature parameter tuning

  • ❏ C. Prompt engineering

  • ❏ D. Model fine-tuning

A mid-sized architecture firm built an internal HR help assistant on Amazon Bedrock to answer benefits and policy questions. The team plans a 60-day pilot and needs to determine whether the assistant increases employee efficiency across the company. What is the most effective way to evaluate the assistant’s impact?

  • ❏ A. BLEU score

  • ❏ B. Track workforce productivity KPIs such as average handle time, questions resolved per hour, and time saved per request

  • ❏ C. ROUGE score

  • ❏ D. Increase retrieval recall for the knowledge base

BlueSky Metrics, a media research startup, is building a knowledge hub with Amazon Bedrock to drive semantic search and summarization. They expect to index about 12 million embeddings for natural language queries and document retrieval and want to avoid managing an external vector store. If they let Knowledge Bases for Amazon Bedrock create the storage automatically, which vector database will be used by default?

  • ❏ A. Amazon DynamoDB

  • ❏ B. Amazon OpenSearch Serverless vector store

  • ❏ C. Amazon Aurora

  • ❏ D. Redis Enterprise Cloud

Meridian Pay is building a customer support assistant on AWS using Amazon SageMaker and Amazon Bedrock. The team wants to ensure the assistant acts responsibly and maintains user trust during conversations. Which guideline should the team prioritize to meet this goal?

  • ❏ A. AWS Shield Advanced

  • ❏ B. Build transparency and explainability into the assistant so users can understand how responses are generated

  • ❏ C. Use proprietary customer datasets even if consent is not explicit

  • ❏ D. Optimize solely for speed and efficiency

A regional fintech cooperative is deploying an AI-based risk scoring platform on AWS across several countries. Regulators require that personal data be stored and processed only in the customer’s home jurisdiction. What is the main advantage of implementing data residency?

  • ❏ A. Data residency guarantees that all model training datasets are automatically recorded and retained for audits within AWS

  • ❏ B. Data residency ensures customer information is stored and processed in a chosen geographic region to satisfy regulatory or contractual obligations

  • ❏ C. Data residency prevents anyone outside a specified IAM role from accessing encrypted data

  • ❏ D. Data residency enables automatic lifecycle deletion of expired Amazon S3 objects across multiple regions to reduce costs

Riverstone Outfitters uses Amazon Bedrock to draft automated replies to customer emails and live chat messages. Leadership wants every message to consistently reflect the company’s friendly and trustworthy brand voice. What change would be the most effective to achieve this?

  • ❏ A. Adjust the temperature setting

  • ❏ B. Prompt engineering with explicit tone instructions

  • ❏ C. Guardrails for Amazon Bedrock

  • ❏ D. Model fine-tuning

A fintech startup that processes mobile payments notices its classifier achieves very high precision but noticeably lower recall. They want a single evaluation metric that fairly balances both. What does the F1 score represent?

  • ❏ A. The arithmetic sum of precision and recall

  • ❏ B. The average of the true positive rate and the true negative rate

  • ❏ C. The harmonic mean of precision and recall

  • ❏ D. Amazon SageMaker

A travel booking startup named SkyWay Tours is launching a generative AI support assistant and wants to enforce fairness, transparency, and continuous safety checks for models in production. Which AWS capability best meets these responsible AI needs?

  • ❏ A. Customers are solely responsible for bias and safety monitoring without any AWS tools

  • ❏ B. AWS infrastructure has no involvement in ethical aspects of AI deployments

  • ❏ C. Use Amazon SageMaker Clarify with Amazon SageMaker Model Monitor to detect bias and continuously track model behavior

  • ❏ D. AWS guarantees that any model hosted on AWS is unbiased by default

A travel-tech startup, AeroTrail, is building an assistant on Amazon Bedrock that must carry out multi-turn reasoning over several steps, select and invoke external tools at runtime, and make HTTP calls to third-party APIs based on intermediate outcomes. Which AWS service or feature should the team choose to meet these needs?

  • ❏ A. AWS Step Functions

  • ❏ B. Amazon Bedrock Agents

  • ❏ C. Amazon SageMaker JumpStart

  • ❏ D. Amazon Augmented AI (A2I)

Aurora Media Group built a multilingual foundation model that translates everyday text well but misses terminology used in oil and gas compliance reports; the team plans to fine-tune the model in Amazon SageMaker with about 1,600 domain documents and a bilingual glossary to improve accuracy; which data preparation approach is most critical to capture this specialized vocabulary?

  • ❏ A. Amazon Translate

  • ❏ B. Randomly sample sentences from the corpus without quality checks

  • ❏ C. Curate and label domain-parallel examples using a consistent glossary

  • ❏ D. Rely solely on the base pre-trained model without any domain data curation

A sporting goods marketplace built a customer help assistant with Amazon Bedrock, and the team notices the assistant sometimes invents product specifications and claims that are not in the catalog. Which approach will best anchor responses to trusted company information?

  • ❏ A. Fine-tuning

  • ❏ B. Retrieval Augmented Generation (RAG)

  • ❏ C. Guardrails

  • ❏ D. Lower temperature

Certification Practice Exam Questions Answered

A digital media startup, BrightWave Studios, is evaluating Amazon Q Developer to speed up coding, automate repetitive work, and streamline development workflows. The team wants to know where they can use Amazon Q Developer during daily development so they can decide whether it fits their toolchain. What should you tell them about its availability across developer tools and AWS interfaces?

  • ✓ B. Amazon Q Developer is available in IDEs and in the AWS Management Console

The Amazon Q Developer is available in IDEs and in the AWS Management Console option is correct because the service is designed to assist developers both inside their coding environments and within the AWS web console.

The Amazon Q Developer is available in IDEs and in the AWS Management Console integration provides in-IDE features for code generation, refactoring, and inline guidance and it also appears in the AWS Management Console to help with cloud specific tasks and resource configuration. This dual availability lets teams use the assistant during local development and when they are working directly with AWS resources in the console.

The Amazon Q Developer is limited to desktop IDEs only option is wrong because it ignores the console experience and excludes the AWS Management Console integration that supports cloud workflows.

The Amazon Q Developer is offered solely through AWS Chatbot integrations like Slack and Amazon Chime option is wrong because Q Developer is not restricted to chat channels and its primary experiences are in IDEs and the console rather than being offered only through chatbot integrations.

The Amazon Q Developer can be used only inside the AWS Management Console option is wrong because Q Developer also integrates into desktop IDEs where developers do most of their coding and day to day workflows.

When an option uses absolute words like only or solely check whether the service spans multiple interfaces because many developer tools appear both in IDEs and in the AWS Management Console.

An online marketplace named NovaMart has collected about 120 TB of unlabeled clickstream and purchase history from its website and mobile app. The marketing team wants to group shoppers into meaningful tiers to target promotions and loyalty rewards. Which approach should the team choose?

  • ✓ B. Unsupervised learning

Unsupervised learning is the correct choice because the dataset is unlabeled and the marketing team wants to discover natural shopper groupings to define tiers for promotions and rewards.

Unsupervised learning methods identify inherent clusters and patterns without predefined labels and they can use algorithms such as k-means or hierarchical clustering to segment customers. These techniques work directly on historical clickstream and purchase records and they are appropriate when the goal is exploration and grouping rather than predicting a labeled outcome.

Reinforcement learning is intended for training agents through interactions with an environment using reward signals and it is not suitable for clustering static unlabeled transaction and clickstream data.

Supervised learning depends on labeled target outcomes and it is used when you have known classes or labels to predict so it does not apply to a purely unlabeled segmentation task.

Amazon SageMaker Ground Truth is a data labeling service that helps create labeled datasets for supervised workflows and while it can be used to label data if you decide to create supervised targets it is unnecessary when segmentation can be achieved directly with unsupervised approaches.

Match the data to the learning type and treat unlabeled datasets as candidates for unsupervised methods like clustering and dimensionality reduction. Use supervised approaches only when labels exist and use reinforcement learning only for sequential decision problems.

PulseWave, a music streaming startup, wants to refine its playlist ranking using reinforcement learning from human feedback so results better match subjective listener tastes. What should the team do first to prepare the reward model before the reinforcement training begins?

  • ✓ C. Collect side-by-side human rankings of alternative model outputs to fit a reward model

Collect side-by-side human rankings of alternative model outputs to fit a reward model is correct because reinforcement learning from human feedback requires an explicit reward model trained from human preference judgments before policy optimization begins.

Evaluators are shown alternative playlist or recommendation outputs for the same input and they indicate which output they prefer or how they rank them. Those pairwise comparisons and full rankings become the labeled targets used to fit a reward model which then scores candidate outputs during reinforcement learning. Training the reward model on human judgments aligns the optimization with subjective listener tastes rather than relying only on indirect engagement metrics.

Analyze historical clickstream and rating logs with statistical methods is insufficient because implicit engagement signals do not reliably provide the explicit pairwise labels that reward models for RLHF need. Historical logs can inform features and baselines but they do not replace deliberate human preference comparisons.

Amazon Personalize is not the right step because it is a managed recommendation service and it does not collect side by side human preference labels nor train reward models for RLHF. You can use it as a conventional recommendation solution but it does not substitute for the RLHF labeling workflow required before reinforcement training.

Generate synthetic reward scores for unlabeled examples using a pre-trained model is premature because pseudo labeling presupposes an existing trustworthy reward model or gold labels. Synthetic scores can introduce bias and they do not replace the initial human preference data that grounds the reward function.

Before starting RLHF gather pairwise human comparisons or rankings and use them to train the reward model so the policy can be optimized against true human preferences.

A language learning app uses a large language model to craft short hint messages during quizzes. The team wants the hints to be concise yet engaging so players do not get overwhelmed. Which parameter change would best enforce shorter responses?

  • ✓ C. Set a lower max tokens limit

Set a lower max tokens limit is the correct option because it directly caps how many tokens the model can produce and therefore enforces shorter hint messages so players do not get overwhelmed.

Set a lower max tokens limit works as a hard limit on output length and it applies regardless of model size or sampling settings. Using a lower max tokens value ensures the model cannot exceed the desired brevity and it is the most reliable parameter to control response length.

Decrease the temperature setting is incorrect because temperature mainly adjusts randomness and creativity and it does not guarantee shorter outputs even when set low.

Choose a smaller model size is incorrect because reducing model size may affect quality and style but it does not explicitly limit how long a response can be and it can introduce unpredictable brevity or verbosity.

Expand the context window is incorrect because increasing the context window gives the model more input to consider and it can enable longer responses rather than restrict them.

Use max tokens to enforce a strict length cap and tune temperature for creativity while keeping the max tokens low for concise hints.

A digital learning startup called NovaScholars uses a foundation model to power an AI study assistant that answers questions in algebra, biology, and world history. Leadership wants to ensure the assistant avoids offensive or biased language and prevents clearly incorrect claims so student trust and organizational ethics are protected. What should the team implement to enforce safe and responsible outputs?

  • ✓ C. Implement guardrails and content filtering for moderated generation

The correct choice is Implement guardrails and content filtering for moderated generation. This option enforces policy controls at generation time so offensive or biased language and clearly incorrect claims can be blocked or flagged before students see them which protects learner trust and supports organizational ethics.

Implement guardrails and content filtering for moderated generation works by applying runtime checks and filters to model outputs and by routing uncertain or risky responses to human review when needed. A layered approach that combines model safety settings with content classifiers and human-in-the-loop review gives a practical way to prevent harmful or misleading answers in real time.

Use prompt engineering to encourage safer replies can help nudge model behavior but it is not a reliable enforcement mechanism on its own because prompts cannot guarantee that a model will never produce unsafe or incorrect outputs.

Amazon SageMaker Clarify is useful for bias detection and explainability during training and evaluation but it does not provide runtime content moderation for generated text so it is not sufficient to block harmful replies at inference time.

Rely on output logging and retrospective analysis after release is a reactive practice that helps you learn and improve over time but it cannot prevent harmful responses from reaching users in the moment so it does not meet a requirement to enforce safety before exposure.

Guardrails and runtime content filters are the safer exam answer when the scenario asks to stop harmful outputs from reaching users in real time

An insurance startup needs to audit how its Amazon VPC security group rules have changed over time across multiple AWS accounts and Regions. Which AWS service provides a detailed, searchable history of these configuration changes?

  • ✓ B. AWS Config

The correct choice is AWS Config because it records configuration states over time for supported resources including security groups and it lets you view and query historical changes and timelines.

AWS Config captures point in time configuration snapshots and maintains a searchable change history for resources and it can aggregate data across multiple accounts and Regions so you can track how security group rules evolved. You can also correlate those Config snapshots with AWS CloudTrail events to determine who made changes and when.

AWS Security Hub centralizes security findings and compliance status from multiple services and it does not store resource by resource configuration histories that let you compare past settings.

AWS CloudTrail records API calls and shows who changed what and when and it is useful for audit trails, yet it does not preserve full configuration snapshots for easy historical comparison of resource state.

AWS Audit Manager helps collect audit evidence and map that evidence to frameworks and it does not provide versioned configuration tracking for individual resources.

When you need a searchable, point in time history of how settings changed choose AWS Config and when you need who made API calls choose AWS CloudTrail.

BlueNova Analytics plans to train a proprietary large language model using only its internal datasets and wants to reduce the carbon footprint of the training process. Which Amazon EC2 instance family should the team choose?

  • ✓ B. Amazon EC2 Trn series

Amazon EC2 Trn series is the correct choice because it uses AWS Trainium accelerators that are purpose built for large scale model training and they offer improved energy efficiency and cost to train which makes them ideal when minimizing environmental impact is a priority.

The Trn series is optimized for deep learning training workloads so Trainium provides high performance per watt compared to general purpose GPUs and CPUs and that leads to lower carbon footprint for large language model training at scale.

Amazon EC2 G series focuses on graphics and inference acceleration and it is not the optimal choice for efficient training of very large language models.

Amazon EC2 C series consists of compute optimized CPU instances for general workloads and they are not specialized for deep learning training at scale.

Amazon EC2 P series offers powerful GPU based training but compared to Trainium it is typically less energy efficient for LLM training when sustainability is emphasized.

When a question highlights training and reducing carbon footprint choose Amazon EC2 Trn series as the best answer.

A drone logistics startup is developing onboard vision for quadcopters to identify landing zones, recognize delivery markers, and avoid hazards. The engineering group must distinguish computer vision from image processing so they apply the right methods to each part of the perception pipeline. Which statement best captures how computer vision differs from image processing?

  • ✓ C. Computer vision targets semantic understanding such as object or scene recognition, while image processing emphasizes pixel-level operations like filtering or edge detection

The correct answer is Computer vision targets semantic understanding such as object or scene recognition, while image processing emphasizes pixel-level operations like filtering or edge detection. This option contrasts high level interpretation with low level pixel manipulations and it best captures the distinction the engineering team needs for landing zone detection and marker recognition.

Computer vision focuses on extracting meaning and structure from images and often relies on machine learning models to detect objects, classify scenes, or segment areas of interest. Image processing is used earlier in the pipeline to enhance images, reduce noise, or compute edges so that downstream computer vision models can work more reliably.

Image processing by itself lets the drone interpret delivery markers without any AI recognition models is incorrect because simple pixel operations can improve contrast or isolate shapes but they do not by themselves provide semantic interpretation. Recognizing symbols or text requires detection or classification models rather than only filters and transforms.

Computer vision only tweaks pixels and does not use machine learning is incorrect because modern computer vision commonly uses machine learning and deep learning to map pixels to labels, bounding boxes, or semantic masks. Pixel level tweaks are part of image processing and they support but do not replace CV models.

Computer vision and image processing are equivalent and interchangeable in drone scenarios is incorrect because the two fields operate at different levels of abstraction and they are complementary. Image processing handles low level enhancement and feature extraction and computer vision provides high level semantic understanding which is required for tasks like landing zone selection and hazard avoidance.

Focus on whether the question contrasts low level pixel work versus high level meaning and pick the answer that describes semantic understanding for recognition tasks rather than just pixel manipulation.

A streaming service plans to use Amazon Bedrock to craft tailored welcome messages that reflect each member’s recent behavior, such as the last two shows watched or a recent plan change. Which prompt engineering approach will most directly improve the personalization of these messages?

  • ✓ B. Supplying user context and recent actions directly in the prompt

Supplying user context and recent actions directly in the prompt is correct because giving the model the user name the last two shows watched and any recent plan changes lets it generate grounded personalized welcome messages without retraining.

By inserting user attributes and recent actions into a prompt the model can reference specific facts and produce timely relevant greetings. This method scales per user and is fast and cost effective because it avoids running a separate fine tuning cycle and instead uses prompt templates or runtime data injection to achieve consistent personalization.

Expanding the response token limit is wrong because increasing the token budget only allows longer outputs and does not add the user specific data needed to shape content.

Switching to a compact model variant is wrong because choosing a smaller or faster model affects cost and latency and does not inherently make outputs more personalized.

Fine-tuning the base model is wrong for this use case because fine tuning changes behavior across many users and requires retraining and maintenance. Fine tuning is useful for broad domain adaptation but it is not the most direct way to produce dynamic per user personalization in real time.

When you need per user personalization prefer adding runtime user context into prompt templates and avoid retraining for each change.

Caldera Insurance needs an AI-driven enterprise search that can index internal knowledge bases, third-party wikis, document repositories, and employee FAQs so staff can quickly retrieve precise answers across sources such as SharePoint Online, Confluence, and Amazon S3. Which AWS service should they choose to deliver this centralized intelligent search experience?

  • ✓ C. Amazon Kendra

The correct option is Amazon Kendra because it is an ML powered enterprise search service that unifies content from many repositories and returns precise answers to natural language questions across sources such as SharePoint Online, Confluence, and Amazon S3.

Amazon Kendra includes built in connectors and crawlers for common content stores so it can index internal knowledge bases, third party wikis, document repositories, and FAQ pages and present relevance ranked results and direct answers rather than only document hits. The service also supports natural language question answering and extracts and highlights snippets that directly address user queries which makes it ideal for a centralized intelligent search experience.

Amazon Comprehend is focused on text analytics tasks such as entity recognition and sentiment analysis and it does not provide a federated enterprise search product with connectors and natural language Q and A.

Amazon SageMaker Data Wrangler is designed to simplify data preparation for machine learning models and it is not used to build an end user search experience across content repositories.

Amazon Textract extracts text and structured data from scanned documents and images which can feed a search index, but it does not provide cross repository indexing, relevance ranking, and natural language Q and A as a complete enterprise search solution.

Map enterprise search scenarios that ask for cross repository indexing and natural language Q and A to Amazon Kendra. Check for built in connectors and direct answer features to distinguish it from analytics or OCR services.

BrightByte Devices, a mid-sized electronics retailer, needs to forecast weekly demand for its gaming accessories across 18 locations. The team must blend internal sales history with third-party market data but has minimal experience with ML and coding. They want a simple, no-code way to build and compare predictive models. Which solution should they use?

  • ✓ C. Import the data into Amazon SageMaker Canvas and create a demand prediction model using its point-and-click workflow

The correct choice is Import the data into Amazon SageMaker Canvas and create a demand prediction model using its point-and-click workflow. This option meets the requirement for a simple no-code method that lets business users combine internal sales history and third party market data and compare predictive models for weekly demand across multiple locations.

Amazon SageMaker Canvas is purpose built for non ML experts and it provides a visual, point and click workflow to prepare tabular datasets build and evaluate forecasting models and generate predictions without writing code. Canvas can ingest data from common sources and it produces results that analysts can review or export for downstream use.

Store data in Amazon S3 and train a forecasting model with Amazon SageMaker built-in algorithms is not ideal because that approach requires setting up training jobs and deployment with SDKs or notebooks and it does not satisfy the no code constraint.

Prepare datasets in Amazon SageMaker Data Wrangler and build models using SageMaker built-in algorithms is unsuitable because Data Wrangler focuses on data preparation and you still need code driven training to run built in algorithms which breaks the requirement for a no code workflow.

Import the data into Amazon SageMaker Data Wrangler and try to generate demand predictions with the Amazon Personalize Trending-Now recipe is incorrect because Amazon Personalize is a recommender service and the Trending Now recipe targets recommendations rather than general time series forecasting or demand prediction.

When a question emphasizes no-code model building for tabular business data choose SageMaker Canvas. Use Data Wrangler for heavy data preparation and use SDKs or notebooks for code based training and deployment.

An engineering group at DeltaNova Labs is building a text summarization feature with foundation models on Amazon Bedrock and is reviewing the required preprocessing steps. In this context, what does tokenization mainly do?

  • ✓ B. Tokenization segments input text into manageable units such as words or subwords so models can learn and infer

The correct choice is Tokenization segments input text into manageable units such as words or subwords so models can learn and infer. In generative AI workflows this preprocessing step converts raw text into discrete tokens that models can embed and process efficiently for both training and inference.

Tokenization maps characters and character sequences into token ids so embeddings and attention mechanisms can operate on fixed units. Using subword methods helps handle rare words and reduces vocabulary size which improves model efficiency for encoding and decoding.

Tokenization calculates probabilities to pick the next token during text generation is incorrect because computing next token probabilities is performed by the model during decoding and not by the preprocessing step.

Tokenization masks sensitive attributes by replacing them with nonreversible stand-ins is incorrect because that describes data masking or anonymization which is a separate data processing activity and not the typical role of NLP tokenization.

AWS Key Management Service is incorrect because KMS manages encryption keys and related cryptographic operations and it does not perform text tokenization for models.

Remember that tokenization breaks text into tokens for model input and not into probabilities or encrypted placeholders for the exam questions.

An online learning platform needs its chatbot to answer organization-specific questions by adapting a general foundation model using proprietary support tickets and FAQ articles. What is the correct sequence of steps to complete this fine-tuning workflow from start to finish?

  • ✓ C. Choose a base foundation model Clean and prepare the training dataset Fine-tune the model on the curated data Validate the tuned model’s performance Release the tuned model to production

Choose a base foundation model Clean and prepare the training dataset Fine-tune the model on the curated data Validate the tuned model’s performance Release the tuned model to production is correct. This sequence follows the standard machine learning lifecycle so you select the most appropriate base model first and then prepare data that matches that model so training can proceed effectively.

Choose a base foundation model Clean and prepare the training dataset Fine-tune the model on the curated data Validate the tuned model’s performance Release the tuned model to production is correct because choosing the base model first informs the data format and preprocessing steps and then fine tuning uses the curated dataset to adapt the model to organization specific content. Validation comes after training so you can measure performance against held out data and then deployment follows successful evaluation.

Clean and prepare the training dataset Choose a base foundation model Fine-tune the model on the curated data Validate the tuned model’s performance Release the tuned model to production is incorrect because preparing data before choosing a base model can lead to mismatches in expected input formats and feature requirements for the selected model.

Choose a base foundation model Fine-tune the model on the curated data Validate the tuned model’s performance Clean and prepare the training dataset Release the tuned model to production is incorrect because attempting to fine tune before cleaning and preparing the data is not feasible and will produce unreliable training outcomes.

Choose a base foundation model Clean and prepare the training dataset Validate the tuned model’s performance Fine-tune the model on the curated data Release the tuned model to production is incorrect because validation before any training means you have nothing trained to evaluate and the assessment becomes invalid.

Map options to the ML lifecycle and pick the one that reads as select, prepare, train, validate, and then deploy when you answer.

An online payments provider is building a fraud-detection ML pipeline on AWS that handles highly sensitive customer records. The security team must restrict who can use the data and also ensure the information is not tampered with as it passes through feature engineering, training, and inference. To select the right controls, how should they differentiate between data access control and data integrity?

  • ✓ C. Data access control governs authentication and authorization for identities, while data integrity ensures information remains accurate, consistent, and untampered

Data access control governs authentication and authorization for identities, while data integrity ensures information remains accurate, consistent, and untampered is the correct option for this scenario.

This option is correct because access control is about who can perform actions and which identities can read or modify data, and it is enforced with mechanisms such as IAM policies, roles, resource policies, and fine grained permissions. Data integrity is about keeping information accurate and proving it has not been altered, and it is achieved with mechanisms such as checksums, digital signatures, immutable or versioned storage, and tamper evidencing in pipelines during feature engineering, training, and inference.

Data access control guarantees data accuracy and consistency, and data integrity decides who can read or change the data is wrong because it reverses the core concepts and mixes up responsibilities for access and for preserving correctness.

Data access control means enabling encryption with AWS KMS, while data integrity is primarily achieved through AWS CloudTrail logging is incorrect because encryption primarily protects confidentiality and CloudTrail provides auditing and visibility rather than a technical enforcement of integrity. Integrity requires checks, signatures, versioning, or tamper proof storage and verification.

Both data access control and data integrity are primarily about encrypting data in transit and at rest is incomplete because encryption helps confidentiality and can assist integrity in limited ways, but access control and integrity require policies, authentication, authorization, and cryptographic or versioning mechanisms to detect or prevent unauthorized changes.

Map who can do what to access control and map unchanged and accurate data to integrity and remember that encryption mainly protects confidentiality.

A fintech analytics company is moving regulated payment records into AWS and needs a clear understanding of the AWS shared responsibility model to satisfy audit requirements. Which statements correctly describe how security responsibilities are divided between AWS and the customer? (Choose 2)

  • ✓ A. AWS operates, secures, and maintains the underlying facilities, hardware, and global network

  • ✓ C. The customer configures encryption, identity and access management, and application-layer protections

AWS operates, secures, and maintains the underlying facilities, hardware, and global network and The customer configures encryption, identity and access management, and application-layer protections are correct because they describe the split where AWS is responsible for the underlying infrastructure and the customer is responsible for configuring and protecting their data and applications.

AWS operates, secures, and maintains the underlying facilities, hardware, and global network covers physical data center security, server hardware lifecycle, and global network integrity. AWS manages the host infrastructure, isolation of the virtualization layer, and the foundational services so customers do not need to provide physical or low level hardware controls.

The customer configures encryption, identity and access management, and application-layer protections means the customer must manage IAM policies and roles, choose and configure encryption and key management, secure guest operating systems when applicable, and implement application level controls and monitoring required for compliance and audits.

AWS defines and enforces the customer’s internal compliance policies and control procedures is incorrect because AWS supplies compliant services and documentation but it is the customer who must define, implement, and operate their internal governance and control procedures to meet regulatory requirements.

AWS automatically performs application-level backups of each customer’s data without customer configuration is incorrect because backup responsibilities and retention settings depend on the service and the customer must configure and manage application level backups and recovery processes.

The customer must provide physical security for AWS data center buildings and server rooms is incorrect because AWS is responsible for physical security of its data centers and customers do not provide or manage those controls.

When preparing audit evidence map each control to whether it is AWS responsibility or customer responsibility and capture configuration artifacts for customer managed controls. Remember that AWS is security of the cloud and you are security in the cloud.

Skyline Outfitters is rolling out an LLM-powered virtual assistant for its customer support center to cut down the number of steps agents perform when answering inquiries, targeting a 25% reduction in agent effort. Which key performance indicator should the company track to best assess the assistant’s impact on agent efficiency?

  • ✓ C. Average handle time (AHT)

The correct KPI to track is Average handle time (AHT). This metric directly measures the amount of time agents spend per contact and will reveal whether the LLM assistant reduces agent effort toward the 25 percent reduction goal.

Average handle time (AHT) aggregates talk time, hold time, and after call work and so a successful assistant that removes steps or speeds decision making should produce a measurable decrease in this metric. Tracking AHT over time also lets Skyline Outfitters quantify improvements and correlate them with assistant rollouts and training updates.

Website session depth measures engagement on the web site and does not reflect call center workflow or time spent per contact. It will not show whether agents are performing fewer steps.

Corporate sustainability goals relate to environmental and social performance and are not useful for assessing agent efficiency or handling time. They are strategic metrics but not operational indicators of contact center workload.

Regulatory compliance posture indicates how well the organization meets laws and standards and does not measure operational speed or the number of actions agents take per interaction. Compliance may influence processes but the posture itself is not the KPI for agent effort.

Match the KPI to the stated outcome and pick operational time or contact metrics such as Average handle time when the goal is to reduce agent steps or workload.

A digital publishing company is building a personalization feature on AWS and wants to move faster this quarter. The product managers need a precise explanation of how a machine learning algorithm differs from a machine learning model so they can assign work for an 8 week sprint. How should you describe the difference?

  • ✓ B. An ML algorithm is a general method that learns from data to fit a function, and an ML model is the learned function with parameters after training

An ML algorithm is a general method that learns from data to fit a function, and an ML model is the learned function with parameters after training is the correct option because the algorithm describes the learning procedure and the model is the trained artifact you use to make predictions on new inputs.

The algorithm is the recipe that specifies how data is processed and how parameters are updated during training. The model is the resulting parameterized function that captures what the algorithm learned and that you deploy for inference and evaluation.

An ML algorithm is responsible for securing the ML pipeline, and an ML model performs feature engineering is incorrect because security controls and feature engineering are operational tasks in the broader ML workflow and they do not define the conceptual roles of algorithm or model.

Amazon SageMaker is incorrect because it is an AWS service used to build, train, and deploy models and it does not explain the conceptual difference between an algorithm and a model.

An ML algorithm is a prebuilt neural network, and an ML model is the raw dataset used for training is incorrect because an algorithm is a method or procedure and not a dataset, and a model is the output produced by training an algorithm on data rather than the input data itself.

Remember that an algorithm is the procedure and the model is the trained artifact. For sprint planning assign algorithm work to research and training tasks and assign model work to evaluation deployment and monitoring tasks.

A regional insurance provider plans to build a generative AI solution to derive insights from policy and claims interactions. The team wants to maximize efficiency by spending less time provisioning, patching, and scaling infrastructure. How can AWS generative AI services help the team meet this objective?

  • ✓ C. By offloading provisioning, patching, and auto scaling of the underlying infrastructure to managed generative AI services

The correct choice is By offloading provisioning, patching, and auto scaling of the underlying infrastructure to managed generative AI services. Services such as Amazon Bedrock and Amazon SageMaker handle the infrastructure so the team can focus on data preparation, prompt design, and model quality rather than managing servers.

Managed generative AI services provide built in provisioning, automatic scaling, monitoring, and patching for compute and model hosting. This reduces operational overhead and speeds time to value because the team does not need to maintain GPU clusters or the orchestration layers that serve models.

By focusing the service on manual data labeling rather than model training and deployment is incorrect because manual labeling does not remove the burden of provisioning and maintaining training and inference infrastructure and it does not cover hosting workflows.

By using AWS Outposts to run all model workloads on-premises under company management is incorrect because running on premises with Outposts brings hardware operations and patching into the team responsibility and usually increases administration instead of minimizing it.

By requiring the team to operate and patch its own GPU fleet for training and inference is incorrect because operating a dedicated GPU fleet adds significant provisioning, patching, and scaling work which conflicts with the goal to reduce operational effort.

When the question stresses reducing operational overhead choose managed or serverless generative AI services that automatically scale and handle maintenance so the team can focus on models and data.

An online education company is rolling out an AI semantic search feature using Knowledge Bases for Amazon Bedrock to support retrieval-augmented answers across its course library. The team wants a vector store with out-of-the-box integration so they can persist and query text embeddings without building custom connectors. Which vector database should they use?

  • ✓ B. Amazon OpenSearch

The correct choice is Amazon OpenSearch. Knowledge Bases for Amazon Bedrock provides native integration with Amazon OpenSearch including OpenSearch Serverless and managed clusters so teams can persist embeddings and run similarity searches without building custom connectors.

Amazon OpenSearch is designed to store vectors, index embeddings, and perform nearest neighbor queries that retrieval augmented generation workflows require. The native integration means the product handles storage, indexing, and similarity search for embeddings out of the box which reduces operational effort for the team.

Amazon Redshift is a data warehouse focused on analytics and reporting and it does not offer native vector similarity features or built in Knowledge Bases integration.

Amazon DynamoDB is a key value and document store and it does not provide built in vector similarity search or an out of the box connector for Knowledge Bases.

Amazon RDS can run extensions such as pgvector on PostgreSQL but it is not a natively supported vector store for Knowledge Bases and would require custom integration and additional operational work.

Native integration is the key phrase to watch for and it usually signals a managed vector capability such as Amazon OpenSearch.

NovaLearn is building an AI tutor that processes sensitive student records and chat messages and retains activity logs for 90 days. Which practice would most effectively support responsible and socially acceptable behavior when handling this data?

  • ✓ B. Enforce comprehensive data privacy policies and perform recurring audits of data access and use

The correct choice is Enforce comprehensive data privacy policies and perform recurring audits of data access and use.

Enforce comprehensive data privacy policies and perform recurring audits of data access and use sets clear expectations for consent, data minimization, and least privilege and recurring audits verify those controls are followed and help detect misuse. Regular reviews also demonstrate compliance with regulations such as GDPR and FERPA and ensure that retaining activity logs for 90 days is justified and monitored.

Amazon SageMaker Model Monitor is useful for detecting model drift and data quality issues and it does not create or enforce privacy policies or consent requirements and it cannot by itself provide acceptable use governance.

AWS Key Management Service provides encryption and key control and that protects data at rest and in transit and it does not by itself ensure ethical use, informed consent, or oversight of who may access student records.

Use proprietary datasets without obtaining user consent is inappropriate and often illegal and it violates privacy expectations and responsible AI principles and therefore it is not a valid practice for handling sensitive student data.

When an exam question focuses on socially responsible handling of sensitive data prioritize governance, consent, and audits over single technical controls and remember that monitoring and encryption support but do not replace policy and oversight.

A digital retailer, NovaMart, is building an AI solution to automatically label items in photos that customers upload to product listings. The ML engineers are reviewing different neural network families to achieve high-accuracy image categorization at scale. Which neural network architecture is the most appropriate for this image classification use case?

  • ✓ C. Convolutional neural networks (CNNs)

The correct option is Convolutional neural networks (CNNs). They are the most appropriate architecture for high accuracy image categorization at scale.

CNNs learn local spatial filters through convolution and pooling which extract edges textures and shapes across an image and build hierarchical feature representations. This weight sharing and local connectivity reduce parameter count and make training efficient and scalable for large image datasets and production inference.

Amazon Rekognition is an AWS managed vision service rather than a neural network family. It can provide ready made image analysis but it does not specify which underlying model architecture is the right choice when designing a custom classifier.

Generative adversarial networks (GANs) are designed to generate realistic images and to learn data distributions. They are not the standard supervised classification architecture and they introduce training complexity that is unnecessary for labeling existing photos.

Recurrent neural networks (RNNs) are built for sequence data like text or time series and they do not capture spatial patterns in static images as effectively as CNNs. They typically underperform CNNs on image classification tasks.

On image questions remember that CNNs are tailored for spatial patterns and that managed services like Amazon Rekognition are products not model families.

NorthBridge Ledger, a fintech firm, uses AI to analyze highly sensitive payment transactions. The company must demonstrate conformance with global security frameworks such as ISO/IEC 27001 and SOC 2 during quarterly audits. Which AWS service provides direct access to AWS compliance reports and certifications to support these reviews?

  • ✓ B. AWS Artifact

The correct choice is AWS Artifact. It provides on demand access to AWS’s audit artifacts including ISO/IEC 27001 and SOC 2 reports which customers can download and present to regulators and auditors.

AWS Artifact is a self service portal that lets customers retrieve official compliance reports third party audit attestations and select AWS agreements that support audit needs. It supplies the formal evidence auditors expect when reviewing conformity to global security frameworks and it is designed for use during periodic compliance reviews.

AWS Config tracks and evaluates resource configurations against rules to help with internal compliance and drift detection but it does not furnish AWS issued certifications or downloadable third party audit reports.

AWS Trusted Advisor provides best practice checks and operational recommendations for cost performance fault tolerance and basic security guidance but it does not supply formal compliance certifications or audit artifacts.

AWS CloudTrail records API activity and supports forensic analysis governance and operational troubleshooting and it can provide evidence of actions but it does not issue certification reports or host third party audit attestations.

When a question mentions access to official audit reports or certifications think AWS Artifact and not configuration or logging services.

A logistics firm is developing an internal assistant to condense driver shift notes and translate safety manuals into eight languages. The engineers plan to evaluate Transformer-based models to process lengthy, unstructured text while preserving context across sentences. Which choice best characterizes the Transformer approach for this scenario?

  • ✓ D. A deep learning architecture that employs self-attention to model relationships among tokens in a sequence

The correct choice is A deep learning architecture that employs self-attention to model relationships among tokens in a sequence. Transformers use self-attention to capture long range dependencies and preserve context across sentences and larger documents which makes them well suited for summarization and translation of lengthy unstructured text.

A deep learning architecture that employs self-attention to model relationships among tokens in a sequence computes attention weights that let the model focus on relevant tokens regardless of their position and this enables handling of long sequences more effectively than many older sequence models. That property supports condensing driver shift notes while retaining important context and it also supports multilingual translation when combined with appropriate training or pre trained language models.

A deterministic rules engine targeted at classifying structured records is not a learned neural approach and it cannot robustly handle the variability and nuanced context present in natural language notes.

A technique restricted to numeric time-series forecasting tasks only is incorrect because Transformers generalize to many sequence modalities and they are widely applied to text not just to numeric data.

Amazon Translate is an AWS managed machine translation service and it provides translation functionality but it is not a description of the Transformer architecture and the question asks about the model concept not a specific managed service.

Look for self-attention and long-range context to identify answers that describe Transformers and avoid options that name a service or a rules based approach.

A product team at Vega Retail has launched a text-generation model on Amazon Bedrock with safety guardrails enabled. During internal red-team exercises, testers craft carefully worded inputs that bypass the guardrails and cause the model to return responses that violate the content policy even though restrictions are in place. Which security risk best describes this behavior?

  • ✓ B. Model jailbreak using adversarial prompting

The correct risk is Model jailbreak using adversarial prompting. The testers are sending carefully worded inputs that intentionally bypass the guardrails so the model returns restricted content and that behavior matches a jailbreak attack.

A model jailbreak uses adversarially crafted prompts at inference to cause a model to ignore safety constraints and produce forbidden outputs. In this case the issue arises from the input text itself and from the model following adversarial instructions rather than from training data or from chaining external content, so the behavior aligns with a jailbreak attack.

Indirect prompt injection through chained context concerns untrusted external data or tool outputs influencing a model across multiple steps. This scenario describes a direct user prompt against the deployed model so it does not match chained context injection.

Training-time backdoor that exposes parameters without authorization involves malicious data or triggers that were inserted during training or fine tuning so a model responds to specific triggers later. The exercise here uses crafted inference prompts and not a planted training backdoor.

Inference-time leakage of memorized sensitive data would mean the model is revealing private training examples or secrets. The problem described is bypassing safety policies rather than the model disclosing memorized sensitive information.

When a crafted input overrides safety controls think jailbreak and when untrusted external content or tool outputs are involved think prompt injection.

A retail startup, Aurora Threads, is deploying a customer support chatbot on Amazon Bedrock and wants every reply to match their brand’s friendly and confident voice. Which technique should the team use to reliably guide the model to use the desired tone without retraining the model?

  • ✓ C. Prompt engineering

The correct choice is Prompt engineering. This approach lets the team instruct the model to adopt Aurora Threads’ friendly and confident brand voice for every reply without changing model weights or retraining.

By writing explicit style guidance such as a persona statement, tone instructions, and do and do not lists in the system or instruction prompt, Prompt engineering steers the model toward consistent wording and phrasing. It is quick to iterate and low cost because it only changes the input prompts sent to Amazon Bedrock rather than creating a new model artifact.

Few-shot prompting can provide examples that nudge tone but it is often less consistent across varied inputs and it consumes more tokens which increases maintenance and cost.

Temperature parameter tuning changes randomness and can make outputs more conservative or more creative but it does not ensure a specific brand voice or persona.

Model fine-tuning can permanently bake a style into model weights but it requires extra time, cost, and governance and it is unnecessary when clear prompts can reliably impose tone.

Use clear system prompts that define persona, tone, and explicit do and do not lists and test variations to ensure consistency across different user inputs.

A mid-sized architecture firm built an internal HR help assistant on Amazon Bedrock to answer benefits and policy questions. The team plans a 60-day pilot and needs to determine whether the assistant increases employee efficiency across the company. What is the most effective way to evaluate the assistant’s impact?

  • ✓ B. Track workforce productivity KPIs such as average handle time, questions resolved per hour, and time saved per request

Track workforce productivity KPIs such as average handle time, questions resolved per hour, and time saved per request is correct because these metrics directly measure employee efficiency and show whether the Amazon Bedrock HR assistant reduces time per task and increases throughput during the 60 day pilot.

Use a before and after comparison or an A/B test to collect a baseline and then measure changes in average handle time, questions resolved per hour, first contact resolution and time saved per request. Quantitative KPIs provide concrete evidence of business impact and you can complement them with qualitative feedback to capture user satisfaction and cases where the assistant failed or succeeded.

BLEU score is not appropriate because it is a text similarity metric for machine translation and it does not indicate changes in workforce productivity or time savings.

ROUGE score is likewise a summarization overlap metric and it will not show whether employees handled more requests or spent less time per question.

Increase retrieval recall for the knowledge base can improve answer coverage but it does not by itself prove increased efficiency unless you also measure outcome KPIs like handle time and resolution rate.

Choose metrics that quantify business outcomes such as time saved, resolution rate and average handle time rather than model quality scores when evaluating an assistant.

BlueSky Metrics, a media research startup, is building a knowledge hub with Amazon Bedrock to drive semantic search and summarization. They expect to index about 12 million embeddings for natural language queries and document retrieval and want to avoid managing an external vector store. If they let Knowledge Bases for Amazon Bedrock create the storage automatically, which vector database will be used by default?

  • ✓ B. Amazon OpenSearch Serverless vector store

Amazon OpenSearch Serverless vector store is the default vector database that Knowledge Bases for Amazon Bedrock provisions when you allow Bedrock to create the store automatically. It uses the OpenSearch Serverless vector engine to store embeddings and perform similarity search for semantic queries.

The OpenSearch Serverless vector store is selected because it provides a managed vector engine that integrates with Knowledge Bases for Bedrock and it scales to large embedding volumes without requiring you to operate a separate vector database. This lets Bedrock handle provisioning, indexing, and similarity search so your team does not need to manage the underlying vector infrastructure.

Amazon DynamoDB is a key value and document database and it is not offered as a native vector store for Knowledge Bases. It lacks the built in vector search engine that OpenSearch Serverless provides and therefore it is not the default.

Amazon Aurora can be configured with pgvector on PostgreSQL as a customer managed integration, but it is not automatically created by Bedrock. That makes Aurora a possible custom option but not the Bedrock managed default.

Redis Enterprise Cloud is available as an external connector for vector storage yet it is not provisioned by Bedrock when the Knowledge Base is created. It remains an external or customer supplied vector service and so it is not the default choice.

When a question emphasizes the default or that Bedrock will create the vector store for you pick Amazon OpenSearch Serverless vector store because it is the Bedrock managed option.

Meridian Pay is building a customer support assistant on AWS using Amazon SageMaker and Amazon Bedrock. The team wants to ensure the assistant acts responsibly and maintains user trust during conversations. Which guideline should the team prioritize to meet this goal?

  • ✓ B. Build transparency and explainability into the assistant so users can understand how responses are generated

The correct choice is Build transparency and explainability into the assistant so users can understand how responses are generated. This guideline helps maintain user trust and ensures the assistant behaves responsibly during conversations.

Build transparency and explainability into the assistant so users can understand how responses are generated supports accountability and helps detect and mitigate bias. Implementing explainability tools and clear user facing disclosures lets customers understand limitations and provenance of responses and enables human oversight and remediation when issues arise.

AWS Shield Advanced is focused on DDoS protection and network security so it does not address model behavior, transparency, or explainability and it is not relevant to responsible AI guidelines.

Use proprietary customer datasets even if consent is not explicit violates privacy and data governance principles and undermines trust. Responsible AI requires proper consent and careful handling of customer data.

Optimize solely for speed and efficiency prioritizes performance at the expense of fairness, accuracy, and user trust. Ethical deployment requires balancing efficiency with transparency, privacy, and human oversight.

When a question focuses on ethical or responsible AI look for answers that highlight transparency and explainability rather than unrelated security services or performance only goals.

A regional fintech cooperative is deploying an AI-based risk scoring platform on AWS across several countries. Regulators require that personal data be stored and processed only in the customer’s home jurisdiction. What is the main advantage of implementing data residency?

  • ✓ B. Data residency ensures customer information is stored and processed in a chosen geographic region to satisfy regulatory or contractual obligations

Data residency ensures customer information is stored and processed in a chosen geographic region to satisfy regulatory or contractual obligations is correct and it directly addresses the regulator requirement to keep personal data within a customer’s home jurisdiction.

Data residency is about the geographic placement and processing location of data so that data sovereignty, local law and contractual restrictions are respected. Implementing residency controls ensures workloads and storage remain in the permitted region while access control, logging and lifecycle management are handled by other AWS services and policies.

Data residency guarantees that all model training datasets are automatically recorded and retained for audits within AWS is incorrect because audit recording and retention require explicit logging and retention configurations such as AWS CloudTrail and service specific logs and are not an inherent feature of residency.

Data residency prevents anyone outside a specified IAM role from accessing encrypted data is incorrect because who can access or decrypt data is determined by IAM policies and key management settings and not by the geographic location of the data.

Data residency enables automatic lifecycle deletion of expired Amazon S3 objects across multiple regions to reduce costs is incorrect because S3 lifecycle rules and cross region replication handle retention and cost optimization and those features are separate from enforcing where data is stored.

Focus your answer on where data is stored and processed when you see data residency in a question and not on who can access it or how it is logged or deleted.

Riverstone Outfitters uses Amazon Bedrock to draft automated replies to customer emails and live chat messages. Leadership wants every message to consistently reflect the company’s friendly and trustworthy brand voice. What change would be the most effective to achieve this?

  • ✓ B. Prompt engineering with explicit tone instructions

The most effective change is Prompt engineering with explicit tone instructions. Specifying the desired voice, style, role, and short exemplars steers the model to produce consistent, friendly, and trustworthy messages across email and live chat without needing additional training.

Prompt engineering works because you can directly tell the model how to phrase greetings, how to express empathy, and which words to prefer or avoid. Clear prompts are fast to iterate and cost effective because they do not require collecting labeled data or redeploying models. You can also combine prompts with runtime settings and monitoring to refine responses in production.

Adjust the temperature setting changes how random or deterministic the model outputs are but it does not reliably enforce a specific brand voice. Temperature is useful for controlling creativity after you have specified tone in the prompt.

Guardrails for Amazon Bedrock are intended for safety, policy enforcement, and content filtering and they do not by themselves shape a nuanced brand voice. Guardrails are complementary and should be used alongside prompt instructions to prevent disallowed content.

Model fine-tuning can imprint style into a model and provide durable behavior but it requires more data, cost, and operational effort. Fine tuning is appropriate for persistent, large scale needs when prompts alone cannot reach the required consistency.

When you need to control tone or style try well crafted prompts and short exemplars first and use temperature only to adjust variability while reserving fine tuning for persistent or large scale requirements.

A fintech startup that processes mobile payments notices its classifier achieves very high precision but noticeably lower recall. They want a single evaluation metric that fairly balances both. What does the F1 score represent?

  • ✓ C. The harmonic mean of precision and recall

The harmonic mean of precision and recall is the correct option because it combines precision and recall into a single score that falls when either component is low.

The F1 score computes the harmonic mean so it favors solutions that have both reasonable precision and reasonable recall rather than one very high value and the other very low. This property makes the F1 score useful for the fintech startup when both false positives and false negatives carry cost and a single balanced metric is needed.

The arithmetic sum of precision and recall is incorrect because simply adding the two can give a misleadingly high value when one measure is poor and it does not penalize imbalance the way the harmonic mean does.

The average of the true positive rate and the true negative rate is incorrect because that describes balanced accuracy and it incorporates true negatives, which the F1 score does not use.

Amazon SageMaker is incorrect because it is an AWS service for building and deploying machine learning models and it is not a metric that represents precision or recall.

When you must balance precision and recall on an exam pick the F1 measure since the harmonic mean penalizes imbalance and so it drops when either value is low.

A travel booking startup named SkyWay Tours is launching a generative AI support assistant and wants to enforce fairness, transparency, and continuous safety checks for models in production. Which AWS capability best meets these responsible AI needs?

  • ✓ C. Use Amazon SageMaker Clarify with Amazon SageMaker Model Monitor to detect bias and continuously track model behavior

The correct choice is Use Amazon SageMaker Clarify with Amazon SageMaker Model Monitor to detect bias and continuously track model behavior.

Amazon SageMaker Clarify provides tools to detect bias and generate explainability reports during data preparation and model evaluation, and Amazon SageMaker Model Monitor continuously inspects inputs and predictions in production to surface drift and quality issues. Together these services enable fairness and transparency and they support continuous safety checks which match the responsible AI needs of SkyWay Tours.

Customers are solely responsible for bias and safety monitoring without any AWS tools is incorrect because AWS supplies services to help detect bias and monitor models in production. Customers retain responsibility for governance and decision making, but they can use Amazon SageMaker Clarify and Amazon SageMaker Model Monitor to automate detection and reporting.

AWS infrastructure has no involvement in ethical aspects of AI deployments is incorrect because AWS offers specific features and services that address responsible AI considerations. AWS provides tooling for explainability and monitoring while customers remain accountable for policies and use of models.

AWS guarantees that any model hosted on AWS is unbiased by default is incorrect because no provider can guarantee a model is bias free without evaluation and mitigation. Customers must assess datasets and models and use tools such as Amazon SageMaker Clarify and Amazon SageMaker Model Monitor to identify and reduce bias.

When a question mentions bias fairness explainability or continuous monitoring think of services that detect bias and monitor production models and map them to SageMaker Clarify and Model Monitor.

A travel-tech startup, AeroTrail, is building an assistant on Amazon Bedrock that must carry out multi-turn reasoning over several steps, select and invoke external tools at runtime, and make HTTP calls to third-party APIs based on intermediate outcomes. Which AWS service or feature should the team choose to meet these needs?

  • ✓ B. Amazon Bedrock Agents

Amazon Bedrock Agents is the correct choice because it extends foundation models with planning, memory, and runtime tool use so an assistant can carry out multi turn reasoning, break tasks into steps, and invoke external APIs based on intermediate outcomes.

Amazon Bedrock Agents supplies agent components that enable decomposition of tasks, state management across turns, and dynamic selection of tools during inference. This makes it possible to integrate HTTP calls to third party APIs as part of the agent workflow so the assistant can decide which API to call after evaluating intermediate results.

AWS Step Functions is incorrect because it orchestrates explicit, predefined workflows and states rather than providing LLM native planning or autonomous tool selection at inference time. Step Functions is better suited to fixed pipelines and long running state machines.

Amazon SageMaker JumpStart is incorrect because it provides prebuilt models and starter solutions for rapid prototyping and deployment rather than an integrated agent framework for multi turn reasoning and dynamic tool or API invocation by a model at runtime.

Amazon Augmented AI (A2I) is incorrect because it focuses on human in the loop review and validation of model outputs and not on enabling autonomous API calls or multi turn decision making by an LLM.

When a question describes an LLM that needs multi turn reasoning together with dynamic tool or API calls prefer solutions designed for agent behavior such as Bedrock Agents rather than generic workflow orchestrators.

Aurora Media Group built a multilingual foundation model that translates everyday text well but misses terminology used in oil and gas compliance reports; the team plans to fine-tune the model in Amazon SageMaker with about 1,600 domain documents and a bilingual glossary to improve accuracy; which data preparation approach is most critical to capture this specialized vocabulary?

  • ✓ C. Curate and label domain-parallel examples using a consistent glossary

The correct choice is Curate and label domain-parallel examples using a consistent glossary. This approach directly targets the vocabulary gap by teaching the model how specialized oil and gas compliance terms map across languages during fine tuning in Amazon SageMaker.

Curate and label domain-parallel examples using a consistent glossary works because parallel sentence pairs provide context for how terms are used and a consistent bilingual glossary enforces the canonical translations for domain terminology. High quality aligned examples let the model learn phrase patterns and term senses that do not appear in general corpora. Preparing focused, labeled pairs also lets you control frequency of rare terms and correct ambiguous translations so the fine tuned model reliably uses the right compliance vocabulary.

Amazon Translate is an inference translation service and not a data preparation method for training a custom model. It can apply custom terminology at runtime but it does not replace creating curated parallel training data for fine tuning.

Randomly sample sentences from the corpus without quality checks is problematic because noisy or misaligned examples dilute signal for rare technical terms. Random sampling without validation is unlikely to show the model consistent, correct usages of specialized vocabulary.

Rely solely on the base pre-trained model without any domain data curation will not adapt the model to the oil and gas compliance lexicon. A base foundation model can provide general language ability but it needs curated domain examples and glossary alignment to close the accuracy gap on specialized reports.

Prioritize quality over quantity when fine tuning. Create aligned sentence pairs that include glossary terms in context and use the glossary to label correct translations before training.

A sporting goods marketplace built a customer help assistant with Amazon Bedrock, and the team notices the assistant sometimes invents product specifications and claims that are not in the catalog. Which approach will best anchor responses to trusted company information?

  • ✓ B. Retrieval Augmented Generation (RAG)

The correct choice is Retrieval Augmented Generation (RAG). This approach anchors assistant replies to trusted company information by fetching authoritative sources at inference time.

By retrieving the live product catalog, manuals, and policy documents when generating answers Retrieval Augmented Generation (RAG) grounds responses in verifiable facts which reduces hallucinations and keeps answers current.

Lower temperature may make wording more consistent but it does not add missing knowledge and cannot guarantee factual accuracy.

Fine-tuning can help with style and domain adaptation but it bakes knowledge into the model making updates harder and it can still hallucinate for new or changing product data.

Guardrails help enforce safety and policy boundaries but they do not retrieve or verify facts so they do not prevent incorrect product claims.

When exam scenarios mention hallucinations or stale facts think RAG and grounding to external sources because changing temperature or adding guardrails affects style and safety not factual correctness.

Darcy Declute

Darcy DeClute is a Certified Cloud Practitioner and author of the Scrum Master Certification Guide. Popular both on Udemy and social media, Darcy’s @Scrumtuous account has well over 250K followers on Twitter/X.


Git, GitHub & GitHub Copilot Certification Made Easy

Want to get certified on the most popular AI, ML & DevOps technologies of the day? These five resources will help you get GitHub certified in a hurry.

Get certified in the latest AI, ML and DevOps technologies. Advance your career today.