AWS Machine Learning Certification Exam Dumps and Braindumps

All questions come from my Udemy AWS ML course, and certificationexams.pro
AWS Machine Learning Exam Topics
Despite the title of this article, this is not an AWS Machine Learning Associate Braindump in the traditional sense. Cheating provides no real learning or professional value. True success comes from understanding and applying machine learning principles to real AWS scenarios.
All of these questions are taken from the AWS ML Associate Udemy course and the certificationexams.pro website, which offers hundreds of AWS ML Associate Practice Questions and Real AWS Machine Learning Exam Questions.
Every item has been carefully written to align with the AWS Machine Learning Associate exam objectives. They reflect the tone, structure, and technical depth of the official certification but are original, instructor-created scenarios that teach as much as they test.
Use the AWS Machine Learning Exam Simulator to practice under timed conditions. Each question in the Machine Learning Associate Questions and Answers collection explains why the correct answer is right and why the others are not. This builds your reasoning and helps you think like a true ML professional.
AWS AI & ML Exam Simulator
If you want a structured and efficient way to prepare, the AWS ML Associate Exam Dump and AWS Machine Learning Sample Questions are designed to reinforce your knowledge through repetition and applied practice. These are study resources built for mastery, not shortcuts.
By understanding concepts such as feature engineering, data bias handling, and model tuning, you will develop the confidence needed to apply your skills both on the exam and in real-world machine learning projects.
Study with integrity, practice consistently, and approach your certification with a learning mindset. True success as an AWS Machine Learning Associate comes from mastering how data, models, and AWS AI services work together to build intelligent, reliable solutions.
Use the AWS Machine Learning Exam Simulator and AWS ML Associate Practice Test to prepare effectively and move closer to your certification. Build your skills, earn your credential, and advance your career in machine learning.
Git, GitHub & GitHub Copilot Certification Made Easy |
---|
Want to get certified on the most popular AI, ML & DevOps technologies of the day? These five resources will help you get GitHub certified in a hurry.
Get certified in the latest AI, ML and DevOps technologies. Advance your career today. |
AWS Machine Learning Braindumps & Exam Dumps
Question 1
Which modeling approach enables cost efficient predictive maintenance that adapts to changing conditions over time?
-
❏ A. Deep neural network with many layers
-
❏ B. Random Forest with curated features and periodic retraining
-
❏ C. SageMaker XGBoost trained once with no retraining
-
❏ D. Amazon Timestream
-
❏ E. Logistic regression
Question 2
Why choose Amazon Neptune for storing and querying highly connected data with fast relationship traversals?
-
❏ A. Optimized for complex SQL joins on relational tables
-
❏ B. Best for large-scale unstructured text ingestion with minimal ops
-
❏ C. Purpose-built graph database with nodes/edges and fast relationship traversals
-
❏ D. DynamoDB key-value access is better for graph traversals
Question 3
For S3 encryption requiring fine-grained key access control, automatic rotation every 12 months, and full audit logs, which AWS KMS approach should be used?
-
❏ A. AWS Secrets Manager with IAM and CloudTrail
-
❏ B. KMS customer managed keys with key policies, IAM, 12-month rotation, and CloudTrail
-
❏ C. SSE-S3 with bucket policies and CloudTrail
-
❏ D. AWS managed keys with bucket policies and CloudTrail
Question 4
Which setup ensures SageMaker training and processing use only private networking and enforce granular S3 access with no internet path?
-
❏ A. Route SageMaker traffic to S3 through a NAT Gateway
-
❏ B. Run SageMaker in private subnets with an S3 VPC endpoint plus endpoint and bucket policies
-
❏ C. Use public S3 endpoints secured with TLS
-
❏ D. Create interface endpoints for SageMaker APIs only and keep S3 public
Question 5
In CNN training, validation accuracy drops after about 36 epochs while training accuracy continues to rise; which actions reduce overfitting and maintain validation performance? (Choose 2)
-
❏ A. Amazon SageMaker Clarify
-
❏ B. Use early stopping on validation loss
-
❏ C. Increase model depth and width
-
❏ D. Amazon SageMaker Automatic Model Tuning
-
❏ E. Add dropout layers for regularization
Question 6
Which mapping of Spot, On-Demand, and Reserved Instances best balances cost, resiliency, and latency for real-time inference, 30-day batch retraining, and large hyperparameter tuning? (Choose 3)
-
❏ A. Hyperparameter tuning on Spot
-
❏ B. Real-time endpoints on Spot
-
❏ C. 30-day retraining on Reserved Instances
-
❏ D. Real-time inference on On-Demand
-
❏ E. Hyperparameter tuning on Reserved Instances
Question 7
In Amazon QuickSight, which feature lets viewers choose values at runtime (for example, territory and product line) to dynamically filter visuals?
-
❏ A. Row-level security
-
❏ B. Parameters with controls
-
❏ C. Drill-through actions
-
❏ D. Drill-down hierarchy
Question 8
What primary capability does Amazon Macie provide for S3 data to reduce PII exposure risk before model training?
-
❏ A. S3 encryption with KMS keys
-
❏ B. Automatic discovery and classification of sensitive data
-
❏ C. Built-in masking or anonymization
-
❏ D. Real-time VPC flow inspection
-
❏ E. AWS Lake Formation fine-grained access control
Question 9
Which AWS deployment fits low-latency bursty recommendations, event-driven fraud detection, and nightly batch forecasting at 02:00 UTC?
-
❏ A. SageMaker Asynchronous Inference for recommender, Kinesis + Lambda for fraud, AWS Batch for forecasting
-
❏ B. Amazon EKS for recommender, SageMaker real-time endpoints for fraud, AWS Batch for forecasting
-
❏ C. SageMaker real-time endpoint for recommender, AWS Lambda for fraud events, Amazon ECS scheduled tasks for forecasting
-
❏ D. SageMaker Serverless Inference for recommender, AWS Batch for fraud, SageMaker Processing for forecasting
Question 10
Which fully managed AWS service best extracts unique key phrases and custom entities from about 150,000 text documents per month?
-
❏ A. Amazon OpenSearch Service
-
❏ B. Amazon Comprehend with custom entity recognition
-
❏ C. Amazon Kendra
-
❏ D. Amazon Bedrock

All questions come from my Udemy AWS ML course, and certificationexams.pro
Question 11
Which AWS approach enables scalable near-real-time SQL analytics on a streaming ingest with minimal data loss?
-
❏ A. Amazon Kinesis Data Firehose to Amazon Redshift
-
❏ B. Amazon MSK to Amazon OpenSearch Service
-
❏ C. Use Amazon Kinesis Data Streams with Kinesis Data Analytics (SQL)
-
❏ D. Amazon Kinesis Data Firehose to Amazon S3 and query with Amazon Athena
Question 12
How should S3 data be partitioned and lifecycle-managed to optimize Athena queries over the latest 15 days while archiving after 90 days?
-
❏ A. Bundle each day into one large gzipped file and query 15-day trends
-
❏ B. Convert to Parquet with Snappy but keep data unpartitioned and filter for 15 days
-
❏ C. Partition by year=YYYY/month=MM/day=DD and transition to S3 Glacier after 90 days
-
❏ D. Store data in S3 Glacier Flexible Retrieval from day 1 and query with Athena after restore
Question 13
Which configurations best meet strong encryption, least-privilege access, and auditability for a batch Amazon Comprehend job processing text in Amazon S3? (Choose 3)
-
❏ A. Attach AmazonS3FullAccess to the execution role
-
❏ B. Turn on CloudTrail and publish Comprehend metrics and logs to CloudWatch
-
❏ C. Skip encrypting the output
-
❏ D. Enforce TLS in transit and KMS encryption for input, output, and logs
-
❏ E. IAM role scoped to only required S3 prefixes and log resources
-
❏ F. Use SSE-S3 instead of KMS to simplify setup
Question 14
For an inference API with under 120 requests most days but spikes to 6,000 within minutes, which SageMaker option auto-scales for bursts and avoids idle instance cost?
-
❏ A. Amazon SageMaker Asynchronous Inference
-
❏ B. Amazon SageMaker Serverless Endpoints
-
❏ C. Amazon SageMaker Real-Time Inference
-
❏ D. Amazon SageMaker Multi-Model Endpoints
Question 15
For highly imbalanced (~0.5% positive) real-time fraud classification on tabular features using SageMaker built-in algorithms, which algorithm best handles imbalance and achieves strong accuracy?
-
❏ A. Random Cut Forest
-
❏ B. XGBoost with scale_pos_weight and F1 tuning
-
❏ C. Linear Learner with class weights
-
❏ D. IP Insights
Question 16
Which AWS service provides managed, low-latency real-time ML inference with autoscaling and high availability?
-
❏ A. Amazon ECS
-
❏ B. Amazon SageMaker Endpoints
-
❏ C. AWS Lambda
-
❏ D. AWS App Runner
-
❏ E. Amazon EC2 Auto Scaling
Question 17
A 25 TB image dataset is on Amazon FSx for NetApp ONTAP in the same VPC as SageMaker; what is the most efficient way for the training job to access it?
-
❏ A. Migrate to Amazon FSx for Lustre and use a file system channel
-
❏ B. Copy to Amazon S3 and read from S3
-
❏ C. Use SageMaker Pipe mode from S3 after syncing the data
-
❏ D. Mount the FSx for ONTAP NFS export to the SageMaker training job
Question 18
How can you automatically discover and remove PII in about 25 TB of CSV and JSON data in Amazon S3 with minimal operations?
-
❏ A. AWS Glue DataBrew jobs for PII scan and cleanup
-
❏ B. Amazon S3 Object Lambda to redact PII on access
-
❏ C. Amazon Macie with EventBridge and AWS Lambda for redaction
-
❏ D. Amazon Comprehend PII detection across S3 objects
Question 19
For large-scale distributed deep learning and real-time inference under 8 ms, which AWS training and inference setup is most performant and cost-efficient?
-
❏ A. SageMaker trn1 for training; ml.g5 for inference
-
❏ B. SageMaker p4d for training; ml.inf1 for inference
-
❏ C. EC2 g4dn for both training and inference
-
❏ D. SageMaker p3 for training; ml.m6i for inference
Question 20
In SageMaker multi-node GPU training, which placement and data-locality actions reduce inter-node latency and improve throughput? (Choose 2)
-
❏ A. Use an EC2 spread placement group
-
❏ B. Co-locate training data and compute in the same Region and AZ
-
❏ C. Place all training instances in one subnet and AZ
-
❏ D. Keep training data in a different Region
-
❏ E. Spread instances across multiple subnets
Question 21
Which approach enables scalable feature engineering and SageMaker training on 50 million DynamoDB items while avoiding high-latency point reads from the source table?
-
❏ A. Query DynamoDB via Athena federated connector and read with Athena
-
❏ B. Run an AWS Glue ETL job to write DynamoDB data to Amazon S3 and train from S3
-
❏ C. Use Boto3 from a SageMaker notebook to read items on demand during preprocessing
-
❏ D. Use SageMaker Pipe mode to stream directly from DynamoDB into the training job
Question 22
Which AWS service monitors Amazon Kendra indexing metrics and lets you set alarms that trigger notifications?
-
❏ A. Amazon SNS
-
❏ B. AWS CloudTrail
-
❏ C. Amazon EventBridge
-
❏ D. Amazon CloudWatch Alarms
-
❏ E. AWS Lambda
Question 23
How can you alert when a SageMaker real-time endpoint’s Invocations exceed a threshold during 8 p.m.–11 p.m.?
-
❏ A. Use AWS CloudTrail Insights to alert on API volume
-
❏ B. Route SageMaker events to Amazon EventBridge and trigger on a count threshold
-
❏ C. Create a CloudWatch Logs metric filter from endpoint logs and alarm on matches
-
❏ D. Set a CloudWatch alarm on the SageMaker endpoint Invocations metric with SNS notification
Question 24
For a regression model predicting a continuous target, which metric reports the average absolute error in the target’s units?
-
❏ A. Amazon SageMaker Clarify
-
❏ B. Area Under the ROC Curve (AUC-ROC)
-
❏ C. Mean Absolute Error (MAE)
-
❏ D. Root Mean Squared Error (RMSE)
Question 25
Which AWS services best provide low-latency, stateful 30-second window aggregations over Kinesis Data Streams with minimal operations? (Choose 2)
-
❏ A. AWS Lambda
-
❏ B. Amazon Managed Service for Apache Flink
-
❏ C. Amazon EMR with Apache Spark Streaming
-
❏ D. Amazon Kinesis Data Analytics
-
❏ E. Amazon MSK
-
❏ F. AWS Glue streaming jobs
AWS Machine Learning Braindump Answers

All questions come from my Udemy AWS ML course, and certificationexams.pro
Question 1
Which modeling approach enables cost efficient predictive maintenance that adapts to changing conditions over time?
-
✓ B. Random Forest with curated features and periodic retraining
The best choice is Random Forest with curated features and periodic retraining. Random Forests handle non-linear relationships and noisy sensor signals well, train efficiently compared to deep nets, and are straightforward to refresh on new data. Periodic retraining enables the model to adapt to concept drift as operating conditions evolve, maintaining accuracy while controlling costs.
The option Deep neural network with many layers is typically compute-intensive, requires extensive tuning, and is harder to operate under strict budget constraints.
Logistic regression is inexpensive but often underfits complex failure dynamics, leading to poor recall on rare events.
Amazon Timestream is a time series database for storage and queries, not a modeling technique, so it does not satisfy the predictive requirement.
SageMaker XGBoost trained once with no retraining locks the model to an initial data distribution; without scheduled updates it cannot adapt to drift, which is critical in predictive maintenance.
When a question emphasizes cost efficiency and adapting to changing conditions, look for approaches that support periodic retraining or pipelines for continuous updates. Be wary of answers that are services unrelated to modeling (databases, streaming engines) or that suggest one-time training with no plan for drift. Simpler ensemble methods plus good feature engineering often beat deep models when budgets and MLOps simplicity matter.
Question 2
Why choose Amazon Neptune for storing and querying highly connected data with fast relationship traversals?
-
✓ C. Purpose-built graph database with nodes/edges and fast relationship traversals
Purpose-built graph database with nodes/edges and fast relationship traversals is correct because Amazon Neptune is designed specifically for graph data models, representing entities and relationships as nodes and edges. It supports Gremlin and SPARQL for efficient multi-hop traversals and low-latency queries on highly connected datasets, which is ideal for relationship-centric workloads.
The option Optimized for complex SQL joins on relational tables is incorrect because Neptune is not a relational database and does not use SQL; relational joins suggest Amazon RDS or Amazon Aurora.
The option Best for large-scale unstructured text ingestion with minimal ops is incorrect because unstructured text search and analytics align more with Amazon OpenSearch Service or data lakes on Amazon S3.
The option DynamoDB key-value access is better for graph traversals is incorrect because while DynamoDB is excellent for key-based lookups, it is not optimized for complex, multi-hop relationship traversals common in graph workloads.
When you see keywords like graph, relationships, nodes and edges, traversals, or Gremlin/SPARQL, think Amazon Neptune. References to SQL joins point to relational databases, and unstructured text often points to OpenSearch or S3-based analytics.
Question 3
For S3 encryption requiring fine-grained key access control, automatic rotation every 12 months, and full audit logs, which AWS KMS approach should be used?
-
✓ B. KMS customer managed keys with key policies, IAM, 12-month rotation, and CloudTrail
KMS customer managed keys with key policies, IAM, 12-month rotation, and CloudTrail is correct because customer managed KMS keys allow you to define precise key policies and use IAM and grants to tightly scope access, enable automatic annual rotation for symmetric keys, and record all KMS API usage in CloudTrail for compliance-grade auditing. For S3, use SSE-KMS pointing to this CMK to meet encryption and governance requirements.
AWS managed keys with bucket policies and CloudTrail is insufficient because AWS controls the key policy, preventing fine-grained, customer-defined permissions and governance.
SSE-S3 with bucket policies and CloudTrail relies on S3 managed keys, which do not support customer key policies or customer-controlled rotation.
AWS Secrets Manager with IAM and CloudTrail is designed for storing secrets, not for managing KMS keys, policies, or rotation used by S3 encryption.
When you see phrases like tightly scoped key permissions, automatic 12-month rotation, and full audit of key operations, map to KMS customer managed keys with key policies plus CloudTrail. Remember that automatic rotation applies to symmetric CMKs and occurs every 365 days, and that AWS managed or S3 managed keys do not give you the same policy control.
Question 4
Which setup ensures SageMaker training and processing use only private networking and enforce granular S3 access with no internet path?
-
✓ B. Run SageMaker in private subnets with an S3 VPC endpoint plus endpoint and bucket policies
Run SageMaker in private subnets with an S3 VPC endpoint plus endpoint and bucket policies is correct because placing SageMaker training and processing jobs in private subnets forces all egress through VPC-controlled paths, and using an Amazon S3 VPC endpoint (Gateway) keeps S3 access on the AWS network without internet traversal. Combining an endpoint policy with a bucket policy that restricts access to the specific VPC endpoint (for example via aws:SourceVpce) provides the granular, governance-grade controls required for confidential data.
Route SageMaker traffic to S3 through a NAT Gateway is wrong because a NAT Gateway uses the public internet, breaking the requirement for fully private networking.
Use public S3 endpoints secured with TLS is wrong because encryption in transit does not eliminate the use of internet paths.
Create interface endpoints for SageMaker APIs only and keep S3 public is wrong because privatizing SageMaker service APIs does not privatize S3 data-plane traffic; S3 would still be accessed via public endpoints.
When a question emphasizes “no internet path” and “granular S3 controls,” think private subnets + S3 VPC endpoint (Gateway) + endpoint policy + bucket policy. Avoid NAT Gateways, internet gateways, or public S3 endpoints. Watch for distractors that add monitoring (for example, flow logs) or encryption (TLS) without removing public routing.
Question 5
In CNN training, validation accuracy drops after about 36 epochs while training accuracy continues to rise; which actions reduce overfitting and maintain validation performance? (Choose 2)
-
✓ B. Use early stopping on validation loss
-
✓ E. Add dropout layers for regularization
Use early stopping on validation loss and Add dropout layers for regularization directly address the pattern of rising training accuracy with declining validation accuracy. Early stopping halts training at the point of best validation performance, preventing the model from continuing to memorize the training set. Dropout injects stochastic regularization by randomly zeroing activations during training, which curbs co-adaptation and improves generalization.
Amazon SageMaker Automatic Model Tuning can find good hyperparameters, but without explicit regularization or a stopping criterion it will not, by itself, stop overfitting within an individual training run.
Increase model depth and width raises capacity and typically intensifies overfitting when validation metrics are already degrading.
Amazon SageMaker Clarify addresses bias and explainability and does not mitigate overfitting during training.
When you see training metrics improving while validation worsens, think capacity-independent regularization and training-time controls such as dropout, weight decay, data augmentation, and early stopping. If a choice only tunes parameters or analyzes data without affecting training dynamics, it likely won’t fix overfitting.
Question 6
Which mapping of Spot, On-Demand, and Reserved Instances best balances cost, resiliency, and latency for real-time inference, 30-day batch retraining, and large hyperparameter tuning? (Choose 3)
-
✓ A. Hyperparameter tuning on Spot
-
✓ C. 30-day retraining on Reserved Instances
-
✓ D. Real-time inference on On-Demand
The best mapping is Real-time inference on On-Demand, 30-day retraining on Reserved Instances, and Hyperparameter tuning on Spot. Real-time endpoints need immediate, reliable capacity with consistent latency, which On-Demand provides without interruption risk. A predictable monthly retraining baseline can commit to RIs to lock in discounts over a 1- or 3-year term. Hyperparameter tuning is batchy and fault-tolerant with checkpointing, making Spot the most cost-effective choice despite possible interruptions.
The option Real-time endpoints on Spot is unsuitable because interruptions can occur at short notice, harming availability and latency SLOs.
The option Hyperparameter tuning on Reserved Instances is inefficient because HPO usage is bursty and variable, so long-term commitments reduce flexibility and may waste reservations.
Map workloads by interruption tolerance and predictability. Use On-Demand for latency-critical services, Reserved Instances (or Savings Plans) for steady or predictable baselines, and Spot for interruptible, checkpointed jobs like training and HPO. Watch for keywords like latency-sensitive (On-Demand), predictable monthly (Reserved), and interruptible/batch/experiments (Spot).
Question 7
In Amazon QuickSight, which feature lets viewers choose values at runtime (for example, territory and product line) to dynamically filter visuals?
-
✓ B. Parameters with controls
The correct choice is Parameters with controls. In QuickSight, parameters can be surfaced through dashboard controls (such as dropdowns or text boxes) so viewers set values at runtime. These parameter values can bind to filters across visuals, enabling dynamic, interactive filtering without redesigning visuals.
The option Row-level security is incorrect because it restricts data access by user but does not give viewers filter inputs.
The option Drill-through actions is incorrect since it enables navigation to another sheet or target with context applied rather than general, user-entered filtering across visuals.
The option Drill-down hierarchy is incorrect because it only changes the level of detail within a visual and does not provide a runtime control for selecting values.
When you see requirements for user-driven, runtime selection of values that affect multiple visuals, look for parameters with controls. If the requirement mentions access restriction per user or group, think row-level security. If it mentions navigating to detail pages, think drill-through/actions; if it mentions changing grain in-place, think hierarchies.
Question 8
What primary capability does Amazon Macie provide for S3 data to reduce PII exposure risk before model training?
-
✓ B. Automatic discovery and classification of sensitive data
Automatic discovery and classification of sensitive data is correct because Amazon Macie’s core function is to scan Amazon S3 and automatically identify, classify, and alert on sensitive data such as PII. This reduces exposure risk by enabling you to act on findings before downstream processing or training.
S3 encryption with KMS keys is incorrect because encryption at rest is provided by S3 and AWS KMS, not Macie.
Built-in masking or anonymization is incorrect since Macie does not redact or transform data; it only reports findings.
Real-time VPC flow inspection is unrelated to Macie and pertains to network telemetry and services like VPC Flow Logs or GuardDuty.
AWS Lake Formation fine-grained access control is incorrect because it governs permissions and table-level controls, not content discovery.
When you see Macie in a question, associate it with automated sensitive data discovery in S3, not encryption, masking, or network monitoring. Look for keywords like PII identification, classification, and S3 scanning.
Question 9
Which AWS deployment fits low-latency bursty recommendations, event-driven fraud detection, and nightly batch forecasting at 02:00 UTC?
-
✓ C. SageMaker real-time endpoint for recommender, AWS Lambda for fraud events, Amazon ECS scheduled tasks for forecasting
The best mapping is SageMaker real-time endpoint for recommender, AWS Lambda for fraud events, Amazon ECS scheduled tasks for forecasting. SageMaker real-time endpoints provide consistently low latency with autoscaling for bursty traffic, Lambda naturally handles event-driven invocation from payment events with spiky patterns, and ECS scheduled tasks (via EventBridge) are a straightforward fit for nightly batch forecasting.
The option SageMaker Asynchronous Inference for recommender, Kinesis + Lambda for fraud, AWS Batch for forecasting is suboptimal because Asynchronous Inference introduces queueing and higher response times, which conflicts with ultra-low-latency recommendations. The choice Amazon EKS for recommender, SageMaker real-time endpoints for fraud, AWS Batch for forecasting misaligns the event-driven fraud pattern with endpoints and does not maximize the managed ML serving benefits for the recommender.
The option SageMaker Serverless Inference for recommender, AWS Batch for fraud, SageMaker Processing for forecasting suffers from serverless cold starts for real-time spikes, Batch not being real-time for fraud, and Processing being aimed at data preparation rather than scheduled inference.
Match workloads to invocation patterns. Use real-time endpoints for low-latency online inference, Lambda for event-driven compute, and scheduled containers (ECS) or Batch for offline jobs. Look for keywords like ultra-low latency, event-driven, and nightly batch to quickly map to the right service.
Question 10
Which fully managed AWS service best extracts unique key phrases and custom entities from about 150,000 text documents per month?
-
✓ B. Amazon Comprehend with custom entity recognition
Amazon Comprehend with custom entity recognition is the right choice because it natively extracts key phrases at scale and allows training a custom entity recognizer for domain-specific entities, all as a fully managed service with low operational overhead. It supports batch processing through asynchronous APIs and scales to large document volumes with minimal maintenance.
The option Amazon OpenSearch Service is incorrect because it is a search and analytics engine that would require custom analyzers and ingestion pipelines and does not provide native semantic key phrase or custom entity extraction.
The option Amazon Kendra is incorrect because it is built for enterprise search and content retrieval, not for generating de-duplicated key phrase or entity outputs from text corpora.
The option Amazon Bedrock is incorrect because it focuses on foundation models and generative AI; while possible with custom prompts and orchestration, it is not a turnkey, managed solution for key phrase and custom entity extraction like Comprehend.
When you see requirements for key phrases plus domain-specific entities with fully managed and minimal operations, prefer Amazon Comprehend. If the need is search over documents, think Kendra or OpenSearch. If the requirement mentions OCR, think Textract. Distinguish task-specific managed NLP (Comprehend) from general search (Kendra/OpenSearch) and generative FM platforms (Bedrock).
Question 11
Which AWS approach enables scalable near-real-time SQL analytics on a streaming ingest with minimal data loss?
-
✓ C. Use Amazon Kinesis Data Streams with Kinesis Data Analytics (SQL)
The correct choice is Use Amazon Kinesis Data Streams with Kinesis Data Analytics (SQL). Kinesis Data Streams provides durable, scalable ingestion with shard-based throughput and at-least-once delivery, minimizing data loss if producers or consumers restart. Kinesis Data Analytics (SQL) runs continuous SQL over the stream with low latency, enabling true near-real-time analytics.
The option Amazon Kinesis Data Firehose to Amazon Redshift is micro-batch oriented with buffering intervals, so it cannot deliver continuous streaming SQL over in-flight data.
The option Amazon MSK to Amazon OpenSearch Service targets search and log analytics, not streaming SQL; OpenSearch is optimized for indexing and search queries rather than continuous SQL processing.
The option Amazon Kinesis Data Firehose to Amazon S3 and query with Amazon Athena is also batch-based; Athena queries data at rest in S3, which introduces buffering and object commit delays unsuited to near-real-time streaming SQL.
Map use cases carefully. Choose Kinesis Data Streams for custom streaming ingestion and low-latency processing, Kinesis Data Analytics for SQL or Flink on streams, Firehose for managed delivery with buffering to destinations like S3/Redshift/OpenSearch, Athena for ad-hoc SQL on S3 data, and OpenSearch for search and log analytics. For minimal data loss and continuous queries, think KDS + KDA (SQL).
Question 12
How should S3 data be partitioned and lifecycle-managed to optimize Athena queries over the latest 15 days while archiving after 90 days?
-
✓ C. Partition by year=YYYY/month=MM/day=DD and transition to S3 Glacier after 90 days
Partition by year=YYYY/month=MM/day=DD and transition to S3 Glacier after 90 days is correct because date-based partitions allow Athena to prune to only the latest 15 days, minimizing scanned bytes and query cost, while an S3 Lifecycle transition to Glacier satisfies the archival requirement at lower storage cost.
The option Convert to Parquet with Snappy but keep data unpartitioned and filter for 15 days is wrong because without partitions, Athena still scans large portions of data even if the query filters by date. Parquet helps, but the lack of partition pruning keeps costs high.
The option Bundle each day into one large gzipped file and query 15-day trends is wrong since compression alone does not provide partition pruning, and very large files can still trigger broad scans. It also omits the required lifecycle archival step.
The option Store data in S3 Glacier Flexible Retrieval from day 1 and query with Athena after restore is incorrect because Athena cannot directly read Glacier storage classes; objects must be restored to S3 first, introducing delays and operational overhead.
Prefer time-based partitions for time-slice queries to enable partition pruning. Use columnar formats like Parquet with Snappy for additional scan reduction. Apply S3 Lifecycle rules to transition older partitions to cheaper tiers. Avoid both many tiny files and single huge files; size files appropriately for Athena and S3 performance.
Question 13
Which configurations best meet strong encryption, least-privilege access, and auditability for a batch Amazon Comprehend job processing text in Amazon S3? (Choose 3)
-
✓ B. Turn on CloudTrail and publish Comprehend metrics and logs to CloudWatch
-
✓ D. Enforce TLS in transit and KMS encryption for input, output, and logs
-
✓ E. IAM role scoped to only required S3 prefixes and log resources
Enforce TLS in transit and KMS encryption for input, output, and logs, IAM role scoped to only required S3 prefixes and log resources, and Turn on CloudTrail and publish Comprehend metrics and logs to CloudWatch together provide defense in depth for data confidentiality, access control, and auditability. Comprehend supports using KMS keys for input and output data, TLS protects data in transit, least-privilege IAM restricts the blast radius, and CloudTrail plus CloudWatch deliver traceability and monitoring needed for compliance.
The option Attach AmazonS3FullAccess to the execution role is overly permissive and violates least-privilege, increasing risk of unintended access. The choice Skip encrypting the output is unsafe because output can contain sensitive derived content and must be encrypted. The distractor Use SSE-S3 instead of KMS to simplify setup may encrypt at rest but lacks the key management and control typically required by strong security policies, making it insufficient for strict compliance.
When a question emphasizes strong protection and auditability for AWS ML services, look for KMS-based encryption at rest, TLS in transit, least-privilege IAM scoped to specific S3 prefixes, and centralized auditing via CloudTrail and CloudWatch. Be wary of broad managed policies and any suggestion to skip encrypting outputs. If a distractor proposes weaker encryption like SSE-S3 when KMS is called for, prefer KMS for tighter key control and compliance.
Question 14
For an inference API with under 120 requests most days but spikes to 6,000 within minutes, which SageMaker option auto-scales for bursts and avoids idle instance cost?
-
✓ B. Amazon SageMaker Serverless Endpoints
The correct choice is Amazon SageMaker Serverless Endpoints. It automatically scales concurrency based on incoming requests and charges only for compute used during invocation, eliminating idle instance costs. This makes it ideal for unpredictable, bursty traffic where you want low operational overhead and no manual capacity planning.
Amazon SageMaker Asynchronous Inference is optimized for long-running or large payload jobs where responses can be retrieved later, not for low-latency interactive APIs during sudden bursts.
Amazon SageMaker Real-Time Inference relies on provisioned instances that accrue cost even when idle, which conflicts with the goal of minimizing idle cost for sporadic traffic.
Amazon SageMaker Multi-Model Endpoints can consolidate multiple models on shared instances to reduce costs, but they still require provisioned instances and therefore do not eliminate idle capacity charges or management.
Look for cues like spiky traffic, avoid idle cost, and no manual capacity planning to map to serverless endpoints. If the prompt mentions long-running or large payloads with delayed responses, think asynchronous inference. When you see consistent low-latency SLAs with predictable load, real-time provisioning is typically correct. If many models must be hosted behind one endpoint, consider multi-model endpoints but remember they still use provisioned capacity.
Question 15
For highly imbalanced (~0.5% positive) real-time fraud classification on tabular features using SageMaker built-in algorithms, which algorithm best handles imbalance and achieves strong accuracy?
-
✓ B. XGBoost with scale_pos_weight and F1 tuning
The best choice is XGBoost with scale_pos_weight and F1 tuning. Gradient-boosted trees are state-of-the-art for tabular, heterogeneous features and support direct class-imbalance handling via scale_pos_weight, enabling tuning toward precision, recall, and F1 for rare-event fraud detection. With labels available, supervised boosting will typically outperform unsupervised anomaly detectors and linear baselines.
Random Cut Forest is incorrect because it is unsupervised and does not leverage labels to optimize supervised metrics like precision/recall for known fraud classes.
Linear Learner with class weights is not ideal because linear models miss nonlinear feature interactions and generally underperform boosted trees on complex fraud patterns even with weighting.
IP Insights is designed for detecting unusual entity-to-IP relationships and is unsupervised, so it is not a fit for supervised transaction classification with labels. On the exam, when you see imbalanced labeled tabular data, prefer boosted-tree methods in SageMaker. Look for cues like scale_pos_weight or tuning for precision/recall/F1, and avoid unsupervised algorithms when labels are available.

All questions come from my Udemy AWS ML course, and certificationexams.pro
Question 16
Which AWS service provides managed, low-latency real-time ML inference with autoscaling and high availability?
-
✓ B. Amazon SageMaker Endpoints
Amazon SageMaker Endpoints is correct because it offers fully managed real-time inference with automatic scaling via Application Auto Scaling, multi-AZ high availability, integrated monitoring, and deployment features purpose-built for production ML workloads.
The option Amazon ECS is incorrect because, while it can host containers, you must implement the model server, autoscaling, health checks, and multi-AZ strategies yourself, increasing operational complexity.
The option AWS Lambda is incorrect due to cold starts, timeouts, and concurrency constraints that make it less suitable for consistent, low-latency, high-throughput inference.
The option AWS App Runner is incorrect because it runs and scales containers but lacks ML-native inference features like model endpoints, variants, and built-in metrics for model serving.
The option Amazon EC2 Auto Scaling is incorrect since it only scales instances and does not provide a managed inference endpoint or deployment workflow.
When you see keywords like managed real-time inference, low latency, autoscaling, and high availability, think SageMaker real-time endpoints. If the scenario mentions offline or large-scale batch predictions, prefer SageMaker Batch Transform. If occasional, event-driven inference is acceptable with some latency tolerance, Lambda can fit. If you need full control over containers and networking and can manage the stack, ECS/EKS may be used but is not ML-native.
Question 17
A 25 TB image dataset is on Amazon FSx for NetApp ONTAP in the same VPC as SageMaker; what is the most efficient way for the training job to access it?
-
✓ D. Mount the FSx for ONTAP NFS export to the SageMaker training job
Mount the FSx for ONTAP NFS export to the SageMaker training job is correct because the data is already on a high-throughput network file system within the same VPC. Mounting and reading in place avoids any bulk transfers, minimizes latency, and maximizes throughput for large image datasets during training.
The option Migrate to Amazon FSx for Lustre and use a file system channel is unnecessary because it introduces a full data migration step even though Lustre is performant.
The option Copy to Amazon S3 and read from S3 adds time, cost, and S3 request latency, which is less efficient for many files compared to direct NFS reads.
The option Use SageMaker Pipe mode from S3 after syncing the data still requires moving data to S3 and may be throughput limited versus in-VPC NFS, making it suboptimal when in-place access is available.
When the dataset already resides on a network file system in the same VPC as SageMaker, prefer in-place access via a mount. When data is in S3 and low-latency POSIX access is needed, consider FSx for Lustre. If the question emphasizes streaming from S3 without local copies, consider Pipe mode, but recognize it still depends on S3 and associated bandwidth and request patterns.
Question 18
How can you automatically discover and remove PII in about 25 TB of CSV and JSON data in Amazon S3 with minimal operations?
-
✓ C. Amazon Macie with EventBridge and AWS Lambda for redaction
Amazon Macie with EventBridge and AWS Lambda for redaction is correct because Macie is a managed service purpose built for discovering sensitive data in Amazon S3. It natively generates findings for PII and integrates with EventBridge so you can invoke Lambda to automatically redact or quarantine objects, achieving minimal ongoing operations while maintaining compliance and automation.
AWS Glue DataBrew jobs for PII scan and cleanup is not ideal because you must build and maintain recipes, schedules, and scaling, which adds operational overhead versus a managed PII discovery service.
Amazon S3 Object Lambda to redact PII on access is not sufficient because it still requires you to implement custom PII detection and transformation logic and does not provide automated discovery across at-rest S3 data.
Amazon Comprehend PII detection across S3 objects would require custom orchestration to crawl and process large S3 datasets and manage retries and scaling, leading to higher complexity and operations.
When the question emphasizes low operations overhead and PII discovery in S3, prefer Amazon Macie. Look for integrations with EventBridge and Lambda to automate remediation. Services like DataBrew or Comprehend can work but typically require more custom pipelines and orchestration.
Question 19
For large-scale distributed deep learning and real-time inference under 8 ms, which AWS training and inference setup is most performant and cost-efficient?
-
✓ B. SageMaker p4d for training; ml.inf1 for inference
SageMaker p4d for training; ml.inf1 for inference is the best match because p4d provides A100 GPUs for highly scalable distributed training, while AWS Inferentia on ml.inf1 is purpose-built for high-throughput, low-latency, and cost-efficient real-time inference. This combination aligns with both the performance and cost goals for sub-8 ms serving. The SageMaker trn1 for training; ml.g5 for inference choice can train effectively, but GPU inference on g5 is generally costlier per throughput and not as latency-optimized as Inferentia for large fleets. The EC2 g4dn for both training and inference setup is inadequate for massive training scale and struggles to hit ultra-low latency at high throughput. The SageMaker p3 for training; ml.m6i for inference choice pairs older training hardware with CPU inference, which is unlikely to meet strict millisecond latency for deep models. On the exam, map distributed training and scale to A100-class instances such as p4d (or p5) and map single-digit ms, cost-efficient inference to Inferentia (Inf1/Inf2). Watch for distractors that use CPUs or general-purpose GPUs for inference when low latency and cost efficiency are emphasized.
Question 20
In SageMaker multi-node GPU training, which placement and data-locality actions reduce inter-node latency and improve throughput? (Choose 2)
-
✓ B. Co-locate training data and compute in the same Region and AZ
-
✓ C. Place all training instances in one subnet and AZ
The best choices are to keep communication paths as short and local as possible. Placing the compute fleet together and keeping the data close reduces hops, routing, and cross-AZ or cross-Region penalties. Therefore, Co-locate training data and compute in the same Region and AZ and Place all training instances in one subnet and AZ together minimize inter-node latency and improve throughput in tight, collective communication patterns common in distributed training.
The option Use an EC2 spread placement group focuses on failure isolation, distributing instances across distinct hardware, which works against node adjacency and can increase latency.
The option Keep training data in a different Region incurs high latency and egress costs, which is counterproductive for frequent parameter synchronization.
The option Spread instances across multiple subnets can add routing overhead and jitter, increasing inter-node latency rather than reducing it.
For latency-sensitive distributed training, look for cues like inter-node communication and all-reduce. Favor same AZ, single subnet, and adjacency. Consider cluster placement groups and EFA for even lower latency and higher throughput. Avoid cross-AZ and cross-Region data paths. Terms like multi-AZ for HA or spread placement usually indicate higher latency, not better performance.
Question 21
Which approach enables scalable feature engineering and SageMaker training on 50 million DynamoDB items while avoiding high-latency point reads from the source table?
-
✓ B. Run an AWS Glue ETL job to write DynamoDB data to Amazon S3 and train from S3
The best choice is Run an AWS Glue ETL job to write DynamoDB data to Amazon S3 and train from S3. Moving data to S3 provides a durable, scalable data lake that SageMaker can read in parallel with high throughput. This avoids high-latency, per-item reads from DynamoDB and decouples your training jobs from the transactional workload. Glue offers managed, scalable ETL for bulk export and transformation, which is ideal for feature engineering at this scale.
The option Query DynamoDB via Athena federated connector and read with Athena is suboptimal for sustained training I/O because federated queries introduce Lambda invocation overhead and are intended for interactive analytics, not continuous, large-scale data feeds.
The option Use Boto3 from a SageMaker notebook to read items on demand during preprocessing relies on point reads which cause high latency and potential throttling, making it unreliable and slow for tens of millions of items.
The option Use SageMaker Pipe mode to stream directly from DynamoDB into the training job is not supported; Pipe mode streams from S3 or supported file systems, and even if it did, it would still place unnecessary load on DynamoDB.
For large-scale training, prefer staging data in S3 (or an S3-backed offline store) to enable parallelized, high-throughput reads. Avoid designs that hit transactional stores during training. Look for keywords like bulk export, decouple, and S3 to identify the best answer.
Question 22
Which AWS service monitors Amazon Kendra indexing metrics and lets you set alarms that trigger notifications?
-
✓ D. Amazon CloudWatch Alarms
Amazon CloudWatch Alarms is correct because Amazon Kendra publishes operational metrics and logs to CloudWatch, where you can create alarms on error or failure metrics and trigger notifications (for example, via Amazon SNS) or automated actions.
The option Amazon SNS is incorrect because it only delivers notifications and requires another service like CloudWatch to detect the error condition.
Amazon EventBridge is incorrect because it focuses on event routing and rules, not on collecting Kendra metrics or creating metric-based alarms.
AWS Lambda is incorrect since it is compute for running code and does not provide native monitoring or alarms.
AWS CloudTrail is incorrect because it logs API activity, not service metrics or alarms for indexing jobs.
When you see keywords like metrics, alarms, or thresholds, think CloudWatch. If the requirement is notification delivery, think SNS. For event routing and rule-based triggers, think EventBridge. For API auditing, think CloudTrail.
Question 23
How can you alert when a SageMaker real-time endpoint’s Invocations exceed a threshold during 8 p.m.–11 p.m.?
-
✓ D. Set a CloudWatch alarm on the SageMaker endpoint Invocations metric with SNS notification
The correct approach is to use Set a CloudWatch alarm on the SageMaker endpoint Invocations metric with SNS notification. SageMaker real-time endpoints publish AWS/SageMaker metrics such as Invocations and Invocation4XXErrors to CloudWatch. You can create a threshold alarm on Invocations and configure an SNS action for alerts. To limit alerts to specific hours (8 p.m.–11 p.m.), enable or disable the alarm on a schedule using EventBridge, or use composite alarms to gate notifications during that window.
The option Use AWS CloudTrail Insights to alert on API volume is incorrect because CloudTrail primarily captures management-plane activity and Insights detects anomalies in those events; it does not monitor data-plane invoke counts for endpoints.
The option Route SageMaker events to Amazon EventBridge and trigger on a count threshold is incorrect since EventBridge lacks built-in metric aggregation and threshold alarms; you would need custom aggregation.
The option Create a CloudWatch Logs metric filter from endpoint logs and alarm on matches is unreliable because endpoints do not produce access logs by default and log-based counting is indirect and brittle compared to native metrics.
When you see terms like Invocations, endpoint metrics, and threshold alerts, think CloudWatch metrics plus alarms with SNS. Distinguish services: CloudWatch = metrics/alarms, EventBridge = event routing, CloudTrail = audit/management APIs. For time-bound alerts, pair alarms with EventBridge schedules to enable/disable during specific hours.
Question 24
For a regression model predicting a continuous target, which metric reports the average absolute error in the target’s units?
-
✓ C. Mean Absolute Error (MAE)
Mean Absolute Error (MAE) is correct because it directly averages the absolute differences between predictions and ground truth and reports the result in the target’s units, providing an intuitive measure of typical error magnitude for regression.
Area Under the ROC Curve (AUC-ROC) is a classification ranking metric and does not apply to continuous regression targets.
Amazon SageMaker Clarify is a service for bias detection and explainability, not an evaluation metric.
Root Mean Squared Error (RMSE) is a valid regression metric and shares the target’s units, but it computes the square root of the mean squared error, emphasizing larger errors rather than reporting an average absolute error as asked.
Watch for keywords like average absolute error and same units, which point to MAE. Distinguish regression metrics (MAE, MSE, RMSE, R2) from classification metrics (accuracy, precision, recall, AUC). When the question emphasizes interpretability in the target’s units and typical error magnitude, MAE is often the best fit.
Question 25
Which AWS services best provide low-latency, stateful 30-second window aggregations over Kinesis Data Streams with minimal operations? (Choose 2)
-
✓ B. Amazon Managed Service for Apache Flink
-
✓ D. Amazon Kinesis Data Analytics
Amazon Managed Service for Apache Flink and Amazon Kinesis Data Analytics are purpose-built for stateful, low-latency stream processing with windowed aggregations over Kinesis Data Streams. Flink provides rich state management and event-time windows with exactly-once semantics, while Kinesis Data Analytics enables concise SQL window functions to compute rolling metrics with minimal operational overhead. Both meet the 30-second rolling window and low-latency requirement without managing clusters.
The option AWS Lambda is not ideal because maintaining cross-shard, stateful 30-second windows is complex, error-prone, and can hit scaling or state limits.
Amazon EMR with Apache Spark Streaming can do windows but increases operational burden and latency due to cluster management and tuning.
Amazon MSK is a managed Kafka ingestion service, not an analytics engine for Kinesis streams, and does not solve windowed aggregation needs here.
AWS Glue streaming jobs focus on micro-batch ETL and generally do not deliver the lowest-latency rolling aggregations compared to Flink or Kinesis Data Analytics.
When you see keywords like stateful windowed aggregations, low latency, and minimal operations on Kinesis Data Streams, favor stream processors like Flink or Kinesis Data Analytics. Avoid ingestion-only services and heavy cluster-managed approaches unless the scenario explicitly requires custom frameworks or large-scale batch/stream hybrids.
Jira, Scrum & AI Certification |
---|
Want to get certified on the most popular software development technologies of the day? These resources will help you get Jira certified, Scrum certified and even AI Practitioner certified so your resume really stands out..
You can even get certified in the latest AI, ML and DevOps technologies. Advance your career today. |
Cameron McKenzie is an AWS Certified AI Practitioner, Machine Learning Engineer, Copilot Expert, Solutions Architect and author of many popular books in the software development and Cloud Computing space. His growing YouTube channel training devs in Java, Spring, AI and ML has well over 30,000 subscribers.