AWS Practice Tests

AWS Certified Machine Learning Engineer – Associate (MLA-C01) Mock Test

Free mock exam for AWS Certified Machine Learning Engineer Associate (MLA-C01)
Written by Arslan Khan

The AWS Certified Machine Learning Engineer – Associate (MLA-C01) certification is a new and vital credential for professionals focused on building, training, tuning, and deploying machine learning (ML) models on AWS. As ML continues to transform industries, this certification validates your expertise in this cutting-edge field. To thoroughly prepare for the MLA-C01 exam, integrating MLA-C01 mock tests into your study plan is indispensable. These practice exams are meticulously designed to align with the MLA-C01 exam guide, covering essential domains such as data engineering for ML, exploratory data analysis, ML modeling, ML implementation and operations, and business understanding for ML.

Engaging with AWS Machine Learning Engineer Associate practice exams offers a realistic simulation of the actual test environment. You’ll tackle questions that assess your ability to use core AWS ML services like Amazon SageMaker for the entire ML lifecycle, alongside data services such as S3, Glue, and Kinesis for preparing and processing data. These mock tests are crucial for identifying your strengths and pinpointing areas where you need further study, whether it’s in feature engineering, model training algorithms, hyperparameter optimization, or deploying models for inference. Regularly working through MLA-C01 practice questions will sharpen your problem-solving skills in real-world ML scenarios.

Beyond just testing your knowledge, these practice exams build your confidence and improve your time-management skills for the actual exam. Familiarizing yourself with the question types and the depth of understanding required for ML engineering on AWS will significantly reduce exam-day anxiety. A robust AWS MLA-C01 preparation strategy involves not only learning the theory behind ML algorithms and AWS services but also understanding their practical application in building and deploying scalable ML solutions. Start leveraging MLA-C01 mock tests today to solidify your expertise and significantly increase your chances of earning your AWS Certified Machine Learning Engineer – Associate certification.

Understanding the AWS Cloud is a valuable asset in today’s tech landscape. For detailed information about the certification, you can always refer to the official AWS Certified Developer – Associate (DVA-C02) page.

Begin your path to certification excellence—click ‘Begin’ to challenge yourself and succeed. You’ve got this!


This is a timed quiz. You will be given 7800 seconds to answer all questions. Are you ready?

7800
0%

A ML Engineer needs to ensure that a SageMaker endpoint is only accessible from within a specific VPC. Which networking configuration should be used?

This involves creating a private connection to the SageMaker service.
Show hint
Correct! Wrong!

You can configure a SageMaker endpoint to be accessible only from within your VPC by creating a VPC endpoint (using AWS PrivateLink) for SageMaker runtime. This keeps traffic within the AWS network.

A ML Engineer needs to ensure that the IAM role used by a SageMaker training job has only the necessary permissions to access specific S3 buckets for input data and output artifacts. This adheres to which security principle?

This principle is about granting minimal necessary permissions.
Show hint
Correct! Wrong!

The principle of least privilege states that an entity (user, role, service) should only be granted the minimum permissions necessary to perform its required tasks. This minimizes potential damage if the entity is compromised.

A ML Engineer is using AWS Glue crawlers to populate the AWS Glue Data Catalog with metadata from data stored in Amazon S3. What does a Glue crawler primarily create in the Data Catalog?

It creates schema definitions in the metadata repository.
Show hint
Correct! Wrong!

AWS Glue crawlers scan your data stores (like S3) and use classifiers to infer schemas and other metadata, then create tables in the AWS Glue Data Catalog.

What is the primary purpose of Amazon SageMaker Pipelines in MLOps?

It helps automate and orchestrate ML workflows.
Show hint
Correct! Wrong!

Amazon SageMaker Pipelines is a continuous integration and continuous delivery (CI/CD) service for machine learning (ML). It helps you automate different steps of your ML workflow, including data preparation, model building, model training, and model deployment.

What is the primary purpose of Amazon SageMaker Feature Store?

It's a centralized repository for ML features.
Show hint
Correct! Wrong!

Amazon SageMaker Feature Store is a fully managed, purpose-built repository to store, update, retrieve, and share machine learning (ML) features. It helps data science teams reuse features and ensure consistency between training and inference.

Which Amazon SageMaker feature allows you to automatically scale the number of instances for a real-time inference endpoint based on workload traffic?

This allows endpoints to handle varying amounts of traffic.
Show hint
Correct! Wrong!

Amazon SageMaker supports automatic scaling for your production variants hosted on an endpoint. Auto scaling dynamically adjusts the number of instances provisioned for a production variant in response to changes in your workload.

In a binary classification problem, what does the 'Precision' metric measure?

It's about the accuracy of positive predictions.
Show hint
Correct! Wrong!

Precision measures the proportion of true positive predictions among all positive predictions made by the model (TP / (TP + FP)). It answers the question: Of all instances predicted as positive, how many were actually positive?

Which Amazon SageMaker feature helps detect bias in your data and machine learning models, and explains model predictions?

This SageMaker tool focuses on fairness and explainability.
Show hint
Correct! Wrong!

Amazon SageMaker Clarify provides machine learning developers with greater visibility into their training data and models so they can identify and limit bias and explain predictions.

A ML Engineer is training a regression model to predict house prices. Which of the following is a common loss function used for regression tasks?

This loss function penalizes larger errors more heavily.
Show hint
Correct! Wrong!

Mean Squared Error (MSE) is a common loss function used for regression problems. It measures the average of the squares of the errors—that is, the average squared difference between the estimated values and the actual value.

Which technique is used to understand the importance of different features in predicting the outcome of a machine learning model?

It helps rank features by their predictive power.
Show hint
Correct! Wrong!

Feature importance techniques (e.g., permutation importance, SHAP values, tree-based feature importance) help identify which input features have the most significant impact on the model's predictions.

Which Amazon SageMaker feature helps you track, organize, and compare your machine learning experiments, including datasets, parameters, and metrics?

This feature is for managing and comparing ML experiments.
Show hint
Correct! Wrong!

Amazon SageMaker Experiments helps you organize, track, compare, and evaluate your machine learning experiments and model versions. It automatically captures input parameters, configurations, and results, and stores them as experiments.

Which AWS service is commonly used to trigger retraining pipelines in an MLOps workflow when model performance degrades or new data becomes available?

This service can react to events and trigger workflows.
Show hint
Correct! Wrong!

Amazon EventBridge (formerly CloudWatch Events) can be used to detect events (e.g., a CloudWatch alarm indicating model degradation, or an S3 PUT event for new data) and trigger downstream actions, such as starting a SageMaker Pipeline for retraining.

Which of the following is a common technique for handling imbalanced datasets in a classification problem?

This involves adjusting the class distribution in the training data.
Show hint
Correct! Wrong!

Oversampling the minority class (e.g., using SMOTE) or undersampling the majority class are common techniques to address class imbalance and help the model learn better from the minority class.

To optimize the cost of a SageMaker real-time endpoint that experiences infrequent and unpredictable traffic, which SageMaker inference option is MOST suitable?

This option scales to zero and charges per invocation.
Show hint
Correct! Wrong!

SageMaker Serverless Inference is designed for workloads with intermittent or unpredictable traffic. You pay only for the compute capacity used to process inference requests, and it automatically scales to zero when there's no traffic.

What is the primary benefit of using Apache Parquet or ORC file formats for storing data in an S3 data lake for ML training and analytics?

These formats are optimized for analytical workloads.
Show hint
Correct! Wrong!

Parquet and ORC are columnar storage file formats optimized for analytical query performance. They allow query engines and ML training jobs to read only the necessary columns, reducing I/O and improving processing speed.

A ML Engineer needs to run a data processing script on a large dataset stored in S3 before training a model. The script is written in Python and uses common libraries like Pandas and NumPy. Which Amazon SageMaker feature is designed for such ad-hoc or scheduled data processing jobs?

This SageMaker feature runs containerized processing scripts.
Show hint
Correct! Wrong!

Amazon SageMaker Processing jobs allow you to run data processing workloads for pre-processing, post-processing, feature engineering, data validation, and model evaluation on Amazon SageMaker. You can use built-in containers or bring your own.

When training a model with Amazon SageMaker, where are the final model artifacts typically stored by default if not specified otherwise?

This is the default durable storage location for SageMaker outputs.
Show hint
Correct! Wrong!

By default, SageMaker training jobs store the output model artifacts (the trained model) in an Amazon S3 bucket that SageMaker creates or that you specify in the training job configuration.

A ML Engineer wants to deploy multiple versions of a model to the same SageMaker endpoint and distribute traffic between them for A/B testing. What SageMaker feature supports this?

This involves deploying different model 'variants' to an endpoint.
Show hint
Correct! Wrong!

SageMaker endpoints support production variants, where you can deploy multiple model versions (or different models) to the same endpoint and configure traffic distribution (e.g., 90% to variant A, 10% to variant B) for A/B testing or canary deployments.

Which of the following is a common strategy to prevent overfitting when training a neural network?

This technique adds a penalty for model complexity or stops training early.
Show hint
Correct! Wrong!

Regularization techniques (like L1/L2 regularization or dropout) and early stopping are common methods to prevent overfitting, where the model performs well on training data but poorly on unseen data.

What is the purpose of a 'validation set' during model training?

It's used for tuning model hyperparameters and preventing overfitting on the training data.
Show hint
Correct! Wrong!

The validation set is used to tune hyperparameters and make decisions about the model architecture. It provides an unbiased evaluation of a model fit on the training dataset while tuning model hyperparameters. The test set is used for the final, unbiased evaluation.

Which Amazon SageMaker mode allows you to bring your own training script (e.g., a Python script using TensorFlow or PyTorch) and run it within a SageMaker-managed framework container?

This mode uses your custom script with SageMaker's containers.
Show hint
Correct! Wrong!

Script mode in Amazon SageMaker allows you to run your custom training scripts using SageMaker's pre-built framework containers (like TensorFlow, PyTorch, MXNet, Scikit-learn). You provide your script, and SageMaker handles the environment setup and execution.

What is 'distributed training' in the context of machine learning?

It involves training across multiple machines or GPUs.
Show hint
Correct! Wrong!

Distributed training involves splitting the model training workload across multiple compute resources (e.g., multiple GPUs or multiple instances) to accelerate the training process for large models or datasets.

If a model has high bias, what does this typically indicate about its performance?

This often means the model is 'underfitting'.
Show hint
Correct! Wrong!

High bias means the model is too simple and makes strong assumptions about the data, leading to underfitting. It performs poorly on both the training data and unseen test data because it fails to capture the underlying patterns.

When training a model, what does the 'learning rate' hyperparameter typically control?

It controls the step size during model optimization.
Show hint
Correct! Wrong!

The learning rate is a hyperparameter that controls how much to change the model in response to the estimated error each time the model weights are updated during training. A small learning rate may result in slow convergence, while a large learning rate may cause the training process to diverge.

Which Amazon SageMaker built-in algorithm is suitable for anomaly detection tasks, such as identifying unusual patterns in time-series data?

This algorithm is designed to find 'odd ones out'.
Show hint
Correct! Wrong!

SageMaker has built-in algorithms like Random Cut Forest (RCF) for anomaly detection. RCF is an unsupervised algorithm that detects anomalous data points within a data set.

What is the primary purpose of hyperparameter tuning (optimization) in machine learning?

It's about finding the best configuration settings for your model.
Show hint
Correct! Wrong!

Hyperparameters are external configuration settings for a learning algorithm. Hyperparameter tuning is the process of finding the optimal set of hyperparameters that yields the best model performance for a given dataset and problem.

A ML Engineer is using Amazon SageMaker Automatic Model Tuning (hyperparameter tuning job). What is the 'objective metric' that the tuning job tries to optimize?

It's the target metric for the tuning process.
Show hint
Correct! Wrong!

The objective metric is the specific model performance metric (e.g., validation:accuracy, validation:auc, validation:mse) that the hyperparameter tuning job aims to maximize or minimize to find the best model.

Which AWS service can be used to build a CI/CD pipeline that automates the build, test, and deployment of the infrastructure and code for an ML application?

This service orchestrates the full release pipeline.
Show hint
Correct! Wrong!

AWS CodePipeline is a continuous delivery service that automates the release process. It can be used to build CI/CD pipelines for ML applications, integrating with services like CodeCommit (source), CodeBuild (build/test), SageMaker (train/deploy), and CloudFormation (infrastructure).

What does the F1-score metric represent in a classification task?

It's a balance between precision and recall.
Show hint
Correct! Wrong!

The F1-score is the harmonic mean of precision and recall. It provides a single score that balances both concerns, and is often useful when you have an uneven class distribution.

What is a confusion matrix used for in evaluating a classification model?

It's a table showing correct and incorrect classifications.
Show hint
Correct! Wrong!

A confusion matrix is a table that summarizes the performance of a classification model by showing the counts of true positives, true negatives, false positives, and false negatives.

A ML Engineer needs to deploy a model for offline predictions on a large dataset that arrives daily. Which SageMaker deployment option is MOST cost-effective and suitable for this scenario?

This option is for non-real-time, large-scale predictions.
Show hint
Correct! Wrong!

SageMaker Batch Transform is ideal for getting inferences from your models for large datasets. It's suitable for offline processing where you don't need sub-second latency.

When dealing with categorical features that have a large number of unique values (high cardinality), which feature engineering technique can be problematic due to creating too many new features?

This technique creates a new column for each unique category.
Show hint
Correct! Wrong!

One-hot encoding creates a new binary feature for each unique category. For high cardinality categorical features, this can lead to a very large number of new features (the curse of dimensionality), potentially harming model performance and increasing computational cost.

What is 'feature scaling' and why is it important for some machine learning algorithms?

It brings features to a similar range of values.
Show hint
Correct! Wrong!

Feature scaling (e.g., normalization or standardization) transforms features to be on a similar scale. This is important for algorithms sensitive to feature magnitudes, like gradient descent-based algorithms (e.g., linear regression, neural networks) and distance-based algorithms (e.g., k-NN, SVM), as it helps them converge faster and perform better.

To ensure the security of model artifacts and data used by Amazon SageMaker, what is a recommended practice regarding network isolation for training jobs and endpoints?

This involves running SageMaker resources within your private network.
Show hint
Correct! Wrong!

Running SageMaker training jobs and hosting endpoints within a VPC without direct internet access (network isolation mode or using VPC endpoints) enhances security by controlling network traffic and reducing exposure.

Which Amazon SageMaker built-in algorithm is suitable for image classification tasks?

This algorithm is designed for categorizing images.
Show hint
Correct! Wrong!

Amazon SageMaker provides a built-in Image Classification algorithm that uses a convolutional neural network (CNN) and can be trained on your own image datasets or fine-tuned from pre-trained models.

Which Amazon SageMaker feature allows you to train machine learning models using built-in algorithms, custom algorithms in Docker containers, or scripts with pre-built framework containers (e.g., TensorFlow, PyTorch)?

This is the core SageMaker capability for model training.
Show hint
Correct! Wrong!

Amazon SageMaker training jobs provide a managed environment for training ML models. You can use SageMaker's built-in algorithms, bring your own custom algorithms packaged in Docker containers, or use script mode with framework containers.

Which evaluation metric is commonly used for regression models to measure the average squared difference between predicted and actual values?

It's the average of squared errors.
Show hint
Correct! Wrong!

Mean Squared Error (MSE) is a standard metric for regression tasks. It calculates the average of the squares of the differences between the predicted and actual values.

Which Amazon SageMaker capability allows you to visually browse, discover, and connect to data sources, and then prepare data for machine learning with over 300 built-in data transformations without writing code?

This SageMaker tool offers a visual interface for data preparation.
Show hint
Correct! Wrong!

Amazon SageMaker Data Wrangler reduces the time it takes to aggregate and prepare data for machine learning (ML) from weeks to minutes. With SageMaker Data Wrangler, you can simplify the process of data preparation and feature engineering, and complete each step of the data preparation workflow, including data selection, cleansing, exploration, and visualization from a single visual interface.

Which type of Amazon EC2 instances are specifically designed and optimized for machine learning training workloads, often featuring powerful GPUs?

These instances often have 'P' or 'Trn' in their family name.
Show hint
Correct! Wrong!

Amazon EC2 P-family instances (e.g., p3, p4d) are designed for general-purpose GPU compute applications and are well-suited for ML training. Trn-family instances are for training, Inf-family for inference.

When deploying a SageMaker model to an endpoint, what does the 'instance type' in the endpoint configuration specify?

It determines the compute resources for the inference endpoint.
Show hint
Correct! Wrong!

The instance type in the SageMaker endpoint configuration specifies the type of EC2 compute instance that will host your model for serving inference requests (e.g., ml.m5.large, ml.g4dn.xlarge).

A ML Engineer is working with a dataset that has many missing values in several numerical columns. Which data imputation technique involves replacing missing values with the central tendency of that column (e.g., mean or median)?

This technique uses the average or middle value of a column.
Show hint
Correct! Wrong!

Mean or median imputation is a common technique where missing values in a numerical column are replaced by the mean (average) or median (middle value) of the non-missing values in that same column.

When training a deep learning model on Amazon SageMaker, what is the role of an 'epoch'?

It represents one full pass through the training data.
Show hint
Correct! Wrong!

An epoch refers to one complete pass of the entire training dataset through the learning algorithm. Training deep learning models typically involves multiple epochs.

Which Amazon SageMaker feature allows you to capture input and output data for your deployed models, and detect deviations in data quality or model quality over time?

This feature helps monitor models in production for drift.
Show hint
Correct! Wrong!

Amazon SageMaker Model Monitor continuously monitors the quality of machine learning models in production. It can detect data drift and concept drift, and alert you when issues arise so you can retrain your models.

What is the purpose of the Amazon SageMaker Model Registry?

It helps catalog and manage model versions and their lifecycle.
Show hint
Correct! Wrong!

SageMaker Model Registry allows you to catalog your ML models, manage model versions, associate metadata (like performance metrics) with models, and manage the approval status of models before deployment, facilitating MLOps and governance.

The Area Under the ROC Curve (AUC) is a common evaluation metric for which type of machine learning problem?

This metric is often used when distinguishing between two classes.
Show hint
Correct! Wrong!

AUC-ROC is a performance measurement for classification problems at various threshold settings. ROC is a probability curve and AUC represents the degree or measure of separability. It tells how much the model is capable of distinguishing between classes.

Which SageMaker hyperparameter tuning strategy explores hyperparameter combinations randomly within the defined ranges?

This strategy doesn't systematically explore the hyperparameter space.
Show hint
Correct! Wrong!

Random search is a hyperparameter tuning strategy where combinations are chosen randomly from the defined search space. Bayesian optimization is more guided, while Grid search exhaustively tries all combinations.

When evaluating a binary classifier, if the cost of a false negative is very high (e.g., failing to detect a critical disease), which metric should be prioritized for optimization?

This metric focuses on minimizing missed positive cases.
Show hint
Correct! Wrong!

Recall (Sensitivity or True Positive Rate) measures the proportion of actual positives that were correctly identified (TP / (TP + FN)). If false negatives are costly, maximizing recall is crucial to minimize missed positive cases.

A ML Engineer has trained a model and now needs to evaluate its performance on data it has never seen before to get an unbiased estimate of its generalization ability. Which dataset should be used for this final evaluation?

This dataset is used for the final, unbiased performance check.
Show hint
Correct! Wrong!

The test set is a separate portion of the data held out from the training and validation processes. It is used only once, at the very end, to provide an unbiased estimate of how well the final chosen model will perform on new, unseen data.

A ML Engineer is using Amazon SageMaker Ground Truth to label a large image dataset for an object detection model. What is a key feature of Ground Truth that helps improve labeling accuracy and efficiency?

This feature can use ML to assist with labeling.
Show hint
Correct! Wrong!

SageMaker Ground Truth offers features like automated data labeling (which uses an ML model to label data automatically after an initial set is labeled by humans) and annotation consolidation to improve accuracy from multiple labelers.

A Machine Learning Engineer needs to prepare a large dataset stored in Amazon S3 for training. The preparation involves cleaning, transforming, and feature engineering. Which AWS service is MOST suitable for performing these data preparation tasks at scale in a serverless manner?

This service is a serverless ETL and data integration service.
Show hint
Correct! Wrong!

AWS Glue is a fully managed ETL (extract, transform, and load) service that makes it simple and cost-effective to categorize your data, clean it, enrich it, and move it reliably between various data stores and data streams. It's well-suited for serverless data preparation.

What is 'concept drift' in the context of a deployed machine learning model?

It's when the underlying relationship the model learned changes.
Show hint
Correct! Wrong!

Concept drift occurs when the statistical properties of the target variable that the model is trying to predict change over time. This means the relationship between input features and the target variable changes, leading to model performance degradation.

Which of the following is a common component of an MLOps pipeline for continuous training (CT)?

Think about automatically retraining models based on triggers.
Show hint
Correct! Wrong!

Continuous training involves automatically retraining models when new data arrives or when model performance degrades. This typically includes automated data validation, model retraining, model evaluation, and potentially model redeployment steps, often orchestrated by a pipeline.

A data engineer needs to ingest real-time sensor data from multiple devices into an AWS data lake for ML model training. The data needs to be durable and allow for multiple applications to consume it. Which AWS service is MOST suitable for this initial ingestion point?

This service is for real-time, durable data streaming.
Show hint
Correct! Wrong!

Amazon Kinesis Data Streams is designed for real-time data ingestion at scale. It provides durable storage for stream records and allows multiple consumer applications to process the data concurrently.

Which AWS service is commonly used in an MLOps pipeline to store and version machine learning model artifacts?

This service is for durable object storage and supports versioning.
Show hint
Correct! Wrong!

Amazon S3 is widely used for storing model artifacts due to its durability, scalability, and versioning capabilities. SageMaker Model Registry also provides model versioning and management on top of S3.

What is 'transfer learning' in the context of training machine learning models?

It involves reusing knowledge from a pre-trained model.
Show hint
Correct! Wrong!

Transfer learning is a technique where a model pre-trained on a large dataset for one task is adapted (fine-tuned) for a second, related task, often with a smaller dataset. This leverages the knowledge learned from the initial task.

What is the benefit of using Amazon SageMaker Neo to compile a trained machine learning model?

This service optimizes models for specific hardware targets.
Show hint
Correct! Wrong!

Amazon SageMaker Neo optimizes models to run up to twice as fast, with less than a tenth of the memory footprint, with no loss in accuracy. It compiles models for specific target hardware (cloud instances or edge devices).

A ML Engineer has trained a model using Amazon SageMaker and now needs to deploy it for real-time inference with low latency. Which SageMaker feature is used for this?

This feature hosts your model for live predictions.
Show hint
Correct! Wrong!

Amazon SageMaker Endpoints provide a way to deploy trained ML models for real-time inference. You create an endpoint configuration and then deploy the model to an endpoint, which can then be invoked by applications.

What is 'data drift' in the context of a deployed machine learning model?

It's when the live input data changes significantly from the training data.
Show hint
Correct! Wrong!

Data drift occurs when the statistical properties of the input data used for inference change over time compared to the data the model was trained on. This can lead to a degradation in model performance.

What is 'checkpointing' in the context of long-running SageMaker training jobs?

It allows training to resume from an intermediate state if interrupted.
Show hint
Correct! Wrong!

Checkpointing involves periodically saving the state of the model during a long training job. If the job is interrupted (e.g., due to a Spot Instance interruption), it can resume training from the last saved checkpoint instead of starting over, saving time and cost.

When using Amazon SageMaker, what is a 'training channel'?

It defines where SageMaker gets its input data from S3.
Show hint
Correct! Wrong!

Training channels in SageMaker specify the S3 locations of the input data for a training job (e.g., 'train' channel for training data, 'validation' channel for validation data).

Amazon SageMaker provides pre-built Docker images for popular ML frameworks. What is the primary benefit of using these framework containers?

They simplify environment setup for common ML frameworks.
Show hint
Correct! Wrong!

SageMaker's pre-built framework containers (e.g., for TensorFlow, PyTorch, Scikit-learn) provide managed environments with the necessary libraries and dependencies, simplifying the setup for training and inference and ensuring compatibility with SageMaker.

What does it mean if a machine learning model is 'overfitting'?

The model performs well on training data but poorly on new data.
Show hint
Correct! Wrong!

Overfitting occurs when a model learns the training data too well, including its noise and random fluctuations, and as a result, performs poorly on new, unseen data (e.g., the validation or test set).

Which type of data store is Amazon S3 primarily considered when used as a data lake for ML?

S3 stores data as files or 'objects'.
Show hint
Correct! Wrong!

Amazon S3 is an object storage service. In the context of data lakes, it stores data in its native format as objects (files), which can then be processed by various analytics and ML services.

What is 'cross-validation' used for in machine learning model evaluation?

It involves splitting data into multiple folds for training and testing.
Show hint
Correct! Wrong!

Cross-validation is a resampling technique used to evaluate ML models on a limited data sample. It helps provide a more robust estimate of model performance on unseen data and helps detect overfitting by training and testing the model on different subsets of the data.

A ML Engineer wants to capture the input data and predictions for a SageMaker real-time endpoint to monitor for data quality issues or model drift. Which SageMaker Model Monitor feature should be configured?

This feature captures inference request/response data.
Show hint
Correct! Wrong!

SageMaker Model Monitor allows you to enable data capture for your endpoints. It captures the request and response payloads and stores them in S3, which can then be analyzed for drift or data quality issues.

AWS Certified Machine Learning Engineer Associate (MLA-C01) Practice Exam
Excellent!
Great job! You're well-prepared for the AWS ML Engineer - Associate level concepts.
Good Effort!
Solid understanding! Review the explanations for any missed questions to strengthen your knowledge.
Needs More Practice
Keep studying the SageMaker documentation, practice hands-on labs, and try more mock tests.

Share your Results:

About the author

Arslan Khan

Arslan is a Senior Software Engineer, Cloud Engineer, and DevOps Specialist with a passion for simplifying complex cloud technologies. With years of hands-on experience in AWS architecture, automation, and cloud-native development, he writes practical, insightful blogs to help developers and IT professionals navigate the evolving world of cloud computing. When he's not optimizing infrastructure or deploying scalable solutions, he’s sharing knowledge through tutorials and thought leadership in the AWS and DevOps space.

Leave a Comment