Amazon AWS Certified Machine Learning Engineer - Associate MLA-C01 Exam (page: 1)
Amazon AWS Certified Machine Learning Engineer - Associate MLA-C01
Updated on: 12-Feb-2026

Case Study

A company is building a web-based AI application by using Amazon SageMaker. The application will provide the following capabilities and features: ML experimentation, training, a central model registry, model deployment, and model monitoring.

The application must ensure secure and isolated use of training data during the ML lifecycle. The training data is stored in Amazon S3.

The company needs to use the central model registry to manage different versions of models in the application.

Which action will meet this requirement with the LEAST operational overhead?

  1. Create a separate Amazon Elastic Container Registry (Amazon ECR) repository for each model.
  2. Use Amazon Elastic Container Registry (Amazon ECR) and unique tags for each model version.
  3. Use the SageMaker Model Registry and model groups to catalog the models.
  4. Use the SageMaker Model Registry and unique tags for each model version.

Answer(s): C

Explanation:

The SageMaker Model Registry is specifically designed to manage the lifecycle of machine learning models, including versioning, deployment, and monitoring. By using model groups, the registry allows cataloging and organizing models based on different criteria, such as use case or project. This approach minimizes operational overhead by providing an integrated solution within SageMaker for model versioning and management.



Case Study

A company is building a web-based AI application by using Amazon SageMaker. The application will provide the following capabilities and features: ML experimentation, training, a central model registry, model deployment, and model monitoring.

The application must ensure secure and isolated use of training data during the ML lifecycle. The training data is stored in Amazon S3.

The company is experimenting with consecutive training jobs.

How can the company MINIMIZE infrastructure startup times for these jobs?

  1. Use Managed Spot Training.
  2. Use SageMaker managed warm pools.
  3. Use SageMaker Training Compiler.
  4. Use the SageMaker distributed data parallelism (SMDDP) library.

Answer(s): B

Explanation:

SageMaker managed warm pools help minimize infrastructure startup times for training jobs by keeping instances warm and ready to be reused for subsequent jobs. This significantly reduces the initialization time that is typically required when starting new training jobs, making it ideal for scenarios involving consecutive training jobs. This approach ensures efficient utilization of resources with minimal delays between jobs.



Case Study

A company is building a web-based AI application by using Amazon SageMaker. The application will provide the following capabilities and features: ML experimentation, training, a central model registry, model deployment, and model monitoring.

The application must ensure secure and isolated use of training data during the ML lifecycle. The training data is stored in Amazon S3.

The company must implement a manual approval-based workflow to ensure that only approved models can be deployed to production endpoints.

Which solution will meet this requirement?

  1. Use SageMaker Experiments to facilitate the approval process during model registration.
  2. Use SageMaker ML Lineage Tracking on the central model registry. Create tracking entities for the approval process.
  3. Use SageMaker Model Monitor to evaluate the performance of the model and to manage the approval.
  4. Use SageMaker Pipelines.
    When a model version is registered, use the AWS SDK to change the approval status to "Approved."

Answer(s): D

Explanation:

SageMaker Pipelines is a purpose-built feature for creating, automating, and managing ML workflows. It integrates seamlessly with the SageMaker Model Registry, which supports setting approval statuses for model versions. By using the AWS SDK to update the model's status to "Approved," the company can implement a manual approval process that ensures only approved models are deployed to production. This approach is efficient and aligns well with the requirement for manual approvals while leveraging SageMaker's built-in capabilities.



Case Study

A company is building a web-based AI application by using Amazon SageMaker. The application will provide the following capabilities and features: ML experimentation, training, a central model registry, model deployment, and model monitoring.

The application must ensure secure and isolated use of training data during the ML lifecycle. The training data is stored in Amazon S3.

The company needs to run an on-demand workflow to monitor bias drift for models that are deployed to real- time endpoints from the application.

Which action will meet this requirement?

  1. Configure the application to invoke an AWS Lambda function that runs a SageMaker Clarify job.
  2. Invoke an AWS Lambda function to pull the sagemaker-model-monitor-analyzer built-in SageMaker image.
  3. Use AWS Glue Data Quality to monitor bias.
  4. Use SageMaker notebooks to compare the bias.

Answer(s): A

Explanation:

SageMaker Clarify is designed to detect and monitor bias in ML models and datasets. By running a Clarify job, the company can analyze the deployed model for bias drift. Configuring the application to invoke an AWS Lambda function to trigger the SageMaker Clarify job allows for on-demand and automated monitoring of bias drift in real-time endpoints. This solution ensures operational efficiency and meets the requirement for secure and automated bias monitoring.



HOTSPOT

A company stores historical data in .csv files in Amazon S3. Only some of the rows and columns in the .csv files are populated. The columns are not labeled. An ML engineer needs to prepare and store the data so that the company can use the data to train ML models.

Select and order the correct steps from the following list to perform this task. Each step should be selected one time or not at all. (Select and order three.)

· Create an Amazon SageMaker batch transform job for data cleaning and feature engineering.
· Store the resulting data back in Amazon S3.
· Use Amazon Athena to infer the schemas and available columns. · Use AWS Glue crawlers to infer the schemas and available columns. · Use AWS Glue DataBrew for data cleaning and feature engineering.

Hot Area:

  1. See Explanation section for answer.

Answer(s): A

Explanation:

The correct steps, in order, are:
1. Use AWS Glue crawlers to infer the schemas and available columns.
- AWS Glue crawlers can automatically scan the .csv files in Amazon S3, detect the schema, and catalog the data for further processing.
2. Use AWS Glue DataBrew for data cleaning and feature engineering.
- AWS Glue DataBrew provides tools for cleaning, transforming, and preparing the data for ML tasks.
3. Store the resulting data back in Amazon S3.
- After cleaning and preparing the data, the resulting dataset can be stored back in Amazon S3 for training ML models.



HOTSPOT

An ML engineer needs to use Amazon SageMaker Feature Store to create and manage features to train a model.

Select and order the steps from the following list to create and use the features in Feature Store. Each step should be selected one time. (Select and order three.)

· Access the store to build datasets for training.
· Create a feature group.
· Ingest the records.

Hot Area:

  1. See Explanation section for answer.

Answer(s): A

Explanation:




The correct steps, in order, are:
4. Create a feature group.
A feature group defines the schema and structure for the features, serving as a container for storing and organizing features.
5. Ingest the records.
Populate the feature group by ingesting data records, which include the features and their associated values.
6. Access the store to build datasets for training.
Retrieve features from the Feature Store to construct datasets for model training.



HOTSPOT

A company wants to host an ML model on Amazon SageMaker. An ML engineer is configuring a continuous integration and continuous delivery (CI/CD) pipeline in AWS CodePipeline to deploy the model. The pipeline must run automatically when new training data for the model is uploaded to an Amazon S3 bucket.

Select and order the pipeline's correct steps from the following list. Each step should be selected one time or not at all. (Select and order three.)

· An S3 event notification invokes the pipeline when new data is uploaded. · S3 Lifecycle rule invokes the pipeline when new data is uploaded. · SageMaker retrains the model by using the data in the S3 bucket.
· The pipeline deploys the model to a SageMaker endpoint.
· The pipeline deploys the model to SageMaker Model Registry.

Hot Area:

  1. See Explanation section for answer.

Answer(s): A

Explanation:

The correct steps, in order, are:
7. An S3 event notification invokes the pipeline when new data is uploaded. Set up an S3 event notification to trigger the pipeline when new training data is added to the S3 bucket.

8. SageMaker retrains the model by using the data in the S3 bucket. The pipeline should include a step to retrain the ML model using the new data in the S3 bucket.

9. The pipeline deploys the model to a SageMaker endpoint. After retraining, the pipeline deploys the updated model to a SageMaker endpoint for inference.



HOTSPOT

An ML engineer is building a generative AI application on Amazon Bedrock by using large language models (LLMs).

Select the correct generative AI term from the following list for each description. Each term should be selected one time or not at all.

· Embedding
· Retrieval Augmented Generation (RAG)
· Temperature
· Token

Hot Area:

  1. See Explanation section for answer.

Answer(s): A

Explanation:




The correct terms for each description are:
10. Text representation of basic units of data processed by LLMs - Token Tokens are the basic units of text (such as words or subwords) that LLMs process.

11. High-dimensional vectors that contain the semantic meaning of text - Embedding Embeddings are numerical representations of text in high-dimensional space, capturing semantic meaning.
12. Enrichment of information from additional data sources to improve a generated response - Retrieval Augmented Generation (RAG)
RAG involves retrieving relevant information from external data sources to enhance the quality of generated responses.



Viewing Page 1 of 30



Share your comments for Amazon AWS Certified Machine Learning Engineer - Associate MLA-C01 exam with other users:

kriah 9/4/2023 10:44:00 PM

please upload the latest dumps
UNITED STATES


ed 12/17/2023 1:41:00 PM

a company runs its workloads on premises. the company wants to forecast the cost of running a large application on aws. which aws service or tool can the company use to obtain this information? pricing calculator ... the aws pricing calculator is primarily used for estimating future costs
UNITED STATES


Muru 12/29/2023 10:23:00 AM

looks interesting
Anonymous


Tech Lady 10/17/2023 12:36:00 PM

thanks! that’s amazing
Anonymous


Mike 8/20/2023 5:12:00 PM

the exam dumps are helping me get a solid foundation on the practical techniques and practices needed to be successful in the auditing world.
UNITED STATES


Nobody 9/18/2023 6:35:00 PM

q 14 should be dmz sever1 and notepad.exe why does note pad have a 443 connection
Anonymous


Muhammad Rawish Siddiqui 12/4/2023 12:17:00 PM

question # 108, correct answers are business growth and risk reduction.
SAUDI ARABIA


Emmah 7/29/2023 9:59:00 AM

are these valid chfi questions
KENYA


Mort 10/19/2023 7:09:00 PM

question: 162 should be dlp (b)
EUROPEAN UNION


Eknath 10/4/2023 1:21:00 AM

good exam questions
INDIA


Nizam 6/16/2023 7:29:00 AM

I have to say this is really close to real exam. Passed my exam with this.
EUROPEAN UNION


poran 11/20/2023 4:43:00 AM

good analytics question
Anonymous


Antony 11/23/2023 11:36:00 AM

this looks accurate
INDIA


Ethan 8/23/2023 12:52:00 AM

question 46, the answer should be data "virtualization" (not visualization).
Anonymous


nSiva 9/22/2023 5:58:00 AM

its useful.
UNITED STATES


Ranveer 7/26/2023 7:26:00 PM

Pass this exam 3 days ago. The PDF version and the Xengine App is quite useful.
SOUTH AFRICA


Sanjay 8/15/2023 10:22:00 AM

informative for me.
UNITED STATES


Tom 12/12/2023 8:53:00 PM

question 134s answer shoule be "dlp"
JAPAN


Alex 11/7/2023 11:02:00 AM

in 72 the answer must be [sys_user_has_role] table.
Anonymous


Finn 5/4/2023 10:21:00 PM

i appreciated the mix of multiple-choice and short answer questions. i passed my exam this morning.
IRLAND


AJ 7/13/2023 8:33:00 AM

great to find this website, thanks
UNITED ARAB EMIRATES


Curtis Nakawaki 6/29/2023 9:11:00 PM

examination questions seem to be relevant.
UNITED STATES


Umashankar Sharma 10/22/2023 9:39:00 AM

planning to take psm test
Anonymous


ED SHAW 7/31/2023 10:34:00 AM

please allow to download
UNITED STATES


AD 7/22/2023 11:29:00 AM

please provide dumps
UNITED STATES


Ayyjayy 11/6/2023 7:29:00 AM

is the answer to question 15 correct ? i feel like the answer should be b
BAHRAIN


Blessious Phiri 8/12/2023 11:56:00 AM

its getting more technical
Anonymous


Jeanine J 7/11/2023 3:04:00 PM

i think these questions are what i need.
UNITED STATES


Aderonke 10/23/2023 2:13:00 PM

helpful assessment
UNITED KINGDOM


Tom 1/5/2024 2:32:00 AM

i am confused about the answers to the questions. do you know if the answers are correct?
KOREA REPUBLIC OF


Vinit N. 8/28/2023 2:33:00 AM

hi, please make the dumps available for my upcoming examination.
UNITED STATES


Sanyog Deshpande 9/14/2023 7:05:00 AM

good practice
UNITED STATES


Tyron 9/8/2023 12:12:00 AM

so far it is really informative
Anonymous


beast 7/30/2023 2:22:00 PM

hi i want it please please upload it
Anonymous