AWS Certified AI Practitioner Practice Test (AIF-C01)
Use the form below to configure your AWS Certified AI Practitioner Practice Test (AIF-C01). The practice test can be configured to only include certain exam objectives and domains. You can choose between 5-100 questions and set a time limit.

AWS Certified AI Practitioner AIF-C01 Information
The AWS Certified AI Practitioner (AIF-C01) certification verifies that you have a strong foundational understanding of artificial intelligence, machine learning, and generative AI, along with exposure to how these are implemented via AWS services. It’s aimed at those who may not be developers or ML engineers but who need to understand and contribute to AI/ML-related decisions and initiatives in their organizations. The exam tests your knowledge of AI/ML concepts, use cases, and how AWS’s AI/ML services map to various business problems.
Topics include basics of supervised/unsupervised learning, generative AI fundamentals, prompt engineering, evaluation metrics, responsible AI practices, and how AWS tools like Amazon SageMaker, Amazon Bedrock, Amazon Comprehend, Amazon Rekognition, and others support AI workflows. Questions are scenario-based and focused on choosing the right service or strategy, rather than hands-on coding or architecture design. After passing, you’ll be able to identify the proper AI/ML tool for a given business challenge, articulate trade-offs, and guide responsible deployment of AI solutions within AWS environments.
Having this certification shows stakeholders that you have a solid conceptual grasp of AI and AWS’s AI/ML ecosystem. It’s well suited for technical leads, solution architects, product managers, or anyone who interacts with AI/ML teams and wants credibility in AI strategy discussions. It also helps bridge the gap between technical teams and business stakeholders around what AI/ML can — and cannot — do in real scenarios.

Free AWS Certified AI Practitioner AIF-C01 Practice Test
- 20 Questions
- Unlimited
- Fundamentals of AI and MLFundamentals of Generative AIApplications of Foundation ModelsGuidelines for Responsible AISecurity, Compliance, and Governance for AI Solutions
Free Preview
This test is a free preview, no account required.
Subscribe to unlock all content, keep track of your scores, and access AI features!
Within the context of machine learning, what does the term "bias" most accurately describe when evaluating a model's predictions against real-world outcomes?
The total number of trainable parameters present in a specific neural network layer.
The iterative procedure of adjusting model weights during backpropagation.
A consistent error introduced by algorithmic assumptions or training data that favors certain outcomes in the model's predictions.
Random fluctuations in predictions that arise solely from noisy data samples.
Answer Description
Bias is the systematic error that causes a model's predictions to be consistently skewed toward certain values or groups. This usually stems from assumptions within the algorithm or from imbalanced or representative issues in the training data. Random noise is variance, not bias. The count of parameters merely reflects model size, and weight updates during backpropagation describe the learning process, not bias itself.
Ask Bash
Bash is our AI bot, trained to help you pass your exam. AI Generated Content may display inaccurate information, always double-check anything important.
What is the difference between bias and variance in machine learning?
How does imbalanced training data contribute to bias in a machine learning model?
What strategies can be used to reduce bias in machine learning models?
An e-commerce retailer adds a generative-AI product-recommendation banner to its website. The team must pick a business metric to confirm that the model is driving more revenue. Which metric is MOST appropriate?
Token usage cost per model inference
Conversion rate of visitors who complete a purchase
Embedding vector dimensionality used by the model
Average handle time in customer-support chats
Answer Description
Conversion rate directly ties customer actions to revenue. If a higher percentage of site visitors who see the recommendation banner complete a purchase, the model is creating measurable value for the business. Average handle time measures support efficiency, not sales impact. Token usage cost tracks model-execution expenses, not income. Embedding vector dimensionality is a technical attribute that has no inherent connection to revenue generation.
Ask Bash
Bash is our AI bot, trained to help you pass your exam. AI Generated Content may display inaccurate information, always double-check anything important.
What is a conversion rate in the context of e-commerce?
How does generative AI impact e-commerce conversion rates?
Why are technical metrics like token usage cost and embedding vector dimensionality not suitable for revenue measurement?
An organization has fine-tuned a text-generation foundation model and now wants to expose it through a scalable HTTPS endpoint without provisioning or managing any GPU instances. Which AWS service is the most appropriate choice for this deployment stage?
AWS CloudFormation
Amazon Bedrock
Amazon S3
Amazon EC2 Auto Scaling
Answer Description
Amazon Bedrock is a fully managed service that hosts foundation models and exposes them through an API. Because Bedrock handles the underlying compute infrastructure, the company can deploy and scale its fine-tuned model without setting up or operating GPU instances.
- Amazon S3 is an object-storage service; it cannot run inference.
- Amazon EC2 Auto Scaling adds or removes virtual machines but still requires customers to build and manage the inference stack on those instances.
- AWS CloudFormation automates infrastructure provisioning but does not provide a managed model-hosting endpoint by itself.
Ask Bash
Bash is our AI bot, trained to help you pass your exam. AI Generated Content may display inaccurate information, always double-check anything important.
What is Amazon Bedrock?
How does Amazon Bedrock differ from Amazon EC2 Auto Scaling?
Why doesn’t Amazon S3 work for deploying an AI model?
An online media company stores thousands of recorded podcasts in Amazon S3. Editors need a managed AWS service that can automatically convert the speech in these audio files to searchable text without building custom ML models. Which service best fits this requirement?
Amazon Rekognition
Amazon Comprehend
Amazon Transcribe
Amazon Polly
Answer Description
Amazon Transcribe is a fully managed AI service that performs automatic speech recognition to convert audio streams or files into accurate text transcripts. Amazon Polly does the opposite (text-to-speech), Amazon Comprehend analyzes existing text for sentiment and entities, and Amazon Rekognition analyzes images and video, so none of those services meet the need to turn speech into text.
Ask Bash
Bash is our AI bot, trained to help you pass your exam. AI Generated Content may display inaccurate information, always double-check anything important.
What is Amazon Transcribe used for?
How does Amazon Polly differ from Amazon Transcribe?
Can Amazon Transcribe handle specialized audio content like medical or technical terms?
A company is deciding whether to invest in an AI/ML project. In which situation is an AI/ML solution least appropriate and likely to add unnecessary cost and complexity?
The security team needs to detect fraudulent credit card transactions that change over time.
The finance team must calculate quarterly sales tax using fixed percentages defined by law.
The marketing team wants to personalize product recommendations on an e-commerce site.
The supply chain team wants to predict next month's inventory requirements using past sales data.
Answer Description
AI/ML adds value when patterns must be learned from historical data and can change over time, such as fraud detection, demand forecasting, or personalized recommendations. Calculating sales tax, however, follows fixed percentages defined by law. Because the rules are explicit, infrequently updated, and produce a single deterministic answer, a simple rule-based program is cheaper, easier to maintain, and guarantees the exact outcome required for regulatory compliance. Therefore, an AI/ML approach would be unnecessary and unjustified for this task, whereas the other scenarios inherently involve prediction under uncertainty where ML can provide clear benefits.
Ask Bash
Bash is our AI bot, trained to help you pass your exam. AI Generated Content may display inaccurate information, always double-check anything important.
Why is AI/ML not suited for calculating quarterly sales tax?
What factors make AI/ML appropriate for tasks like fraud detection?
How does AI/ML help in predicting inventory requirements?
In the ML development lifecycle, a data scientist wants to automatically search different learning rates, batch sizes, and tree depths to find the most accurate model before deployment. Which Amazon SageMaker feature directly supports this stage?
Amazon SageMaker Automatic Model Tuning
Amazon SageMaker Batch Transform
Amazon SageMaker Model Package
Amazon SageMaker Ground Truth
Answer Description
The automatic search of multiple hyperparameter combinations corresponds to the hyperparameter tuning stage of the ML pipeline. Amazon SageMaker Automatic Model Tuning (sometimes called hyperparameter tuning jobs) runs many training jobs in parallel, evaluates them against a chosen metric, and returns the model that performs best. SageMaker Ground Truth is for data labeling, Batch Transform is for offline inference, and a Model Package is a container format used for sharing or deploying an already trained model, so none of these handle hyperparameter optimization.
Ask Bash
Bash is our AI bot, trained to help you pass your exam. AI Generated Content may display inaccurate information, always double-check anything important.
What are hyperparameters in machine learning?
How does Amazon SageMaker Automatic Model Tuning work?
What is the difference between Automatic Model Tuning and Ground Truth in SageMaker?
Within the AWS Cloud learning path, which of the following statements best defines artificial intelligence (AI) in general computing?
Computer systems designed to carry out tasks that typically require human intelligence, such as understanding language or recognizing images.
A specialized branch of machine learning that relies on multi-layer neural networks to process very large datasets.
A cloud-based database service optimized for storing structured, text, and image data for analytics workloads.
Algorithms that automatically adjust their weights through back-propagation to fit labeled datasets.
Answer Description
Artificial intelligence is the overarching discipline focused on building computer systems that can perform tasks normally requiring human intellect, such as perception, reasoning, and decision-making. This broad goal may be achieved with many techniques, including-but not limited to-machine learning and deep learning. Choice describing AI as the ability to learn from data using multi-layer neural networks is actually deep learning, a subset of machine learning. The option stating that AI automatically tunes weights with back-propagation also refers to machine or deep learning specifics, not AI as a whole. A database technology that stores text and image data is unrelated to the definition of AI.
Ask Bash
Bash is our AI bot, trained to help you pass your exam. AI Generated Content may display inaccurate information, always double-check anything important.
How does artificial intelligence differ from machine learning and deep learning?
What are some common tasks artificial intelligence can perform?
What is the role of perception, reasoning, and decision-making in AI?
A video-streaming company wants to automatically propose relevant movies to each user, based on the titles they have previously watched and rated. Which type of AI application best suits this requirement?
A fraud detection model that flags unusual account activity
A recommendation system that generates personalized content suggestions
A speech recognition service that transcribes dialogues in real time
An image classification system that labels movie poster images
Answer Description
Recommender systems analyze users' historical interactions-such as viewing and rating behaviour-to predict and surface items each user is likely to enjoy. This personalization objective is not addressed by fraud detection (focused on anomalous activity), speech recognition (turning audio into text), or image classification (labelling visual content).
Ask Bash
Bash is our AI bot, trained to help you pass your exam. AI Generated Content may display inaccurate information, always double-check anything important.
What is a recommendation system, and how does it work?
What is the difference between collaborative filtering and content-based filtering?
How does a recommendation system handle new users or items (cold-start problem)?
A retail startup wants to produce high-quality product photos simply by entering short text prompts that describe each desired scene. Which generative AI use case best fits this need?
Translation
Recommendation engine
Summarization
Image generation
Answer Description
The scenario describes creating new images from textual descriptions, which is characteristic of the image generation use case. Summarization condenses text, translation converts language, and recommendation engines suggest items based on user data; none of these tasks involve synthesizing new visual content from prompts.
Ask Bash
Bash is our AI bot, trained to help you pass your exam. AI Generated Content may display inaccurate information, always double-check anything important.
How does image generation work in generative AI?
What are some real-world applications of image generation?
How does text input turn into an image in tools like DALL·E?
A retailer launches a generative-AI chatbot that suggests personalized add-on items during checkout. The company wants one metric that shows whether the bot is boosting revenue from ongoing customer relationships, not just single orders. Which metric should they track?
Customer acquisition cost (CAC)
Average order value (AOV)
Net promoter score (NPS)
Customer lifetime value (CLV)
Answer Description
Customer lifetime value (CLV) captures the total revenue a business expects from an individual customer across the entire relationship and therefore shows whether generative-AI recommendations are increasing long-term spending. Average order value (AOV) looks only at a single transaction, customer acquisition cost (CAC) measures how much it costs to obtain a new buyer, and net promoter score (NPS) reflects satisfaction rather than revenue. None of these alternatives directly indicate the cumulative revenue impact per customer that CLV provides.
Ask Bash
Bash is our AI bot, trained to help you pass your exam. AI Generated Content may display inaccurate information, always double-check anything important.
What is Customer Lifetime Value (CLV)?
Why isn’t Average Order Value (AOV) the correct metric in this scenario?
How does a generative-AI chatbot influence Customer Lifetime Value (CLV)?
Which risk of prompt engineering occurs when attackers insert malicious instructions into a model's context, such as data retrieved by RAG, so future prompts are influenced to produce incorrect or harmful answers?
Prompt poisoning
Prompt hijacking
Model drift
Jailbreaking
Answer Description
The described scenario is prompt poisoning. In a prompt-poisoning attack, adversaries inject harmful or misleading instructions into the data a foundation model later consumes (for example, documents indexed for RAG). When the model retrieves this tainted content, the malicious instructions become part of the prompt, leading the model to generate compromised outputs. Prompt hijacking instead refers to an attacker taking over an in-progress conversation, while jailbreaking tries to bypass a model's safety filters. Model drift is an unrelated performance change that happens gradually as data or requirements evolve, not from deliberate malicious instructions.
Ask Bash
Bash is our AI bot, trained to help you pass your exam. AI Generated Content may display inaccurate information, always double-check anything important.
What is RAG in the context of prompt poisoning?
How does prompt poisoning differ from prompt hijacking?
What safeguards can mitigate the risk of prompt poisoning?
A startup needs a fully managed, MongoDB-compatible database that can store text embeddings and support similarity searches with a native vector index. Which AWS service should they choose?
Amazon DynamoDB
Amazon DocumentDB
Amazon ElastiCache for Redis
Amazon Neptune
Answer Description
Amazon DocumentDB offers native vector search indexes, letting teams persist embeddings and run similarity queries while using familiar MongoDB drivers and tools. Amazon Neptune is a graph database. Amazon DynamoDB is a key-value/NoSQL service that does not have native vector search capabilities but can integrate with other services like Amazon OpenSearch Service to provide it. Amazon ElastiCache for Redis provides in-memory caching rather than being primarily a managed document database.
Ask Bash
Bash is our AI bot, trained to help you pass your exam. AI Generated Content may display inaccurate information, always double-check anything important.
What is a native vector index in databases?
How does Amazon DocumentDB support MongoDB compatibility?
What are text embeddings, and why are they useful in similarity searches?
A startup wants to train a machine learning model that detects customer sentiment in live chat transcripts. Which type of training data should the team collect and prepare for this use case?
Labeled text documents annotated with sentiment categories
Unlabeled time-series sensor readings collected at regular intervals
Labeled JPEG images showing different facial expressions
Numeric tabular data of customer purchase amounts with column headers
Answer Description
Detecting sentiment in chat transcripts is a natural language processing task, so the model must learn from examples of written language. Labeled text documents that include sentiment tags (for example, positive, neutral, or negative) provide the necessary input-output pairs for supervised learning. Time-series sensor data, numeric tabular records, and labeled images do not contain the textual information required to teach a model about language sentiment, making them unsuitable for this scenario.
Ask Bash
Bash is our AI bot, trained to help you pass your exam. AI Generated Content may display inaccurate information, always double-check anything important.
Why is sentiment analysis considered a natural language processing task?
Why are labeled text documents essential for supervised machine learning in this use case?
Why are other types of data, such as images or tabular data, not suitable for this problem?
Within artificial intelligence terminology, which statement best describes the core idea of machine learning (ML)?
Specialized hardware acceleration used to run deep neural networks more efficiently.
A technique that allows computers to automatically learn patterns from data and improve without explicit programming of each rule.
The process of manually labeling datasets so that an algorithm can be trained later.
A programming style where every decision rule is explicitly encoded by developers.
Answer Description
Machine learning is defined as the field of study that enables computers to improve their performance at a task by learning patterns from data, rather than relying on explicitly coded instructions. This distinguishes ML from approaches where developers manually encode every decision rule. Manually labeling data is often part of preparing supervised ML, but it is not the definition of ML itself. Likewise, ML does not specifically refer to specialized hardware acceleration.
Ask Bash
Bash is our AI bot, trained to help you pass your exam. AI Generated Content may display inaccurate information, always double-check anything important.
What is supervised learning in ML?
Why are patterns important in machine learning?
What is the role of datasets in machine learning?
An ecommerce startup is building a customer-support chatbot on Amazon Bedrock. The bot must answer questions by using the company's product manuals stored in Amazon S3. To follow the Retrieval Augmented Generation (RAG) pattern, which additional component should the team add to the workflow?
Configure an Amazon S3 Access Point so the model can read all manuals directly during inference.
Run an Amazon SageMaker training job to fine-tune the foundation model on the S3 documents.
Add an AWS Lambda function that sets the model's temperature to 0 for deterministic answers.
Create a vector store such as an Amazon OpenSearch Service index that holds embeddings and returns relevant passages before invoking the model.
Answer Description
RAG solutions first retrieve content relevant to the user's question and then pass that content to a foundation model for generation. Amazon Bedrock supports this by connecting to a vector database where embeddings of the source documents are stored. When a query arrives, the vector store performs similarity search, returns the most relevant passages, and those passages are appended to the prompt that Bedrock sends to the model. Fine-tuning the model, granting the model blanket S3 access, or only adjusting inference parameters does not provide retrieval capabilities. Therefore, adding a vector store such as an Amazon OpenSearch Service index for embeddings is required to implement RAG.
Ask Bash
Bash is our AI bot, trained to help you pass your exam. AI Generated Content may display inaccurate information, always double-check anything important.
What is Retrieval Augmented Generation (RAG)?
What is a vector store and how does it support RAG workflows?
How does Amazon Bedrock integrate with a vector store for RAG solutions?
While estimating generative AI costs, a team learns that Amazon Bedrock bills for model inference using a token-based pricing model. Which factor will have the greatest direct impact on those token charges?
The total count of IAM users authorized to invoke the model endpoint
The total number of input and output tokens processed in each model invocation
The number of AWS Regions where the Bedrock API is called
The amount of data stored in Amazon S3 buckets that the application reads
Answer Description
Amazon Bedrock charges by the number of tokens processed by the model during inference. This includes both the input tokens contained in the prompt and the output tokens returned by the model. Storage, regional deployment, and IAM user counts incur separate charges (such as S3 storage or data transfer) or no charge at all, but they do not change the token fee that Bedrock applies for each model invocation.
Ask Bash
Bash is our AI bot, trained to help you pass your exam. AI Generated Content may display inaccurate information, always double-check anything important.
What are input and output tokens in generative AI?
How does Amazon Bedrock's token-based pricing model work?
What other factors besides token charges might impact generative AI costs in AWS?
A team will fine-tune a foundation language model for a retailer's customer-service chatbot. Which data selection practice best prepares the dataset for responsible fine-tuning?
Collect recent chat transcripts, redact personal identifiers, and remove conversations that violate company policy.
Keep all historical chat logs unchanged to maximize the size of the training set.
Use only publicly available Wikipedia articles about retail terminology.
Generate entirely synthetic conversations with the current model and include them without manual review.
Answer Description
Removing personal identifiers and filtering out policy-violating or off-topic conversations creates a domain-specific, compliant corpus. This improves the model's ability to answer retail queries accurately while reducing privacy risks and the chance that it learns undesirable behavior. Simply retaining all raw logs, relying on unrelated public articles, or using synthetic data without review either adds irrelevant content, preserves sensitive data, or risks compounding model errors, all of which hurt fine-tuning quality.
Ask Bash
Bash is our AI bot, trained to help you pass your exam. AI Generated Content may display inaccurate information, always double-check anything important.
What is fine-tuning in machine learning?
Why is it important to redact personal identifiers in datasets for AI training?
What risks are associated with training AI models on synthetic or unrelated datasets?
A startup is testing a new text-generation foundation model to confirm that its replies sound polite and helpful. Which evaluation method is most appropriate for measuring these subjective qualities?
Compute the BLEU score of each response compared with a reference answer.
Record the model's average inference latency in milliseconds.
Have human reviewers rate the model's responses against a qualitative rubric.
Count the total number of output tokens produced per response.
Answer Description
Politeness and helpfulness are subjective, qualitative attributes that automated metrics such as BLEU, latency, or token counts cannot reliably capture. Instead, organizations typically rely on human evaluation: reviewers read sample responses and rate them against a rubric that defines desired qualities (for example, politeness levels or usefulness of information). Human judgment can account for nuance, tone, and context in ways automated statistics cannot. BLEU focuses on n-gram overlap with reference text, latency measures performance speed, and token counts track output length-none of which directly indicates whether answers are polite or helpful.
Ask Bash
Bash is our AI bot, trained to help you pass your exam. AI Generated Content may display inaccurate information, always double-check anything important.
Why can't automated metrics like BLEU score capture subjective qualities like politeness and helpfulness?
What is a qualitative rubric, and how is it used for evaluating AI responses?
How does human evaluation account for nuances in context and tone compared to automated metrics?
A small startup plans to add an AI-powered brainstorming tool to its web app. The team would rather invoke a managed foundation model with plain-language prompts than build, train, and host its own machine-learning pipeline. Which generative AI benefit does this scenario highlight?
Regional coverage
Simplicity
Interpretability
Redundancy
Answer Description
Using a fully managed foundation model through simple prompt calls removes the need for data scientists, infrastructure setup, and model training. This ease of adoption demonstrates the simplicity benefit of generative AI. Redundancy and regional coverage are infrastructure considerations, while interpretability refers to model transparency-not to how easy a model is to consume.
Ask Bash
Bash is our AI bot, trained to help you pass your exam. AI Generated Content may display inaccurate information, always double-check anything important.
What are foundation models in generative AI?
Why is simplicity important when using a managed foundation model?
What is the difference between managed foundation models and custom-trained models?
A developer must prevent an LLM from returning personal data in its responses. Which prompt-engineering practice provides this guardrail at the prompt level?
Insert a negative prompt that instructs the model to exclude any personal data.
Increase the maximum output tokens so responses can fully elaborate.
Provide chain-of-thought examples so the model explains its reasoning step-by-step.
Set the temperature parameter close to 0 to make outputs more deterministic.
Answer Description
Adding a negative instruction (negative prompt) inside the prompt explicitly tells the model what it must not include, such as personal data. This operates as a guardrail before generation begins. Simply lowering temperature, using chain-of-thought examples, or enlarging the token limit affects creativity or verbosity but does not directly block disallowed content.
Ask Bash
Bash is our AI bot, trained to help you pass your exam. AI Generated Content may display inaccurate information, always double-check anything important.
What is a negative prompt in AI?
How does temperature affect model outputs in AI?
What are chain-of-thought prompts, and how do they work?
Gnarly!
Looks like that's it! You can go back and review your answers or click the button below to grade your test.