AWS Certified AI Practitioner (AIF-C01)

- 324 exam-style questions
- Detailed explanations and references
- Simulation and custom modes
- Custom exam settings to drill down into specific topics
- 180-day access period
- Pass or money back guarantee
What is in the package
CertVista AIF-C01 content, tone, and depth precisely mirror the questions in the AWS Certified AI Practitioner (AIF-C01) exam. Our comprehensive materials include detailed explanations and practical exam-taker tips, thoroughly referencing AWS documentation to prepare you for all domain areas of the AI Practitioner certification.
Please consider this course the final pit-stop so you can cross the winning line with absolute confidence and get AWS AI Certified! Trust our process; you are in good hands.
Complete AIF-C01 domains coverage
Our practice exams fully align with the official AWS courseware and the Certified AI Practitioner exam objectives.
1. Fundamentals of AI and ML
Covers core AI/ML concepts, terminology, and practical applications. This domain focuses on understanding basic AI concepts, identifying suitable use cases, and comprehending the ML development lifecycle. It includes knowledge of different learning types, data types, and AWS's managed AI services. The domain also emphasizes understanding MLOps concepts and model performance evaluation.
2. Fundamentals of Generative AI
Focuses on generative AI essentials, including foundational concepts like tokens, embeddings, and prompt engineering. This domain explores use cases for generative AI, its lifecycle, capabilities, and limitations. It also covers AWS's infrastructure and technologies for building generative AI applications, including cost considerations and service selection.
3. Applications of Foundation Models
Explores the practical aspects of working with foundation models, including design considerations, prompt engineering techniques, and model customization. This domain covers model selection criteria, Retrieval Augmented Generation (RAG), training processes, and performance evaluation methods. It emphasizes understanding both technical implementation and business value assessment.
4: Guidelines for Responsible AI
Addresses the ethical and responsible development of AI systems. This domain covers important aspects like bias, fairness, inclusivity, and transparency in AI systems. It includes understanding tools for responsible AI development, recognizing legal risks, and implementing practices for transparent and explainable AI models.
5. Security, Compliance, and Governance for AI Solutions
Focuses on securing AI systems and ensuring regulatory compliance. This domain covers AWS security services, data governance strategies, and compliance standards specific to AI systems. It includes understanding security best practices, privacy considerations, and governance protocols for AI implementations.
CertVista's AI Practitioner question bank contains hundreds of exam-style questions that accurately replicate the certification exam environment. Practice with diverse question types, including multiple-choice, multiple-response, and scenario-based questions focused on real-world AI implementation challenges. CertVista exam engine will familiarize you with the real exam environment so you can confidently approach your certification.

Each CertVista question comes with a detailed explanation and references. The explanation outlines the underlying AI principles, references official AWS documentation, and clarifies common misconceptions. You'll learn why the correct answer satisfies the scenario presented in the question and why the other options do not.

CertVista offers two effective study modes: Custom Mode is for focused practice on specific AWS domains and is perfect for strengthening knowledge in targeted areas. Simulation Mode replicates the 90-minute exam environment with authentic time pressure and question distribution, building confidence and stamina.

The CertVista analytics dashboard helps you gain clear insights into your AWS exam preparation. You can monitor your performance across all exam domains and identify knowledge gaps. This will help you create an efficient study strategy and know when you're ready for certification.

What's in the AIF-C01 exam
The AWS Certified AI Practitioner exam validates your understanding of artificial intelligence and machine learning fundamentals within the AWS ecosystem. This foundational certification demonstrates your ability to identify appropriate AWS AI services for various business scenarios, understand generative AI applications, and implement AI solutions responsibly and securely.
To pass the AWS AI Practitioner exam, candidates must demonstrate proficiency in several key areas. You'll need to explain core AI and ML concepts, including the differences between various types of machine learning and their applications. The exam tests your ability to understand and articulate the capabilities and limitations of AWS AI services, particularly in the context of business solutions.
You'll be expected to show competency in generative AI concepts, including foundation models, prompt engineering, and AWS's generative AI infrastructure. The exam also emphasizes responsible AI practices, requiring you to understand bias, fairness, and transparency in AI systems. Additionally, you'll need to demonstrate knowledge of security and compliance considerations specific to AI implementations on AWS.
Exam Format and Scoring
- Total Questions: 65 questions
- Time Limit: 100 minutes
- Question Types:
- Multiple choice: One correct answer from four options
- Multiple response: Two or more correct answers from five or more options
- Exam Cost: USD 100
- Passing Score: 750 out of 1000
- Language: Available in English
- Validity: Three years from the date of certification
The exam includes unscored questions that are used for statistical purposes. These questions are indistinguishable from scored questions and are randomly placed throughout the exam. Any unanswered questions are marked as incorrect.
Upon completion, you'll receive a detailed score report on your performance across each domain. While you'll see your performance in individual domains, the certification is awarded based on your overall score, not domain-specific performance. This means you don't need to achieve a minimum score in each domain – only your total score needs to meet or exceed the passing threshold.
The exam tests theoretical knowledge and practical understanding, with questions ranging from basic concept identification to complex scenario-based problem-solving. Many questions will present real-world situations where you'll need to identify the most appropriate AWS AI services or solutions for specific business challenges.
Remember, the exam is regularly updated to reflect the latest AWS AI services and best practices, so staying current with AWS's AI/ML offerings and industry developments is important during your preparation.
Sample AIF-C01 questions
Get a taste of the AWS Certified AI Practitioner exam with our carefully curated sample questions below. These questions mirror the actual exam's style, complexity, and subject matter, giving you a realistic preview of what to expect. Each question comes with comprehensive explanations, relevant AWS documentation references, and valuable test-taking strategies from our expert instructors.
While these sample questions provide excellent study material, we encourage you to try our free demo for the complete exam preparation experience. The demo features our state-of-the-art test engine that simulates the real exam environment, helping you build confidence and familiarity with the exam format. You'll experience timed testing, question marking, and review capabilities – just like the actual AWS certification exam.
Which SageMaker service helps split data into training, testing, and validation sets?
Amazon SageMaker Feature Store
Amazon SageMaker Clarify
Amazon SageMaker Ground Truth
Amazon SageMaker Data Wrangler
Preparing data for machine learning often involves splitting the dataset into subsets for training, validation, and testing. This ensures that the model is trained on one portion of the data, tuned on another, and finally evaluated on unseen data.
Amazon SageMaker Data Wrangler is designed to simplify the process of data preparation for ML. It provides a visual interface and a comprehensive set of built-in data transformations to help data scientists and engineers aggregate, prepare, and transform data. Among its capabilities, Data Wrangler allows users to apply various transformations, including those needed to split a dataset into training, validation, and testing sets based on specified criteria or proportions.
Amazon SageMaker Feature Store is a repository for storing, retrieving, and managing ML features, but it doesn't perform the splitting operation itself.
Amazon SageMaker Clarify focuses on detecting potential bias in data and explaining model predictions, not on splitting datasets for training.
Amazon SageMaker Ground Truth is a data labeling service used to create labeled datasets, which are often the input for the preparation phase, but it doesn't perform the train/test/validation split.
Think of Data Wrangler as the primary SageMaker tool for preparing raw data before training. This includes cleaning, transforming, and structuring data, which often involves splitting it. Associate Feature Store with managing features, Clarify with bias and explainability, and Ground Truth with labeling.
Your company wants to use machine learning to better understand customer and sales patterns using unlabeled data.
Which two methods would help you discover patterns automatically? (Select two.)
Clustering
Dimensionality reduction
Sentiment analysis
Neural network
Decision tree
To understand customer and sales patterns using unlabeled data, your company should employ unsupervised learning methods that can automatically discover inherent structures and relationships within the data.
Clustering is an unsupervised learning technique ideal for this scenario. It works by grouping data points—in this case, customers or sales transactions—into clusters based on their similarities. For instance, clustering could reveal distinct customer segments based on purchasing behavior (e.g., high-value customers, infrequent buyers, customers who prefer specific product categories) or identify common patterns in sales data (e.g., product bundles frequently bought together). These discovered clusters represent patterns that were not explicitly defined beforehand.
Dimensionality reduction is another valuable unsupervised learning method for discovering patterns. It aims to reduce the number of features (variables) in a dataset while retaining essential information. By transforming high-dimensional data (e.g., customer data with many attributes) into a lower-dimensional space, dimensionality reduction can help uncover underlying patterns and relationships that might be obscured in the original data. For example, it could reveal the most influential factors driving customer behavior or simplify complex sales data to make trends more apparent.
Sentiment analysis, while useful for understanding customer opinions, is typically a supervised learning task that requires labeled text data (e.g., reviews labeled as positive, negative, or neutral) to train a model.
A neural network is a type of model architecture that can be used for various tasks, including supervised and unsupervised learning, but it's not a specific method for pattern discovery in unlabeled data without further context (like an autoencoder for dimensionality reduction).
A decision tree is a supervised learning algorithm used for classification or regression tasks, which requires labeled data to learn the decision rules.
Remember that when dealing with unlabeled data and the goal is to discover hidden patterns or structures automatically, unsupervised learning techniques like clustering and dimensionality reduction are the primary methods to consider.
References:
- What is Unsupervised Learning? - AWS
- Clustering data - Amazon SageMaker
- Dimensionality Reduction - Amazon SageMaker (Context for PCA, a common dimensionality reduction technique)
A company needs to implement an AI solution that can convert natural language input into SQL queries for their large-scale database analysis. The solution should be user-friendly for employees with limited technical expertise.
Which AI model would be most appropriate for this use case?
Generative pre-trained transformers (GPT)
Residual neural network
Support vector machine
WaveNet
Generative pre-trained transformers (GPT) is the most suitable solution for this scenario. GPT models excel at understanding and processing natural language input, making them ideal for converting plain English requests into structured SQL queries. They can understand context and intent behind user queries, which is crucial for employees with minimal technical experience.
The practical effectiveness of GPT for this use case has been demonstrated by major companies like Uber, which have successfully implemented GPT-based solutions for SQL query generation in enterprise environments. GPT models have consistently shown superior performance in text-to-SQL tasks compared to other AI models. Furthermore, GPT can handle complex database schemas and generate accurate SQL queries for large-scale data analysis. The model can be fine-tuned to understand specific business contexts and database structures, making it highly adaptable to different enterprise needs.
The other options are not suitable for this specific use case. Residual Neural Network, while powerful for image processing and deep learning tasks, is not specifically designed for natural language understanding and SQL generation. Support Vector Machine is a traditional machine learning algorithm better suited for classification and regression tasks, not complex language processing and query generation. WaveNet is a deep neural network primarily designed for audio generation and speech synthesis, making it inappropriate for text-to-SQL conversion.
When evaluating AI solutions for natural language processing tasks, particularly those involving text transformation or generation, GPT models are often the strongest candidates due to their advanced language understanding capabilities.
A company needs an AI assistant that can help employees by answering questions, creating summaries, generating content, and securely working with internal company data.
Which Amazon Q solution is designed for this kind of general business use across an organization?
Amazon Q Business
Amazon Q in Connect
Amazon Q Developer
Amazon Q in QuickSight
Amazon Q Business is specifically designed as an AI assistant for work. It can connect to various enterprise data sources, understand company-specific information (like organizational structure, product names, and internal jargon), and help employees with tasks such as answering questions based on internal knowledge bases, summarizing documents, drafting emails, generating reports, and more, all while respecting existing security permissions.
Amazon Q in Connect is tailored for customer service agents using Amazon Connect, helping them with real-time responses and actions during customer interactions.
Amazon Q Developer is focused on assisting developers throughout the software development lifecycle, including code generation, debugging, and optimization, integrated into IDEs and AWS environments.
Amazon Q in QuickSight enhances the business intelligence experience within Amazon QuickSight, allowing users to build dashboards and gain insights using natural language queries.
Given the requirements for a general-purpose assistant for employees across the organization to work securely with internal data for various tasks like Q&A, summarization, and content generation, Amazon Q Business is the appropriate solution.
Amazon Q has several variations tailored to specific use cases. Pay attention to the context described in the question. If it's about general employee productivity and accessing internal business data, think Amazon Q Business. If it's about coding, think Amazon Q Developer. If it's about contact centers, think Amazon Q in Connect. If it's about BI dashboards, think Amazon Q in QuickSight.
A media company is deploying machine learning models with Amazon SageMaker to provide personalized content recommendations. They have intermittent workloads and don't want to manage the underlying infrastructure. They are looking for a deployment model that offers cost savings through cold starts.
Which deployment model should they choose?
Asynchronous Inference
Serverless Inference
Real-time hosting services
Batch Transform
For a media company with intermittent workloads that wants to provide personalized content recommendations using Amazon SageMaker, and is looking for a deployment model that offers cost savings through mechanisms like cold starts without requiring infrastructure management, Serverless Inference is the most suitable choice.
Amazon SageMaker Serverless Inference is designed specifically for workloads that are intermittent or have infrequent traffic patterns. It automatically provisions, scales, and turns off compute resources based on the volume of inference requests. When there are no requests, it can scale down to zero, meaning the company doesn't pay for idle compute capacity. When a new request comes in after a period of inactivity, a 'cold start' might occur as resources are provisioned, but this is the trade-off for significant cost savings during idle times. Because it's serverless, AWS manages the underlying infrastructure, so the company's development team doesn't need to worry about provisioning or managing servers.
Asynchronous Inference is suitable for large payloads and long processing times, where clients don't need an immediate response. While it can handle intermittent traffic, its cost model and primary design aren't focused on the 'scale to zero' and cold-start-for-cost-saving dynamic in the same way Serverless Inference is. Real-time hosting services (persistent SageMaker endpoints) involve provisioned instances that run continuously, incurring costs even during idle periods, which is not ideal for highly intermittent workloads where cost optimization is key. Batch Transform is used for offline processing of large datasets and is not suitable for providing real-time or near real-time personalized content recommendations.
For the AIF-C01 exam, remember that 'serverless' in the context of SageMaker Inference implies automatic scaling, pay-per-use (even scaling to zero), and no infrastructure management. The mention of 'cold starts' as a factor in cost savings strongly points towards a serverless paradigm where resources are spun up on demand after being idle.
A software development company is building generative AI solutions and needs to understand the distinctions between model inference and model evaluation.
Which option best summarizes these differences in the context of generative AI?
Model inference is the process of evaluating and comparing model outputs to determine the model that is best suited for a use case.
Model evaluation is the process of a model generating an output (response) from a given input (prompt).
Both model inference and model evaluation refer to the process of evaluating and comparing model outputs to determine the model that is best suited for a use case
Model evaluation is the process of evaluating and comparing model outputs to determine the model that is best suited for a use case.
Model inference is the process of a model generating an output (response) from a given input (prompt).
Both model inference and model evaluation refer to the process of a model generating an output from a given input.
Understanding the distinction between model inference and model evaluation is crucial when working with generative AI models.
Model Inference: This is the process where a trained foundation model (FM) takes an input, typically called a prompt, and generates an output or response. For example, providing a text prompt like "Write a short story about a friendly robot" to a large language model (LLM) and receiving the generated story is an inference operation. It's the act of using the model to produce results.
Model Evaluation: This involves assessing the performance and quality of a model's outputs to determine its suitability for a specific task or use case. In generative AI, evaluation can be complex and often involves both quantitative metrics (if available) and qualitative assessments (human judgment). This process might compare outputs from different models based on criteria like relevance, coherence, accuracy, safety, or adherence to specific instructions. The goal is to understand how well the model performs and select the best one for the intended application.
Therefore, model evaluation is about judging the model's output quality and suitability, while model inference is the process of the model generating that output.
Remember the flow: You first perform inference (generate output using the model), and then you evaluate that output (assess its quality/suitability). Think of inference as the action of the model and evaluation as the judgment of that action's result.
A company wants to use large language models (LLMs) with Amazon Bedrock to develop a chat interface for the company's product manuals. The manuals are stored as PDF files.
Which solution meets these requirements most cost-effectively?
Use prompt engineering to add one PDF file as context to the user prompt when the prompt is submitted to Amazon Bedrock.
Use prompt engineering to add all the PDF files as context to the user prompt when the prompt is submitted to Amazon Bedrock.
Use all the PDF documents to fine-tune a model with Amazon Bedrock. Use the fine-tuned model to process user prompts.
Upload PDF documents to an Amazon Bedrock knowledge base. Use the knowledge base to provide context when users submit prompts to Amazon Bedrock.
Using Amazon Bedrock knowledge base is the most cost-effective solution for this use case. Knowledge bases implement a managed Retrieval Augmented Generation (RAG) architecture that efficiently retrieves relevant information from the uploaded documents when needed. This approach optimizes both performance and cost by only retrieving and using relevant portions of the documents for each query.
The other approaches have significant drawbacks:
Adding a single PDF file as context through prompt engineering would limit the chatbot's ability to access information from other manuals, requiring multiple queries and increasing costs. This would also provide incomplete responses if the information spans multiple manuals.
Including all PDF files as context in every prompt would be highly inefficient and expensive. This approach would unnecessarily increase token usage and processing costs for each query, even when only a small portion of the documentation is relevant.
Fine-tuning the model with all PDF documents would be the most expensive option. It requires significant computational resources and typically takes longer compared to other approaches. Additionally, updating the model with new or modified documentation would require repeated fine-tuning, increasing costs further.
When evaluating solutions involving document processing with LLMs, consider both the immediate implementation costs and long-term operational efficiency. Knowledge bases often provide the best balance between functionality and cost-effectiveness for document-heavy applications.
A company is using a pre-trained large language model (LLM) to build a chatbot for product recommendations. The company needs the LLM outputs to be short and written in a specific language.
Which solution will align the LLM response quality with the company's expectations?
Adjust the prompt.
Choose an LLM of a different size.
Increase the temperature.
Increase the Top K
value.
The most direct and common method to guide the behavior of a pre-trained LLM without modifying the model itself is through prompt engineering.
Adjusting the prompt involves carefully crafting the input text given to the LLM. By including explicit instructions within the prompt, the company can guide the model's response. For example, the prompt could be structured like: "You are a helpful chatbot recommending products. Respond in Spanish and keep your answer to one or two sentences. Recommend a product similar to [product name]
."
This prompt clearly states the desired constraints: the language (Spanish) and the length (one or two sentences). The LLM will use these instructions to generate an output that aligns with the company's expectations.
Choosing an LLM of a different size might impact overall capability or performance, but it doesn't offer fine-grained control over the output length or language for a specific request.
Increasing the temperature parameter makes the output more random and creative, often leading to longer and less predictable responses, which is contrary to the requirement for short, specific outputs.
Increasing the Top K
sampling parameter allows the model to consider more words at each step, increasing diversity but not specifically controlling length or language. It's a method for controlling randomness, similar to temperature.
Remember that prompt engineering is a powerful technique for controlling the output of generative AI models. When requirements involve specific output formats, styles, languages, or lengths, explicitly stating these constraints in the prompt is often the first and most effective approach.
A company wants to build an interactive application for children that generates new stories based on classic stories. The company wants to use Amazon Bedrock and needs to ensure that the results and topics are appropriate for children.
Which AWS service or feature will meet these requirements?
Amazon Rekognition
Amazon Bedrock playgrounds
Guardrails for Amazon Bedrock
Agents for Amazon Bedrock
Guardrails for Amazon Bedrock provide a mechanism to implement safeguards for generative AI applications. They allow users to define specific policies to control the interaction between users and foundation models (FMs). Key features relevant to this scenario include:
- Denied Topics: Administrators can define topics that the application should not engage with. For a children's application, this could include sensitive or adult themes.
- Content Filters: Guardrails offer configurable filters to detect and block harmful content across categories like hate speech, insults, sexual content, and violence, based on specified thresholds.
- Word Filters: Specific words can be blocked.
- PII Redaction: Personally Identifiable Information can be filtered out.
By configuring Guardrails, the company can enforce content policies consistently across the different FMs available through Bedrock, ensuring the generated stories align with the requirement of being appropriate for children.
Amazon Rekognition is an image and video analysis service and is not used for moderating text generated by Bedrock models.
Amazon Bedrock playgrounds are environments for experimenting with models, not for implementing runtime safety controls in a deployed application.
Agents for Amazon Bedrock enable the creation of applications that can perform tasks using APIs, but they do not inherently provide the content filtering capability required; Guardrails are used to provide safety for agents and direct model invocations.
Therefore, Guardrails for Amazon Bedrock is the appropriate feature to meet the requirement for content safety and appropriateness.
When questions involve ensuring safety, appropriateness, or filtering harmful content in generative AI applications built on AWS, immediately think of Guardrails for Amazon Bedrock. It's the purpose-built feature for implementing these types of policies.
A company wants to use a large language model (LLM) to develop a conversational agent. The company needs to prevent the LLM from being manipulated with common prompt engineering techniques to perform undesirable actions or expose sensitive information.
Which action will reduce these risks?
Create a prompt template that teaches the LLM to detect attack patterns.
Increase the temperature parameter on invocation requests to the LLM.
Avoid using LLMs that are not listed in Amazon SageMaker.
Decrease the number of input tokens on invocations of the LLM.
Creating a prompt template that teaches the LLM to detect attack patterns is the correct approach. This method provides a robust defense mechanism against prompt injection attacks. Well-designed prompt templates with security guardrails can detect and prevent various attack patterns, including prompted persona switches, attempts to extract prompt templates, and instructions to ignore security controls.
The template can incorporate specific guardrails that validate input, sanitize prompts, and establish secure communication parameters. This approach is particularly effective because it addresses security at the foundational level of the LLM's interaction with users, creating a first line of defense against malicious inputs.
As for the incorrect options, increasing the temperature parameter would actually make the model's outputs less predictable and potentially more vulnerable to manipulation. Limiting LLM selection to those listed in SageMaker doesn't address the core security concerns, as security depends on implementation rather than the model source. Reducing input tokens is an ineffective approach since sophisticated attacks can be executed with minimal tokens while this restriction would unnecessarily limit the model's legitimate functionality.
When evaluating LLM security measures, focus on solutions that directly address the specific security concern at the interaction level rather than general model parameters or arbitrary restrictions.
Frequently Asked Questions
The AWS Certified AI Practitioner is a foundational-level certification that validates your understanding of artificial intelligence, machine learning, and generative AI concepts within the AWS ecosystem. This certification demonstrates that you can effectively understand and work with AI/ML technologies, regardless of your job role.
The exam lasts 90 minutes and consists of 65 questions. To earn the credential, you must achieve a passing score of 700 out of 1000. The exam is in English and Japanese, and the registration fee is USD 100.
The AI Practitioner exam covers five essential domains: AI and ML fundamentals, generative AI basics, foundation model applications, responsible AI guidelines, and security/compliance considerations for AI solutions. You should understand how these concepts apply within the AWS ecosystem.
A solid foundation in AWS services is crucial. You should know core services like EC2, S3, Lambda, and SageMaker. Understanding the AWS shared responsibility model, IAM, global infrastructure, and service pricing models will also be beneficial.
You should understand fundamental AI terminology, the machine learning development lifecycle, and generative AI basics. Knowledge of foundation models, responsible AI practices, and security/compliance in AI systems is essential. The exam emphasizes practical applications rather than theoretical depth.
Focus on Amazon SageMaker and its ecosystem, Amazon Bedrock, and core AWS AI services like Comprehend, Transcribe, and Translate. Understanding security services and monitoring tools is also essential for the exam.
If you don't pass on your first attempt, you must wait 14 calendar days before trying again. While there's no limit on the number of attempts, each try requires paying the full registration fee. After passing, you must wait two years before retaking the same exam version.