Prepare for AWS Certified AI Practitioner (AIF-C01) with free sample questions, a full-length diagnostic, topic drills, timed practice, AI and machine-learning foundations, generative AI use cases, responsible-AI controls, governance scenarios, and detailed explanations in IT Mastery.
AIF-C01 is AWS’s AI Practitioner certification for candidates who need practical AI, generative AI, service-selection, and responsible-AI judgment without going deep into custom model building. If you are searching for AIF-C01 sample questions, a practice test, mock exam, or exam simulator, this is the main IT Mastery page to start on web and continue on iOS or Android with the same IT Mastery account.
Start a practice session for AWS Certified AI Practitioner (AIF-C01) below, or open the full app in a new tab. For the best experience, open the full app in a new tab and navigate with swipes/gestures or the mouse wheel—just like on your phone or tablet.
Open Full App in a New TabA small set of questions is available for free preview. Subscribers can unlock full access by signing in with the same app-family account they use on web and mobile.
Prefer to practice on your phone or tablet? Download the IT Mastery – AWS, Azure, GCP & CompTIA exam prep app for iOS or IT Mastery app on Google Play (Android) and use the same IT Mastery account across web and mobile.
Free diagnostic: Try the 65-question AWS AIF-C01 full-length practice exam before subscribing. Use it as one AI Practitioner baseline, then return to IT Mastery for timed mocks, topic drills, explanations, and the full AWS AI question bank.
AIF-C01 questions usually reward the option that applies the right AWS AI service, understands generative AI limits, and protects responsible-AI and governance expectations without overengineering the solution.
| Domain | Weight |
|---|---|
| Fundamentals of AI and ML | 20% |
| Fundamentals of Generative AI | 24% |
| Applications of Foundation Models | 28% |
| Guidelines for Responsible AI | 14% |
| Security, Compliance, and Governance for AI Solutions | 14% |
Use these filters when two AWS AI answers sound reasonable:
| Area | What strong readiness looks like |
|---|---|
| AI and ML fundamentals | You can distinguish supervised learning, unsupervised learning, deep learning, inference, training, evaluation, and common metrics. |
| Generative AI fundamentals | You can explain foundation models, prompts, embeddings, tokens, hallucinations, model modalities, and model selection. |
| Foundation model applications | You can identify when to use RAG, agents, knowledge bases, prompt templates, customization, and managed AWS AI services. |
| Responsible AI | You can apply fairness, privacy, transparency, human oversight, and safety controls to realistic business scenarios. |
| Security and governance | You can protect AI workloads with least privilege, encryption, monitoring, data controls, and compliant service configuration. |
| Day | Practice focus |
|---|---|
| 7 | Take the free full-length diagnostic and group misses by AI concept, AWS service, and governance topic. |
| 6 | Drill AI/ML fundamentals, model lifecycle, evaluation concepts, and data-quality reasoning. |
| 5 | Drill generative AI concepts, prompt patterns, embeddings, RAG, agents, and foundation-model applications. |
| 4 | Drill responsible AI, privacy, security, compliance, and governance scenarios. |
| 3 | Complete a timed mixed set and explain the use-case/service fit for every miss. |
| 2 | Review confusing service pairs and AI terms, especially Bedrock, SageMaker AI, Transcribe, Translate, Comprehend, and Rekognition. |
| 1 | Do a light confidence pass; avoid learning new AI terminology at the last minute. |
If you can score above roughly 75% on several unseen mixed attempts and explain why each answer fits the use case, service boundary, and governance constraint, further drilling may produce memorization more than readiness. At that point, schedule the exam and keep the final review light.
Use these child pages when you want focused IT Mastery practice before returning to mixed sets and timed mocks.
Need concept review first? Read the AWS AIF-C01 Cheat Sheet on Tech Exam Lexicon, then return here for timed mocks, topic drills, and full IT Mastery practice.
These are original IT Mastery practice questions aligned to AI and ML fundamentals, generative AI, foundation models, responsible AI, AWS AI services, and business-use-case decisions. They are not AWS exam questions and are not copied from any exam sponsor. Use them to check readiness here, then continue in IT Mastery with mixed sets, topic drills, and timed mocks.
Topic: Content Domain 2: Fundamentals of GenAI
Which option best describes a GenAI agent use case (for example, using Agents for Amazon Bedrock) rather than a simple Q&A assistant?
Best answer: C
Explanation: A GenAI agent is designed to go beyond answering questions by selecting tools and taking actions to achieve a user goal. The best example is a workflow where the model calls approved business APIs (for example, ticketing or payments) to complete a multi-step task on the user’s behalf.
The core concept is the difference between an assistant and an agent. A GenAI assistant primarily generates text responses (Q&A, summarization, drafting) and may use retrieval (RAG) to ground answers in enterprise content. A GenAI agent adds “action-taking” behavior: it can break a request into steps, decide which tools to use, and invoke APIs or workflows to complete tasks with guardrails and permissions. In AWS, this pattern is commonly implemented with Agents for Amazon Bedrock, where the model orchestrates tool calls to perform operations like checking order status, creating tickets, or processing returns. The key distinguishing feature is autonomous tool use to accomplish an outcome, not just producing an answer.
Topic: Content Domain 1: Fundamentals of AI and ML
A team is using Amazon SageMaker AI to build a binary classification model. The team is writing a standard operating procedure (SOP) that describes what happens during model training, hyperparameter tuning, and final model evaluation, along with basic governance controls.
Which TWO SOP statements are INCORRECT (anti-patterns or unsafe practices)? (Select TWO)
s3:* on the dataset bucket to avoid permission issuesBest answers: B, C
Explanation: Training fits model parameters (weights) using the training data, hyperparameter tuning selects configuration values using validation results, and evaluation measures the final model on an untouched test set. Good governance also requires traceability of experiments and least-privilege access to sensitive datasets. The incorrect statements either compromise evaluation integrity or weaken security controls.
In the ML lifecycle, model training updates the model’s learned parameters using the training dataset. Hyperparameter tuning runs repeated training jobs with different settings (for example, learning rate, regularization strength) and chooses the best configuration based on validation performance. Final model evaluation should be performed on a separate, held-out test set exactly once to estimate real-world performance.
From a governance perspective, you should preserve provenance (track datasets, hyperparameters, metrics, and model versions) and enforce least privilege for access to training and evaluation data. Using the test set during tuning leaks evaluation information into model selection and invalidates the trustworthiness of reported test metrics; overly broad permissions increase the chance of accidental or malicious data exposure or tampering.
Topic: Content Domain 1: Fundamentals of AI and ML
Which scenario is the best fit for using speech recognition to create a text transcript?
Best answer: C
Explanation: Speech recognition is used to convert spoken audio into written text. Creating transcripts from call recordings is a common application because the resulting text can be searched, summarized, and analyzed (for example, for sentiment or compliance checks). The other scenarios involve computer vision, anomaly detection on log data, or text translation rather than transcription.
Speech recognition (also called automatic speech recognition, ASR) converts human speech in an audio stream or recording into text. Transcription is valuable when you need a written record of spoken interactions, such as contact center calls, meetings, or videos, because text is easier to index, search, analyze for themes, and use for captions/subtitles. In AWS, services like Amazon Transcribe provide managed speech-to-text capabilities for these use cases. Tasks like object detection in video, anomaly detection in network logs, or language translation operate on different data modalities (images/video, structured logs, or existing text) and therefore are not speech recognition problems.
Key takeaway: choose speech recognition when the input is audio speech and the desired output is a text transcript.
Topic: Content Domain 2: Fundamentals of GenAI
Which TWO statements about choosing a foundation model’s modality are INCORRECT? (Select TWO)
Best answers: A, C
Explanation: Model modality must match the input and output types your use case requires. Text-only models cannot natively process audio or images, so those modalities require either multimodal models or pre-/post-processing with specialized services. The incorrect statements are the ones claiming text-only models can directly handle images or audio transcription.
A key model selection factor is modality: what the model can accept as input and produce as output. If your use case includes non-text content, you must choose an image/audio/multimodal-capable model, or convert the content into text before using a text-only model.
Always align modality to the data you actually have and the output you need.
Topic: Content Domain 3: Applications of Foundation Models
Which TWO statements accurately describe Amazon Bedrock Model Evaluation and when it is used? (Select TWO.)
Best answers: B, E
Explanation: Amazon Bedrock Model Evaluation is used to assess and compare foundation model performance for a specific use case before choosing a model. You provide prompts (and optionally reference answers) and select evaluation approaches/metrics to understand which model best meets your quality requirements.
Amazon Bedrock Model Evaluation is a feature for evaluating and comparing foundation models on your workload, typically during model selection and iteration (for example, before launching a summarization or Q&A application). At a high level, you run evaluation using your prompts/datasets and chosen evaluation approach to measure output quality and help decide which model is the best fit.
It is not a training or deployment feature: it does not fine-tune models as part of evaluation, it is not limited to models you trained yourself, and it does not automatically push a “winner” to production. The key takeaway is that Model Evaluation supports structured comparison of model quality for your specific task so you can pick an appropriate FM with evidence.
Topic: Content Domain 2: Fundamentals of GenAI
A company is building a retrieval-augmented generation (RAG) chatbot using Amazon Bedrock. The team generates embeddings for a user query and searches a vector index.
Exhibit: Vector search results (cosine similarity)
| Item | Type | Cosine similarity to query |
|---|---|---|
| chunk-17 | document chunk embedding | 0.82 |
| chunk-04 | document chunk embedding | 0.31 |
| chunk-09 | document chunk embedding | 0.76 |
Which option best interprets the exhibit in terms of vectors and vector similarity search?
Best answer: B
Explanation: An embedding represents text as a numeric vector in a high-dimensional space. Vector similarity search compares the query vector to stored vectors using a metric such as cosine similarity and returns the closest matches. Because the exhibit’s highest similarity is 0.82, that item is the best retrieval candidate.
Vectors (embeddings) are arrays of numbers that represent the meaning of text so that semantically similar text ends up “near” each other in vector space. Vector similarity search works by comparing a query embedding to stored embeddings with a distance/similarity metric (here, cosine similarity) and returning the nearest neighbors.
In the exhibit, the system should pick the document chunk with the largest cosine similarity value because that vector is closest to the query vector:
chunk-17 has 0.82chunk-09 has 0.76chunk-04 has 0.31The key takeaway is that the score is produced by comparing vectors, not by keyword matching or model retraining.
Topic: Content Domain 4: Guidelines for Responsible AI
A company is building a customer-support chatbot using an Amazon Bedrock foundation model. The company needs a managed control that can screen both user prompts and model responses to help prevent harmful content and to detect/redact sensitive data (for example, PII) without changing the underlying model.
Which solution meets this requirement?
bedrock:InvokeModel for the chatbot roleBest answer: D
Explanation: Amazon Bedrock Guardrails are designed to apply policy-based controls to both incoming prompts and outgoing model responses. They can filter categories of harmful content and detect or redact sensitive information such as PII. This directly supports responsible AI by constraining inputs and outputs without changing the foundation model.
The discriminating requirement is a managed mechanism that enforces safety constraints on both the user input and the model output. Amazon Bedrock Guardrails provide centralized, configurable policies (such as content filters, topic restrictions, and sensitive-information/PII handling) that are evaluated on prompts and on generated responses.
Encryption and IAM address security and access control, but they do not evaluate text for harmful or sensitive content. Prompt-only instructions can reduce risk, but they are not reliable enforcement because the model can still produce disallowed content under some inputs. The key takeaway is to use Guardrails when you need policy enforcement over model inputs and outputs.
Topic: Content Domain 3: Applications of Foundation Models
A company uses an Amazon Bedrock text model to generate summaries of customer support tickets. Different agents type their own prompts, and the downstream application fails when the model output is not consistent.
Exhibit: Current workflow
1 User prompt varies (no template)
2 Example output A: "Summary:... Next steps:..."
3 Example output B: "- Issue...\n- Action..."
4 Example output C: "{ 'sum':..., 'priority': 'high' }"
5 Requirement: downstream app expects JSON keys:
6 summary, priority, customer_impact, next_action
Based on the exhibit, what is the best next step to standardize outputs across users and workflows?
Best answer: B
Explanation: The exhibit shows inconsistent formats across agents (lines 1-4) while the downstream app requires specific JSON keys (lines 5-6). The most direct way to standardize output is to use a shared prompt template that fixes the instructions and output structure (for example, “return only JSON with these keys”) and parameterizes only the ticket content.
Prompt templates standardize results by separating variable user inputs from the fixed instructions and response format the model must follow. In the exhibit, outputs vary widely because agents use different prompts (line 1), producing different structures (lines 2-4). Meanwhile, the consuming application expects a specific schema with four JSON keys (lines 5-6).
A best-practice next step is to define a single reusable prompt template that:
{{ticket_text}}).summary, priority, customer_impact, and next_action.This addresses the formatting failure directly without changing the model or adding extra data sources.
Topic: Content Domain 5: Security, Compliance, and Governance for AI Solutions
A company collects customer chat transcripts to evaluate a generative AI assistant. The company defines how long each data type is kept (for example, raw transcripts for 30 days, redacted aggregates for 1 year) and automates transition and expiration using storage lifecycle rules to ensure timely deletion.
Which governance principle does this practice MOST directly demonstrate?
Best answer: D
Explanation: This practice is about governing data from creation through storage, retention, and end-of-life deletion. Defining retention durations by data type and automating expiration are core data governance lifecycle strategies that reduce risk and support compliance expectations for AI systems.
The core principle is data lifecycle governance: managing data from collection to disposal. In AI solutions, this includes defining what data is stored, how long it is retained, and how it is securely deleted when it is no longer needed. Automating retention and deletion (for example, with storage lifecycle/expiration rules) helps ensure policies are consistently enforced, supports compliance requirements, and reduces exposure of sensitive training/evaluation data. This is distinct from security access controls or model interpretability, which address different risk areas.
Topic: Content Domain 2: Fundamentals of GenAI
A company deploys a GenAI-powered virtual agent (using Amazon Bedrock) to help customers resolve support issues before creating a ticket. The team agrees to evaluate the rollout using average issue resolution time, self-service deflection rate, and customer satisfaction (CSAT).
Which principle is the team applying?
Best answer: A
Explanation: The team is focusing on how to measure whether the GenAI assistant improves the support experience. Resolution time, deflection rate, and CSAT are outcome-oriented KPIs that indicate efficiency, reduced agent workload, and customer sentiment. This reflects the principle of defining business-aligned success metrics for the use case.
For a customer support GenAI use case, you should evaluate success with metrics that reflect the desired business outcomes. Resolution time indicates how quickly issues are solved, deflection rate indicates how often self-service avoids creating a ticket or contacting an agent, and CSAT measures customer-perceived quality of the experience.
Defining these KPIs up front helps you:
Other principles like least privilege, transparency, and robustness are important, but they don’t directly represent the act of selecting and using these support KPIs.
Topic: Content Domain 2: Fundamentals of GenAI
A company is building an internal HR assistant using Amazon Bedrock. The team wants to steer the foundation model to (1) answer only from an approved HR policy excerpt provided in the prompt, (2) refuse requests for employee PII, and (3) return answers in a consistent JSON format for downstream processing.
Which TWO actions are examples of prompt engineering to achieve these goals? (Select TWO.)
bedrock:InvokeModel and encrypt the HR documents with AWS KMS.Best answers: B, D
Explanation: Prompt engineering means designing and iterating on prompts (instructions, context, and examples) to steer a model’s behavior and outputs. Using clear system-level instructions and providing examples of the desired JSON and refusal patterns are direct prompt changes that shape how the model responds. Security, governance, and model selection are important, but they are not prompt engineering techniques.
Prompt engineering is the practice of crafting the input prompt (instructions, context, formatting constraints, and examples) to steer an FM’s responses without changing the model weights. In this scenario, the goal is to constrain the assistant to an approved excerpt, refuse PII-related requests, and produce consistent structured output; those are best addressed by prompt instructions and in-prompt examples.
| Action | Prompt engineering? | Why |
|---|---|---|
| Add system prompt with role/refusal/“use excerpt only” | Directly steers behavior and constraints via instructions | |
| Add few-shot JSON + refusal examples | Demonstrates the target format and refusal style | |
| Enable Bedrock Guardrails | Safety control outside the prompt (policy/filters) | |
| IAM least privilege + KMS encryption | Security controls, not prompt design | |
| CloudTrail logging | Audit/governance capability, not prompt design | |
| Switch to a different FM | Model selection, not prompt design |
Key takeaway: prompt engineering changes what you ask and show the model to influence outputs; other controls wrap or govern the system but don’t redesign the prompt.
Topic: Content Domain 5: Security, Compliance, and Governance for AI Solutions
A team runs a generative AI customer-support app on AWS. The application stores chat transcripts in Amazon S3 and must meet an internal compliance requirement to continuously prove that data stores remain encrypted and not publicly accessible.
During a review, the team discovers that an S3 bucket briefly became public, but they have no reliable configuration history or automated compliance evaluation to detect and document this drift.
Which action addresses the root cause?
Best answer: A
Explanation: The issue is missing continuous configuration recording and compliance evaluation for AWS resources. AWS Config provides configuration history (configuration items) and can automatically evaluate resources against rules such as S3 public access and encryption settings. This produces ongoing compliance status and evidence, not just point-in-time findings.
This is a governance/compliance gap: the team cannot demonstrate what the S3 bucket configuration was over time or automatically assess whether it stayed compliant. AWS Config is designed to record resource configuration changes (creating a history) and to evaluate those configurations against compliance requirements using AWS Config Rules or conformance packs. In this scenario, Config can track when the bucket’s public access settings changed and continuously evaluate encryption and public access posture, generating a compliance timeline and alerts. Logging or threat-detection services help with auditing access or detecting suspicious activity, but they do not provide continuous configuration state and compliance evaluation as the primary control.
Topic: Content Domain 5: Security, Compliance, and Governance for AI Solutions
Which statement best describes AWS Key Management Service (AWS KMS) and how it supports encryption key management for AI solutions on AWS?
Best answer: A
Explanation: AWS KMS is a managed service for creating and managing encryption keys and defining who can use them through key policies and IAM. In AI solutions, KMS keys are commonly used to encrypt data at rest (such as datasets, prompts, and outputs) and to control and audit key usage.
AWS KMS provides centralized encryption key management in AWS. You can create AWS managed keys or customer managed keys and then use them with integrated AWS services to encrypt and decrypt data used in AI solutions (for example, data stored in Amazon S3 or encrypted volumes and databases that store model inputs and outputs). Access to use a KMS key is controlled with KMS key policies and IAM permissions, which helps enforce least privilege. KMS also supports auditing of key usage through AWS logging, helping with governance and compliance evidence. KMS manages key material lifecycle and helps you avoid building and operating your own key management infrastructure.
Topic: Content Domain 3: Applications of Foundation Models
A company wants an LLM-based assistant to draft responses to customer disputes in a highly regulated insurance domain. A team proposes training a new model from scratch on Amazon SageMaker AI, but they have only 8,000 labeled examples, a 4-week deadline, and a limited GPU budget. Compliance also prohibits creating a persistent vector index for retrieval (no embeddings store).
Which change best optimizes cost and time while still adapting the model to the insurance domain?
Best answer: C
Explanation: Transfer learning adapts a pretrained model to a new domain by fine-tuning on domain-specific examples, which usually needs much less data and compute than training from scratch. Given the small labeled dataset, tight timeline, and limited GPU budget, fine-tuning is the most practical way to achieve domain adaptation while controlling cost and delivery time.
Domain adaptation is the goal of making a model perform well on a specific target domain (here, regulated insurance language and dispute-response style). Transfer learning is the common way to do this in practice: start from a pretrained foundation model that already learned general language patterns, then fine-tune it on representative domain examples so it learns domain terminology, tone, and decision patterns with far less compute and data.
In this scenario, fine-tuning optimizes:
Because a persistent vector index is prohibited, approaches that rely on retrieval-backed embeddings don’t meet the constraints; fine-tuning adapts the model without needing an embeddings store at inference time.
Topic: Content Domain 2: Fundamentals of GenAI
A company is building a GenAI assistant on AWS. Users can submit chat text, screenshots of error dialogs, and short voice messages. The team will use Amazon Bedrock and wants to choose appropriate model modalities while following good governance.
Which TWO approaches are INCORRECT? (Select TWO.)
bedrock:* on * so all developers can use any modelBest answers: A, F
Explanation: Modality selection should match the input type (text, image, audio, or multimodal) while still enforcing basic security and governance controls. Broad access to all Bedrock actions and storing sensitive inputs in public, unencrypted storage are unsafe practices that increase the risk of unauthorized use and data leakage.
When selecting GenAI model modalities, align the model (or preprocessing) to the data type: text models for text, transcription before text modeling for audio, and multimodal models (or OCR services) for images. In AWS, you can combine managed services (for example, Amazon Transcribe or Amazon Textract) with Amazon Bedrock models to fit the modality needed.
Governance must be applied regardless of modality:
The key takeaway is that correct modality choices don’t justify weakening access control or data protection.
Topic: Content Domain 2: Fundamentals of GenAI
A company built an internal HR chatbot using an Amazon Bedrock foundation model. The team currently runs a weekly fine-tuning job so the model “learns” the latest policy PDFs stored in Amazon S3.
Constraints: Policies change multiple times per day, updates must be reflected within hours, operational effort and cost must decrease, and the solution must reduce hallucinations by grounding answers in approved content.
Which change best optimizes the solution while meeting all constraints?
Best answer: D
Explanation: Replacing frequent fine-tuning with retrieval-augmented generation (RAG) lets the application use updated S3 documents immediately while keeping the underlying model unchanged. This shifts the problem from the fine-tuning stage to data selection and deployment-time retrieval, which typically lowers operational effort and cost. Grounding responses on retrieved, approved passages also reduces hallucinations.
In the foundation model lifecycle, pre-training and fine-tuning are model adaptation stages that are slower and operationally heavier than updating the knowledge used at inference time. Because the policies change multiple times per day, repeatedly fine-tuning to “bake in” new content is inefficient and risks serving stale answers.
Using Amazon Bedrock Knowledge Bases implements RAG during deployment:
Key takeaway: when content changes frequently, prefer RAG over fine-tuning for freshness and lower ops.
Topic: Content Domain 3: Applications of Foundation Models
Which statement is INCORRECT about high-level qualitative criteria used to evaluate GenAI outputs (helpfulness, accuracy, relevance, safety)?
Best answer: D
Explanation: The incorrect statement is the one that treats safety as optional when other criteria look good. Safety is a separate qualitative dimension that checks for harmful, disallowed, or sensitive outputs even when the response is useful, on-topic, and factually correct. Evaluations typically consider all four criteria together to understand overall output quality.
Qualitative evaluation of GenAI outputs commonly scores multiple dimensions because they capture different failure modes. In the scenario, helpfulness, accuracy, relevance, and safety are complementary: a response can be helpful but irrelevant to the user’s request, relevant but inaccurate (hallucinated), or accurate and relevant but still unsafe (for example, providing instructions for wrongdoing or disclosing sensitive data).
A simple way to apply these criteria is:
Key takeaway: safety is not implied by the other criteria and must be evaluated explicitly.
Topic: Content Domain 2: Fundamentals of GenAI
A company wants a GenAI chatbot for its internal HR help desk. The answers must include citations back to the exact policy paragraphs, and policies change weekly so updates must be reflected quickly without retraining a model. The HR documents must remain in the company’s AWS account.
Which solution is the BEST fit?
Best answer: C
Explanation: Retrieval-augmented generation (RAG) is designed to ground an LLM on up-to-date enterprise content while keeping the model unchanged. Using Amazon Bedrock Knowledge Bases lets the chatbot retrieve relevant HR policy passages from documents stored in the company’s AWS account and return responses with citations. This best matches the requirements for source attribution and frequent content updates.
The key decision is choosing between in-context prompting, RAG, and fine-tuning. When the business requires answers to be based on specific, changing internal documents with citations, RAG is typically the best strategy because it retrieves authoritative passages at request time and uses them as context for generation. With Amazon Bedrock Knowledge Bases, the HR documents can stay in Amazon S3 in the company account, the knowledge base can be updated as documents change, and the model can cite the retrieved sources.
Fine-tuning is better for changing the model’s style/behavior or improving performance on a stable task, but it does not inherently provide per-answer citations and is not ideal when content changes weekly. Pure prompting can help format and tone, but it cannot reliably ensure factual grounding or citations from a large document set.
Topic: Content Domain 4: Guidelines for Responsible AI
Which TWO statements about training dataset characteristics that support responsible AI are INCORRECT?
Best answers: D, F
Explanation: Responsible datasets aim for diverse, inclusive coverage and balanced representation so model performance is more equitable across groups. It is unsafe to drop an underrepresented group to “fix” bias, and it is also unsafe to assume historical outcomes are neutral when they may reflect biased processes. Better approaches include improving coverage, validating labels, and documenting known limitations.
A key responsible-AI dataset characteristic is whether the data represents the real-world populations and use cases the model will affect, including groups that are often under-sampled. “Balanced representation” typically means reducing harmful skews that cause systematically worse performance for some groups.
Two dataset practices are unsafe here:
Stronger signals of responsibility include collecting more representative data (or carefully re-sampling), validating label quality across groups, and documenting data provenance and known gaps so stakeholders understand limits and risks.
Topic: Content Domain 2: Fundamentals of GenAI
A company wants to add generative text summarization to an existing application. The team wants to call a fully managed service that provides API access to multiple foundation models and does not require provisioning or managing model training infrastructure or inference endpoints.
Which AWS service best fits this requirement?
Best answer: A
Explanation: The discriminating factor is whether the service offers managed foundation model inference through a simple API without managing model endpoints. Amazon Bedrock is designed for accessing and invoking multiple foundation models as managed APIs. This matches the team’s requirement to avoid provisioning and operating inference infrastructure.
The deciding attribute is “managed FM APIs vs. broader ML platform.” Amazon Bedrock is the AWS service that provides managed access to multiple foundation models through a consistent API, so teams can integrate generative capabilities (like summarization) without standing up or operating their own model hosting.
Amazon SageMaker AI is a broader ML platform used to build, train, customize, and deploy ML models (including hosting endpoints you manage). While it can be used for generative workloads, it is not primarily the single-API, multi-FM managed access layer described in the scenario. Key takeaway: choose Bedrock when you want managed FM invocation; choose SageMaker AI when you need a full ML lifecycle platform.
Topic: Content Domain 4: Guidelines for Responsible AI
A retail company wants to use a generative AI solution on Amazon Bedrock to draft marketing copy. Legal leadership is most concerned about intellectual property (IP) infringement claims and reputational damage from outputs that are too similar to copyrighted third-party content.
Which approach best reduces this specific risk?
Best answer: B
Explanation: The deciding factor is whether the content used to generate responses is licensed or company-owned with clear usage rights. Building a RAG solution that retrieves only approved, properly licensed material reduces exposure to copyright infringement allegations and the reputational harm that can follow. Controls like safety filters or encryption do not address IP ownership of the source content.
A key legal and reputational risk in GenAI is generating content that infringes on third-party intellectual property, especially when the system is trained on or grounded with copyrighted material that the organization does not have rights to use. The most direct mitigation is controlling data provenance: restrict the system’s grounding/training sources to content that is company-owned or properly licensed and maintain traceability of what sources are used.
Using Amazon Bedrock with a RAG pattern (for example, a knowledge base backed by approved documents in Amazon S3) helps because the model is guided by retrieved, authorized references rather than unvetted third-party text. In contrast, safety guardrails and security controls improve safety and confidentiality but do not resolve whether the underlying content can be legally used.
Topic: Content Domain 2: Fundamentals of GenAI
A company is building a customer-support chatbot using Amazon Bedrock. Most users are in Germany and France, and a new compliance requirement states that all customer content (prompts and model responses) must be processed and stored only within the EU.
Which approach best satisfies this requirement?
Best answer: A
Explanation: The deciding attribute is regional data residency: prompts and responses must not leave the EU. Invoking Amazon Bedrock in an EU Region and keeping storage in the EU meets the compliance constraint and typically improves latency for users located in nearby EU countries.
Regional coverage affects GenAI solutions because the AWS Region you choose determines where requests are processed (latency), what happens during regional disruptions (availability), and whether data residency/compliance requirements are met. In this scenario, the single discriminator is the requirement that customer content (prompts and responses) be processed and stored only within the EU.
To satisfy that requirement at a high level:
Any design that sends prompts/responses to a non-EU Region, even as a performance optimization or failover path, violates the stated compliance requirement.
Topic: Content Domain 1: Fundamentals of AI and ML
A company is building a customer-support chatbot using Amazon Bedrock. Two foundation models are being considered.
The product owner wants to choose the model with the best expected business ROI while minimizing negative customer impact.
Which TWO actions should the team take? (Select TWO)
bedrock:InvokeModel to determine which model has higher ROIBest answers: B, E
Explanation: To compare models by business value, the team should translate model performance (resolution rate) into financial impact (avoided agent cost) and subtract the model’s inference costs to estimate ROI. They should also validate real user impact through a controlled pilot that captures customer feedback and safety/quality signals that can change the realized business outcome.
The core idea is to connect offline model metrics to business outcomes by expressing them in money and customer experience. Here, resolution rate determines how often the chatbot produces measurable savings, while token-based inference pricing determines cost per chat; together they produce an expected net value. Because offline test performance can differ from production behavior, a limited A/B pilot helps measure real CSAT and escalation rates, and adding guardrails plus targeted human review helps quantify safety-related quality issues that can reduce ROI through rework, churn, or brand risk.
Convert resolution rate and token cost into monthly net ROI
A/B pilot with guardrails + CSAT/escalation + human review
Higher resolution alone ignores per-chat cost and net value
CloudTrail supports auditability, not ROI comparison
SSE-KMS improves security, not inference cost per user
Fine-tuning adds dev cost/time and isn't required to compare ROI
Key takeaway: choose actions that quantify net value (savings minus cost) and validate customer impact, not just technical scores or controls.
Topic: Content Domain 4: Guidelines for Responsible AI
Which statement correctly describes a legal and reputational risk related to intellectual property (IP) when an organization uses generative AI to produce public-facing content (for example, product images or marketing copy)?
Best answer: B
Explanation: A key GenAI risk is that outputs can be substantially similar to copyrighted or trademarked material, which can lead to infringement claims and reputational damage if published. This risk exists even when using managed models, so organizations should implement review and usage controls before releasing content.
Generative AI can produce novel-looking text or images that still resemble protected works (copyright) or brand identifiers (trademark). If an organization publishes that output externally, it can face IP infringement claims and associated reputational harm. This is not eliminated simply by choosing a managed foundation model; under AWS’s shared responsibility model, customers remain responsible for how they use outputs and whether published content complies with applicable laws and licensing obligations. Practical mitigations include human review workflows, content policies, and guardrails to reduce risky generations, but these controls reduce risk rather than guarantee zero infringement.
Use this map after the sample questions to connect individual items to the AWS AI Practitioner fundamentals decisions these practice samples test.
flowchart LR
S1["Business AI scenario"] --> S2
S2["Identify AI ML or GenAI workload type"] --> S3
S3["Choose AWS AI service family"] --> S4
S4["Apply responsible AI and security basics"] --> S5
S5["Review cost data and governance"] --> S6
S6["Select next AWS AI route"]
| Cue | What to remember |
|---|---|
| Workload type | Separate predictive ML, generative AI, computer vision, language, speech, and recommendation scenarios. |
| AWS service fit | Know where Bedrock, SageMaker, Comprehend, Rekognition, Textract, Lex, Polly, and Transcribe fit at a high level. |
| Responsible AI | Watch for privacy, bias, explainability, safety, governance, and human oversight cues. |
| Data basics | AI quality depends on relevant, clean, governed, and sufficiently representative data. |
| Business value | Connect AI choices to measurable outcomes, risk, and cost rather than tool novelty. |