☁️ AWS

Free AWS Certified AI Practitioner (AIF-C01) Study Resources

Comprehensive AWS Certified AI Practitioner (AIF-C01) hive provides study notes, question bank with practice tests, flashcards, and hands-on labs, all supported by a personal AI tutor to help you master the AWS Certified AI Practitioner (AIF-C01) certification.

353
Practice Questions
5
Mock Exams
145
Study Notes
340
Flashcard Decks
2
Source Materials

AWS Certified AI Practitioner (AIF-C01) Study Notes & Guides

145 AI-generated study notes covering the full AWS Certified AI Practitioner (AIF-C01) curriculum. Showing 10 complete guides below.

Curriculum Overview785 words

Curriculum Overview: AI Concepts and Terminology (AIF-C01)

AI Concepts and Terminology

Read full article

Curriculum Overview: AI Concepts and Terminology

This curriculum is designed to prepare learners for the AWS Certified AI Practitioner (AIF-C01) exam. It covers the foundational pillars of Artificial Intelligence, Machine Learning, and Generative AI, specifically focusing on how these technologies are implemented within the AWS ecosystem.

Prerequisites

Before starting this curriculum, learners should have a basic understanding of the following:

  • Basic Cloud Literacy: Familiarity with cloud computing concepts (e.g., storage, compute, and APIs).
  • Data Fundamentals: An understanding of what data is (structured vs. unstructured) and how it is used in a business context.
  • Business Logic: Ability to identify business problems that might benefit from automation or prediction.
  • No Coding Required: While technical, this curriculum focuses on high-level concepts and managed services rather than deep programming or advanced calculus.

Module Breakdown

The curriculum is structured into six core units, progressing from theoretical foundations to practical AWS implementations.

UnitTitleFocus AreaDifficulty
1Fundamentals of AI & MLHierarchy of AI/ML/DL, Data Types, and Lifecycle⭐⭐ (Foundation)
2Fundamentals of GenAITransformers, Tokens, and LLM basics⭐⭐⭐ (Intermediate)
3Applications of Foundation ModelsPrompt Engineering and Model Adaptation⭐⭐⭐ (Intermediate)
4Guidelines for Responsible AIEthics, Bias, Fairness, and Explainability⭐⭐ (Critical)
5Security & GovernanceData Protection, Compliance, and Shared Responsibility⭐⭐⭐ (Advanced)
6AWS AI ServicesManaged services (Bedrock, SageMaker, Q)⭐⭐ (Practical)

Learning Objectives per Module

Unit 1: Fundamentals of AI and Machine Learning

  • The Hierarchy: Distinguish between Artificial Intelligence (broadest), Machine Learning (subset), and Deep Learning (specialized subset using neural networks).
  • Data Savvy: Identify different data formats such as Structured (tabular data like Excel), Unstructured (images/video), and Time-Series (stock prices over time).
  • Inferencing: Explain the difference between Batch Inference (processing bulk data at intervals) and Real-time Inference (immediate response to a user request).
Loading Diagram...

Unit 2 & 3: Generative AI and Foundation Models

  • Mechanics: Define Tokens (the basic units of text processed by models) and Embeddings (numerical vector representations of data).
  • Prompt Engineering: Apply techniques like Few-shot prompting (providing examples in the prompt) to improve model accuracy.
  • Adaptation: Compare RAG (Retrieval-Augmented Generation) vs. Fine-tuning for providing models with specialized knowledge.

Unit 4 & 5: Responsible AI and Security

  • Fairness & Bias: Identify types of bias in datasets and how they affect model outcomes (e.g., demographic subgroup inaccuracy).
  • Explainability: Use tools like Amazon SageMaker Model Cards to document model intentions and limitations.
  • Security: Implement the AWS Shared Responsibility Model to secure AI systems, protecting against threats like Prompt Injection.

Success Metrics

To demonstrate mastery of this curriculum, the learner must be able to:

  1. Categorize Use Cases: Correctly identify if a problem requires a Classification (Labeling), Regression (Predicting a number), or Clustering (Finding patterns) approach.
  2. Service Selection: Choose the correct AWS service for a task (e.g., use Amazon Rekognition for image analysis or Amazon Bedrock for accessing LLMs).
  3. Lifecycle Mapping: Describe the steps of the ML Development Lifecycle using the correct terminology.
Loading Diagram...

Real-World Application

Understanding AI concepts and AWS terminology is not just for passing an exam; it has direct career applications:

  • Predictive Maintenance: Using time-series data from factory sensors to predict when a machine will fail before it happens.
  • Customer Experience: Implementing Amazon Lex to build conversational chatbots that handle 80% of routine customer queries.
  • Fraud Detection: Utilizing ML models to analyze transaction patterns in real-time to flag suspicious activity.
  • Responsible Innovation: Ensuring that a company's AI tools do not accidentally discriminate against specific groups by monitoring for bias using SageMaker Clarify.

[!IMPORTANT] AI is a rapidly evolving field. This curriculum focuses on the "Foundational" layer, meaning it prioritizes understanding the what and the why over the specialized implementation of the how.

Hands-On Lab940 words

Hands-On Lab: Exploring AI Concepts with AWS Managed Services

AI Concepts and Terminology

Read full article

Hands-On Lab: Exploring AI Concepts with AWS Managed Services

Welcome to this guided lab! In this session, we will bridge the gap between theoretical Artificial Intelligence (AI) concepts and practical implementation. You will work with unstructured data, perform real-time inferencing, and see the difference between Computer Vision (CV) and Natural Language Processing (NLP) using AWS managed services.

Prerequisites

Before you begin, ensure you have the following ready:

  • AWS Account: An active AWS account with an IAM user or role that has AdministratorAccess (or permissions for Amazon S3, Amazon Rekognition, and Amazon Comprehend).
  • AWS CLI: The AWS Command Line Interface installed and configured (aws configure) with your credentials and a default region (e.g., us-east-1).
  • Basic Terminal Knowledge: Familiarity with running commands in a bash/zsh or PowerShell terminal.
  • Conceptual Understanding: Basic awareness of AI, ML, Computer Vision, and NLP as covered in the study guide.

Concept Review: The AI Hierarchy

Before we build our architecture, let's review the relationship between AI, Machine Learning (ML), and Deep Learning (DL). The services we will use today abstract away the deep learning layers, allowing you to focus on the AI application.

Compiling TikZ diagram…
Running TeX engine…
This may take a few seconds

Learning Objectives

By the end of this lab, you will be able to:

  1. Differentiate Data Types: Handle unstructured data (images and raw text) in an AI pipeline.
  2. Execute Real-Time Inferencing: Send single-request data to pre-trained foundation models and receive immediate predictions.
  3. Apply Computer Vision (CV): Use Amazon Rekognition to extract Optical Character Recognition (OCR) data from an image.
  4. Apply Natural Language Processing (NLP): Use Amazon Comprehend to determine the sentiment of a text string.

Architecture Overview

The following flowchart visualizes the data flow for this lab. We will use the CLI to interact with unstructured data (an image), extract text using a CV model, and then analyze text using an NLP model.

Loading Diagram...

Step-by-Step Instructions

Step 1: Create a Storage Bucket for Unstructured Data

AI models require data. In this step, we will create an Amazon S3 bucket to hold our unstructured data (images).

bash
aws s3 mb s3://brainybee-lab-ai-concepts-<YOUR_ACCOUNT_ID>

[!TIP] S3 bucket names must be globally unique. Replace <YOUR_ACCOUNT_ID> with your actual AWS account number or a random string of numbers to ensure the name is available.

Console alternative
  1. Navigate to the S3 Console.
  2. Click Create bucket.
  3. Enter brainybee-lab-ai-concepts-<YOUR_ACCOUNT_ID> as the Bucket name.
  4. Leave all other settings as default and click Create bucket.

📸 Screenshot: S3 Create Bucket Form

Step 2: Download and Upload an Image

We need an image containing text to test our Computer Vision model. We will download a sample image and upload it to our new S3 bucket.

bash
# Download a sample image containing text curl -o sample-text.jpg https://raw.githubusercontent.com/aws-samples/amazon-rekognition-code-samples/master/images/text.jpg # Upload it to your S3 bucket aws s3 cp sample-text.jpg s3://brainybee-lab-ai-concepts-<YOUR_ACCOUNT_ID>/
Console alternative
  1. Download the image from this link to your computer.
  2. Navigate to your newly created S3 bucket in the console.
  3. Click Upload, select your downloaded sample-text.jpg, and click Upload.

Checkpoints

Let's verify our unstructured data is safely stored in S3.

bash
aws s3 ls s3://brainybee-lab-ai-concepts-<YOUR_ACCOUNT_ID>/

Expected Output: You should see sample-text.jpg listed with its timestamp and file size.

Step 3: Perform Real-Time Inferencing with Computer Vision (CV)

Now, we will use Amazon Rekognition, a managed Deep Learning service for Computer Vision. We will perform real-time inferencing to detect text (OCR) within our unstructured image.

bash
aws rekognition detect-text \ --image '{"S3Object":{"Bucket":"brainybee-lab-ai-concepts-<YOUR_ACCOUNT_ID>","Name":"sample-text.jpg"}}' \ --region us-east-1

[!NOTE] Look at the JSON output. This represents a model prediction. The service returns DetectedText and a Confidence score (a percentage indicating how certain the ML algorithm is about its prediction).

Console alternative
  1. Navigate to the Amazon Rekognition Console.
  2. In the left sidebar, choose Text in image.
  3. Expand the Upload dropdown and upload your local sample-text.jpg.
  4. View the extracted text results and confidence scores in the Results pane on the right.

📸 Screenshot: Rekognition Text Detection Results

Step 4: Perform Real-Time Inferencing with Natural Language Processing (NLP)

Next, let's explore Natural Language Processing (NLP). We will use Amazon Comprehend to analyze the sentiment of a sentence. This is an example of taking unstructured text and turning it into structured, labeled insights.

bash
aws comprehend detect-sentiment \ --text "Artificial Intelligence is transforming the world in amazing ways!" \ --language-code en \ --region us-east-1

[!TIP] The output will show a Sentiment (e.g., POSITIVE) and a SentimentScore breakdown for Positive, Negative, Neutral, and Mixed. This highlights how AI models deal with probability rather than absolute certainty.

Console alternative
  1. Navigate to the Amazon Comprehend Console.
  2. In the left sidebar, click Real-time analysis.
  3. Under Input text, paste: Artificial Intelligence is transforming the world in amazing ways!
  4. Click Analyze.
  5. Scroll down to the Insights tab and view the Sentiment results.

Teardown

[!WARNING] Remember to run the teardown commands to avoid ongoing charges. Leaving data in S3 incurs minor storage costs.

To clean up your AWS environment, empty and delete the S3 bucket, and remove your local file.

bash
# Delete the image from the S3 bucket aws s3 rm s3://brainybee-lab-ai-concepts-<YOUR_ACCOUNT_ID>/sample-text.jpg # Delete the S3 bucket aws s3 rb s3://brainybee-lab-ai-concepts-<YOUR_ACCOUNT_ID> # Remove the local file rm sample-text.jpg

Troubleshooting

Common ErrorCauseFix
AccessDenied or InvalidAccessKeyIdYour AWS CLI is not configured correctly or your IAM user lacks permissions.Run aws configure and ensure you are using an IAM user with AdministratorAccess or specific service policies.
BucketAlreadyExistsS3 bucket names are globally unique. Someone else has used this name.Change <YOUR_ACCOUNT_ID> to a random string of numbers/letters in Steps 1, 2, and 3.
InvalidS3ObjectException in RekognitionThe bucket name or file name in the JSON string is incorrect.Double-check the spelling of the bucket and ensure sample-text.jpg was successfully uploaded in Step 2.
UnrecognizedClientException in ComprehendOften caused by specifying an unsupported region.Ensure you append --region us-east-1 (or another supported region) to your CLI command.

Cost Estimate

  • Amazon S3: First 50 TB/month is fractions of a cent per GB. Uploading one small image falls well within the AWS Free Tier.
  • Amazon Rekognition: Free tier allows 1,000 images per month. Outside free tier, ~$0.001 per image.
  • Amazon Comprehend: Free tier allows 50,000 units of text (100 characters per unit) per month. Outside free tier, ~$0.0001 per unit.
  • Total Expected Cost: $0.00 (Assuming free tier eligibility or just a few cents otherwise).
Hands-On Lab962 words

Hands-On Lab: Exploring Basic AI Concepts and Terminology

AI Concepts and Terminology

Read full article

Hands-On Lab: Exploring Basic AI Concepts and Terminology

Welcome to this guided hands-on lab! This module bridges the theoretical concepts from Chapter 1 (Fundamentals of AI and ML) with practical reality. You will interact with pre-trained Artificial Intelligence (AI) models using AWS managed services to perform Computer Vision and Natural Language Processing (NLP) tasks.

By completing this lab, abstract terms like unstructured data, real-time inferencing, and confidence scores will become tangible.

Prerequisites

Before starting, ensure you have the following:

  • AWS Account: An active AWS account with administrative or developer access.
  • CLI Tools: AWS CLI installed and configured (aws configure) on your local machine or AWS CloudShell.
  • Prior Knowledge: Familiarity with basic terminal commands and a conceptual understanding of AI vs. Machine Learning (ML).

Learning Objectives

  1. Differentiate AI fields in practice: Use distinct services for Computer Vision (Amazon Rekognition) and Natural Language Processing (Amazon Comprehend).
  2. Understand Data Types: Work with unstructured data formats (images and raw text) rather than tabular data.
  3. Perform Real-Time Inferencing: Send data to an ML model and receive immediate probabilistic predictions.
  4. Analyze Model Outputs: Interpret JSON responses containing confidence scores to understand the probabilistic nature of AI.

Architecture Overview

This lab uses a simple serverless architecture to demonstrate real-time inferencing on unstructured data.

Loading Diagram...

Step-by-Step Instructions

Step 1: Prepare Unstructured Data (Amazon S3)

In Machine Learning, images and text are considered unstructured data. Deep learning models are particularly adept at processing this format. Let's create a storage bucket and upload a sample image.

bash
# 1. Create a globally unique S3 bucket (replace <YOUR_ACCOUNT_ID>) aws s3 mb s3://brainybee-lab-vision-<YOUR_ACCOUNT_ID> # 2. Download a sample image of a city skyline (or use any local JPEG) curl -o sample.jpg https://raw.githubusercontent.com/aws-samples/amazon-rekognition-developer-guide/master/images/city.jpg # 3. Upload the unstructured data to S3 aws s3 cp sample.jpg s3://brainybee-lab-vision-<YOUR_ACCOUNT_ID>/

📸 Screenshot: A terminal showing the successful upload: sample.jpg to s3://... output.

Console alternative
  1. Navigate to Amazon S3 in the AWS Console.
  2. Click Create bucket, name it brainybee-lab-vision-<YOUR_ACCOUNT_ID>, and click Create bucket.
  3. Select your new bucket, click Upload, and upload a local image file named sample.jpg.

Step 2: Real-Time Inferencing with Computer Vision

Computer Vision is an AI field enabling systems to derive information from images. We will use Amazon Rekognition to perform real-time inferencing (getting immediate predictions) on our unstructured image.

bash
# Run real-time inferencing to detect labels (objects, scenes) in the image aws rekognition detect-labels \ --image '{"S3Object":{"Bucket":"brainybee-lab-vision-<YOUR_ACCOUNT_ID>","Name":"sample.jpg"}}' \ --max-labels 5

[!TIP] Look at the JSON response. Notice the Confidence attribute next to each label. AI models do not return absolute facts; they return probabilities based on their training data.

Console alternative
  1. Navigate to Amazon Rekognition in the AWS Console.
  2. In the left sidebar, click Label detection.
  3. Under the demo section, expand the Upload panel and upload your sample.jpg.
  4. Observe the detected labels and their percentage scores (confidence) in the Results pane.

Step 3: Natural Language Processing (NLP)

Natural Language Processing (NLP) allows computers to interpret human language. We will use Amazon Comprehend to analyze the sentiment of unstructured text.

bash
# Perform NLP inferencing to determine text sentiment aws comprehend detect-sentiment \ --text "Artificial intelligence is a fascinating and transformative technology, though data drift can occasionally cause headaches!" \ --language-code "en"

📸 Screenshot: The JSON output from Comprehend showing a Sentiment of MIXED or POSITIVE, along with SentimentScore breakdowns.

Console alternative
  1. Navigate to Amazon Comprehend in the AWS Console.
  2. Scroll down to the Real-time analysis section.
  3. Paste the text: "Artificial intelligence is a fascinating and transformative technology, though data drift can occasionally cause headaches!"
  4. Click Analyze and view the Sentiment results in the Insights tab.

Checkpoints

Verify your progress after completing the steps above:

  • Checkpoint 1: Run aws s3 ls s3://brainybee-lab-vision-<YOUR_ACCOUNT_ID>. You should see sample.jpg listed.
  • Checkpoint 2: In your Rekognition JSON output, did you receive a list of Labels? If you used a city image, you should see labels like City, Building, or Urban.
  • Checkpoint 3: In your Comprehend JSON output, locate the SentimentScore object. Notice how the scores for Positive, Negative, Neutral, and Mixed all add up to approximately 1.0 (100%).

Concept Review

To solidify your understanding of Chapter 1, let's visually review the hierarchy of AI and the services you just used.

Compiling TikZ diagram…
Running TeX engine…
This may take a few seconds

Terminology Mapping Table

Textbook ConceptLab ApplicationDescription
Unstructured Datasample.jpg, Raw TextData that does not fit neatly into rows and columns.
Inferencingdetect-labels, detect-sentimentThe process of passing new data through a trained model to make a prediction.
Real-time InferenceCLI Command ResponseReceiving immediate, synchronous predictions with low latency.
Computer VisionAmazon RekognitionDeep learning applied to visual data extraction.
NLPAmazon ComprehendDeep learning applied to human language understanding.

Clean-Up / Teardown

[!WARNING] Remember to run the teardown commands to avoid ongoing charges. While Rekognition and Comprehend charge per-request, S3 storage incurs ongoing costs.

Execute the following commands to delete the resources created in this lab:

bash
# Empty the S3 bucket aws s3 rm s3://brainybee-lab-vision-<YOUR_ACCOUNT_ID> --recursive # Delete the S3 bucket aws s3 rb s3://brainybee-lab-vision-<YOUR_ACCOUNT_ID>

Note: Amazon Rekognition and Amazon Comprehend APIs used in this lab do not provision permanent infrastructure; they are stateless API calls.

Troubleshooting

Error MessageLikely CauseSolution
AccessDeniedIAM User lacks permissions.Ensure your IAM user has AmazonS3FullAccess, AmazonRekognitionFullAccess, and ComprehendFullAccess policies attached.
BucketAlreadyExistsS3 bucket names are globally unique.Change <YOUR_ACCOUNT_ID> to something highly unique (e.g., your name plus random numbers).
InvalidImageFormatExceptionRekognition doesn't support the file.Ensure the image uploaded to S3 is a .jpg or .png file under 5MB.
Could not connect to the endpoint URLCLI region configuration issue.Run aws configure and ensure you have a valid default region set (e.g., us-east-1).

Stretch Challenge

Want to test your skills further without step-by-step guidance?

Challenge: Try passing an image containing printed text (like a picture of a street sign or a book page) to Amazon Rekognition to extract the words. Hint: Look up the detect-text API method for Rekognition.

Show solution
bash
# Upload an image with text first aws s3 cp sign.jpg s3://brainybee-lab-vision-<YOUR_ACCOUNT_ID>/ # Run the detect-text API (Optical Character Recognition - OCR) aws rekognition detect-text \ --image '{"S3Object":{"Bucket":"brainybee-lab-vision-<YOUR_ACCOUNT_ID>","Name":"sign.jpg"}}'
Curriculum Overview820 words

Curriculum Overview: Mastering Amazon Bedrock and Amazon Q

Amazon Bedrock and Amazon Q

Read full article

Curriculum Overview: Mastering Amazon Bedrock and Amazon Q

This curriculum provides a comprehensive roadmap for understanding and implementing AWS's primary generative AI platforms. Learners will move from foundational model access via Amazon Bedrock to specialized AI assistance with Amazon Q.

Prerequisites

Before starting this curriculum, students should possess:

  • Cloud Fundamentals: Basic knowledge of AWS infrastructure (IAM, VPCs, and S3).
  • AI Literacy: Understanding of basic GenAI terms (Tokens, LLMs, and Hallucinations).
  • Security Basics: Familiarity with the AWS Shared Responsibility Model.

Module Breakdown

ModuleTitlePrimary FocusDifficulty
1Bedrock FoundationsAccessing FMs, Playgrounds, and Model SelectionBeginner
2Advanced BedrockRAG, Knowledge Bases, and AI AgentsIntermediate
3Amazon Q BusinessOrganizational search, Q Apps, and connectivityIntermediate
4Amazon Q DeveloperIDE integration, code optimization, and CLIAdvanced
5Security & GovernanceGuardrails, private data, and complianceIntermediate

Module Objectives per Module

Module 1: Amazon Bedrock Foundations

  • Model Access: Explain how to access multiple Foundation Models (FMs) through a single API.
  • Parameter Tuning: Master the effects of TemperatureTemperature and TopPTop P on model creativity.
  • Playgrounds: Utilize text, image, and video playgrounds for rapid prototyping.

Module 2: Building with Bedrock

  • Retrieval-Augmented Generation (RAG): Implement Knowledge Bases for proprietary data.
  • Agentic AI: Configure Amazon Bedrock Agents to execute multi-step business tasks.
  • Model Evaluation: Use ROUGE and BLEU scores to assess model performance.

Module 3: Amazon Q Business

  • Unified Search: Index corporate data across Slack, Microsoft 365, and SharePoint.
  • Q Apps: Create no-code applications for content generation and workflow automation.
  • Transparency: Utilize citations and references to ensure response accuracy.

Module 4: Amazon Q Developer

  • Development Speed: Increase coding velocity by up to 80% using IDE plugins.
  • Modernization: Use agents for heavy-duty tasks like Java or .NET migrations.
  • AWS Integration: Query AWS account resources and billing directly via the console.
Loading Diagram...

Success Metrics

How to know you have mastered the curriculum:

  1. Deployment Proficiency: Successfully deploy an Amazon Bedrock Agent that triggers a Lambda function.
  2. Productivity Gains: Demonstrate a measurable reduction in coding time using Amazon Q Developer.
  3. Accuracy Verification: Maintain a hallucination rate below a defined threshold using Negative Prompts and Guardrails.
  4. Financial Efficiency: Optimize model selection based on token pricing and performance needs.

[!IMPORTANT] Success is not just building a model; it is building a safe model. Ensure all applications utilize Amazon Bedrock Guardrails to prevent prompt injection and data leakage.


Real-World Application

Why this curriculum matters in a career:

  • Legacy Modernization: Amazon used Q Developer to migrate tens of thousands of applications to Java 17, saving $260 million and 4,500 years of manual labor.
  • Employee Efficiency: Amazon Q Business acts as a 24/7 subject matter expert, reducing the time employees spend searching for internal documentation.
  • Lower Entry Barrier: Platforms like PartyRock allow non-developers to create AI tools, democratizing innovation across business units.
Loading Diagram...

Comparison: Bedrock vs. Q

FeatureAmazon BedrockAmazon Q
Target AudienceDevelopers building custom appsBusiness users & Software engineers
Core FunctionInfrastructure/API for FMsSpecialized AI Assistant
CustomizationFine-tuning, RAG, AgentsConnectors to 40+ Enterprise tools
Key OutcomeBuilding new GenAI servicesIncreasing workflow productivity
Click to expand: Specific Business Use Cases
  • Automated Inventory: Using Bedrock to monitor supply chain data and recommend reorder points.
  • Code Debugging: Using Q Developer to identify security vulnerabilities in Python or Java code.
  • Content Creation: Using Q Apps to generate marketing emails based on internal product manuals.
Hands-On Lab1,058 words

Hands-On Lab: Building and Automating with Amazon Bedrock and Amazon Q

Amazon Bedrock and Amazon Q

Read full article

Hands-On Lab: Building and Automating with Amazon Bedrock and Amazon Q

Welcome to this guided lab. In this session, you will explore the powerful synergy between Amazon Bedrock (a managed service for Foundation Models) and Amazon Q Developer (a generative AI-powered assistant). You will use Amazon Q to help generate commands and solve coding challenges while directly interacting with Bedrock to test model inference parameters like Temperature and Top P.


Prerequisites

Before starting this lab, ensure you have the following:

  • AWS Account: Active account with Administrator or PowerUser access.
  • CLI Tools: AWS CLI v2 installed and configured (aws configure).
  • IAM Permissions: Policies allowing bedrock:InvokeModel, bedrock:ListFoundationModels, and access to Amazon Q Developer.
  • Prior Knowledge: Basic familiarity with terminal commands and JSON structures.

[!IMPORTANT] Amazon Bedrock models are not enabled by default. You must request access to the models in your AWS region before invoking them.


Learning Objectives

By completing this lab, you will be able to:

  1. Navigate and use Amazon Q Developer to generate accurate AWS CLI commands.
  2. Enable and query Foundation Models (FMs) using Amazon Bedrock.
  3. Understand and manipulate inference parameters (Temperature, Top P) to control model determinism.
  4. Analyze the JSON request and response payloads of Bedrock's InvokeModel API.

Architecture Overview

The following diagram illustrates the flow of our lab. We will use Amazon Q as our intelligent assistant to formulate the correct API calls, which we will then send to Amazon Bedrock to interact with the Titan Foundation Model.

Loading Diagram...

Step-by-Step Instructions

Step 1: Request Model Access

Before using Amazon Bedrock, you must request access to the specific Foundation Models you intend to use. For this lab, we will use Amazon Titan Text G1 - Lite.

bash
aws bedrock list-foundation-models --by-output-modality TEXT --query "modelSummaries[?modelId=='amazon.titan-text-lite-v1']"
Console alternative (Required for first-time setup)
  1. Open the AWS Management Console and navigate to Amazon Bedrock.
  2. In the left navigation pane, scroll down to Model access.
  3. Click Enable specific models (or Manage model access).
  4. Check the box next to Titan Text G1 - Lite under the Amazon provider.
  5. Click Request model access and wait for the status to change to "Access granted".

📸 Screenshot: Look for the green "Access granted" badge next to the model name.

[!TIP] Some models (like Anthropic Claude) require submitting an additional use-case justification form. Amazon Titan models are usually granted instantly.

Step 2: Use Amazon Q Developer to Generate Commands

Instead of memorizing complex CLI syntax, let's ask Amazon Q to help us figure out how to invoke the Bedrock model.

If you have the Amazon Q CLI integration installed in your terminal, or if you are using the Amazon Q chat pane in the AWS Console, type the following prompt:

"Write an AWS CLI command to invoke the amazon.titan-text-lite-v1 model in Amazon Bedrock to explain what Generative AI is. Output the response to a file called response.txt."

Amazon Q should provide an explanation and a command similar to the one we will use in the next step.

Step 3: Invoke the Foundation Model

Now, let's execute the command to query the Foundation Model. We will pass a JSON payload containing our prompt and configuration.

bash
aws bedrock-runtime invoke-model \ --model-id amazon.titan-text-lite-v1 \ --content-type application/json \ --accept application/json \ --body '{"inputText": "Explain generative AI in two sentences.", "textGenerationConfig": {"temperature": 0.1, "topP": 0.9}}' \ response.json
Console alternative
  1. In the Bedrock console, go to Playgrounds > Text.
  2. Select Amazon as the category and Titan Text G1 - Lite as the model.
  3. Type "Explain generative AI in two sentences." in the prompt area.
  4. Click Run to see the output.

Step 4: Inspect the Output and Adjust Parameters

The output is saved to response.json. Let's inspect it.

bash
cat response.json

[!NOTE] The inference parameters we used were temperature: 0.1 and topP: 0.9. A low temperature (closer to 0) makes the model more deterministic and focused.

Let's experiment by increasing the temperature to make the model more creative (and potentially less predictable).

bash
aws bedrock-runtime invoke-model \ --model-id amazon.titan-text-lite-v1 \ --content-type application/json \ --accept application/json \ --body '{"inputText": "Write a creative poem about cloud computing.", "textGenerationConfig": {"temperature": 0.9, "topP": 1.0}}' \ creative_response.json

Review the new output using cat creative_response.json.


Checkpoints

Verify that you have successfully completed the core lab steps:

  1. Check 1: Run aws bedrock list-foundation-models --region <YOUR_REGION> | grep titan-text-lite-v1. Does it return the model ID?
  2. Check 2: Inspect the response.json file. It should contain a structured JSON response with a results array and an outputText field containing the generated text.

Visualizing the Parameter Effect

Below is a conceptual diagram illustrating how parameters like Temperature and Top-P influence the model's token selection process during inference.

Compiling TikZ diagram…
Running TeX engine…
This may take a few seconds

Teardown

Because this lab primarily uses serverless, on-demand inference, there are no long-running EC2 instances or provisioned endpoints to delete. However, to keep your environment clean, remove the local files generated.

bash
# Remove generated JSON files rm response.json creative_response.json

[!WARNING] If you configured Provisioned Throughput for Amazon Bedrock (not covered in this lab but possible in production), you must delete it via the console or CLI, as it incurs high hourly charges.


Troubleshooting

Error MessageCauseFix
AccessDeniedException: You don't have access to the model...Model access has not been granted in your region.Navigate to the Bedrock console > Model access, and request access to Titan Text G1 - Lite.
UnrecognizedClientException: The security token included in the request is invalid.AWS CLI credentials are not configured or expired.Run aws configure or refresh your temporary session tokens.
ValidationException: The provided model identifier is invalid.Typo in the --model-id parameter.Ensure you are using amazon.titan-text-lite-v1 exactly as written.

Stretch Challenge

Now that you understand the CLI interactions, try automating this process with code!

Goal: Write a Python script using the boto3 library that accepts a user string, sends it to the same Titan model, and prints only the outputText string to the console (omitting the rest of the JSON wrapper).

Constraint: You must use Amazon Q Developer (either in your IDE or the console) to help you write the code.

Click here to reveal a solution
python
import boto3 import json def invoke_titan(prompt): client = boto3.client('bedrock-runtime', region_name='us-east-1') payload = { "inputText": prompt, "textGenerationConfig": { "temperature": 0.7, "topP": 0.9 } } response = client.invoke_model( modelId='amazon.titan-text-lite-v1', contentType='application/json', accept='application/json', body=json.dumps(payload) ) response_body = json.loads(response.get('body').read()) # Extract just the text from the Titan response structure print(response_body['results'][0]['outputText']) invoke_titan("What are the benefits of AWS Bedrock?")

Cost Estimate

This lab is extremely cost-effective and designed to be accessible:

  • Amazon Q Developer: The Free Tier allows up to 50 chat interactions per month. This lab uses 1-2 interactions.
  • Amazon Bedrock: On-demand inference is charged per 1,000 input/output tokens. The Titan Text Lite model costs fractions of a cent ($0.0003 per 1K input tokens). Completing this lab will cost less than $0.01.

Concept Review

To solidify your understanding, here is a breakdown of the AWS generative AI tools mentioned in this lab and the study material:

Service / ToolPrimary Use CaseTarget AudienceKey Feature
Amazon BedrockBuilding and scaling generative AI applications.Developers & Data EngineersSingle API access to multiple Foundation Models (Claude, Titan, Llama).
Amazon Q DeveloperSoftware development, debugging, and cloud infrastructure management.Developers & IT ProfessionalsIDE integrations, CLI assistance, and legacy code upgrading.
Amazon Q BusinessEnterprise assistant for internal knowledge search and workflow automation.All EmployeesConnects to 40+ enterprise data sources with built-in access controls.
Hands-On Lab894 words

Hands-On Lab: Building with Amazon Bedrock and Amazon Q

Amazon Bedrock and Amazon Q

Read full article

Hands-On Lab: Building with Amazon Bedrock and Amazon Q

Welcome to this guided hands-on lab! In this session, you will explore AWS's primary generative AI platforms. You will invoke foundation models using Amazon Bedrock, experiment with inference parameters, and interact with Amazon Q Developer to accelerate your cloud workflows.


Prerequisites

Before starting this lab, ensure you have the following:

  • AWS Account: Access to an AWS account with AdministratorAccess or sufficient IAM permissions to use Amazon Bedrock and Amazon Q.
  • CLI Tools: The AWS CLI installed and configured (aws configure) with your access keys.
  • Region Selection: Set your default region to us-east-1 (N. Virginia) or us-west-2 (Oregon), as these regions have the broadest Amazon Bedrock model availability.
  • Prior Knowledge: Basic familiarity with navigating the AWS Management Console and executing terminal commands.

Learning Objectives

By completing this lab, you will be able to:

  1. Enable and configure foundation model access in Amazon Bedrock.
  2. Invoke an Amazon Bedrock foundation model using both the AWS CLI and the AWS Management Console.
  3. Adjust inference parameters (like Temperature) to alter model creativity.
  4. Utilize Amazon Q Developer to ask AWS-specific architectural questions and generate code.

Architecture Overview

The following diagram illustrates the two distinct workflows you will execute in this lab:

Loading Diagram...

Step-by-Step Instructions

Step 1: Request Model Access in Amazon Bedrock

Before you can use any foundation model in Amazon Bedrock, you must explicitly request access to it. This is a one-time setup step per region.

📸 Screenshot: The "Model access" page in the Amazon Bedrock console, showing "Manage model access".

bash
# Check currently available foundation models in your region aws bedrock list-foundation-models --query "modelSummaries[*].modelId" --output table
Console alternative (Required for granting access)
  1. Log in to the AWS Management Console and search for Amazon Bedrock.
  2. In the left navigation pane, scroll down to Bedrock configurations and click Model access.
  3. Click the Manage model access button at the top right.
  4. Check the box next to Amazon Titan Text G1 - Lite (or similar available Titan Text model).
  5. Scroll to the bottom and click Save changes.
  6. Wait for the Access status to change to Access granted.

[!TIP] Model access is granted almost instantly for Amazon's own models (like Titan), but third-party models (like Anthropic Claude) may require submitting a use-case details form.

Step 2: Invoke a Model Using Amazon Bedrock

Now that you have access, we will send a prompt to the Amazon Titan model using the AWS CLI.

📸 Screenshot: Terminal window showing a successful JSON response from the Bedrock API.

bash
# Invoke the Titan Text Lite model aws bedrock-runtime invoke-model \ --model-id amazon.titan-text-lite-v1 \ --body '{"inputText": "Explain cloud computing in two sentences.", "textGenerationConfig": {"temperature": 0.0}}' \ --cli-binary-format raw-in-base64-out \ --accept application/json \ --content-type application/json \ response.json # View the output cat response.json
Console alternative
  1. In the Amazon Bedrock console, under Playgrounds in the left menu, select Text.
  2. Click Select model, choose Amazon, and select Titan Text G1 - Lite.
  3. Type Explain cloud computing in two sentences. in the prompt box.
  4. Click Run to generate the response.

Step 3: Experiment with Temperature Settings

In Amazon Bedrock, adjusting the temperature affects the randomness of the model's responses. A temperature of 0.0 makes responses more deterministic, while higher values (up to 1.0) increase creativity and randomness.

bash
# Invoke the model with a higher temperature for a more creative response aws bedrock-runtime invoke-model \ --model-id amazon.titan-text-lite-v1 \ --body '{"inputText": "Write a creative haiku about cloud computing.", "textGenerationConfig": {"temperature": 0.9}}' \ --cli-binary-format raw-in-base64-out \ --accept application/json \ --content-type application/json \ creative_response.json # View the creative output cat creative_response.json

Here is a visual representation of how temperature impacts token selection:

Loading Diagram...

Step 4: Interact with Amazon Q Developer

Amazon Q Developer is tightly integrated into the AWS Console to help you understand services, troubleshoot errors, and write code.

📸 Screenshot: The Amazon Q chat panel docked on the right side of the AWS Management Console.

  1. Open the AWS Management Console in your browser.
  2. On the right side of the screen, click the Amazon Q icon (a colorful letter Q).
  3. In the chat interface, type the following prompt:
text
What is Amazon EC2, and how does it work? Please list all my running EC2 instances in the us-east-1 region.

[!NOTE] Amazon Q can inspect your active AWS environment (if permissions allow) to answer specific questions about your running resources and costs.


Checkpoints

After completing the steps, verify your progress:

  1. Verify Bedrock CLI Access: Run aws bedrock list-foundation-models | grep amazon.titan-text-lite-v1. Expected Output: The console should return details of the Titan Text Lite model.

  2. Verify Bedrock Invocation: Run cat response.json. Expected Output: A JSON string containing a "results" array with the generated text about cloud computing.

  3. Verify Amazon Q Interaction: Expected Output: The Amazon Q side panel should provide an explanation of EC2 and a list/summary of your running instances.


Clean-Up / Teardown

While Amazon Bedrock charges on a pay-per-request basis (meaning no resources are actively left running), it is best practice to clean up local files. If you signed up for the Amazon Q Developer Pro tier ($19/month), remember to downgrade if you do not wish to be billed.

[!WARNING] Remember to run the teardown commands to avoid clutter and potential ongoing charges if premium subscriptions were activated.

bash
# Remove local output files generated during the lab rm response.json creative_response.json

To cancel Amazon Q Developer Pro (if enabled):

  1. Navigate to the Amazon Q Developer console.
  2. Select Subscriptions and choose to downgrade to the free tier.

Troubleshooting

Error MessageLikely CauseSolution
AccessDeniedExceptionIAM user lacks permissions, or Model Access was not granted in the Bedrock console.Complete Step 1 to request Model Access. Ensure your IAM user has bedrock:InvokeModel permissions.
ValidationExceptionThe JSON payload in the --body parameter is malformed.Check for missing quotes or brackets. Ensure you use --cli-binary-format raw-in-base64-out.
ThrottlingExceptionToo many API requests made in a short period.Wait a few seconds and retry the command.
UnrecognizedClientExceptionAWS CLI is not configured correctly.Run aws configure and provide your valid access keys and default region.
Hands-On Lab875 words

Hands-On Lab: Getting Started with Amazon Bedrock and Amazon Q Developer

Amazon Bedrock and Amazon Q

Read full article

Hands-On Lab: Getting Started with Amazon Bedrock and Amazon Q Developer

Welcome to this guided hands-on lab! In this session, you will explore AWS's primary generative AI platforms. You will learn how to enable and interact with Foundation Models (FMs) using Amazon Bedrock, understand the impact of inference parameters like Temperature, and use Amazon Q Developer as your intelligent coding and AWS assistant.

Prerequisites

Before you begin, ensure you have the following:

  • An active AWS Account with Administrator or PowerUser access.
  • The AWS CLI installed and configured on your local machine (aws configure).
  • IAM Permissions allowing AmazonBedrockFullAccess.
  • Basic familiarity with JSON and terminal commands.

[!WARNING] Some Amazon Bedrock Foundation Models incur charges based on the number of input and output tokens processed. Remember to follow the teardown instructions to remove any files, though simply having model access enabled does not accrue hourly charges.

Learning Objectives

By completing this lab, you will be able to:

  1. Request and configure access to Foundation Models in Amazon Bedrock.
  2. Invoke a generative AI model directly via the AWS CLI to generate text.
  3. Adjust inference parameters (Temperature and Top P) to control model output.
  4. Leverage Amazon Q Developer to ask AWS-specific architectural questions.

Architecture Overview

The following diagram illustrates how you will interact with both Amazon Bedrock and Amazon Q during this lab.

Loading Diagram...

Step-by-Step Instructions

Step 1: Request Model Access in Amazon Bedrock

Before you can use a Foundation Model in Amazon Bedrock, you must explicitly request access to it. This ensures you review and accept the End User License Agreement (EULA) for the specific model provider.

bash
# Note: Due to EULA acceptance requirements, model access # must initially be requested via the AWS Console. aws bedrock list-foundation-models --by-provider Amazon --query "modelSummaries[*].modelId"
Console alternative (REQUIRED for first-time setup)
  1. Log in to the AWS Management Console and navigate to Amazon Bedrock.
  2. In the left navigation pane, select Model access.
  3. Click the Manage model access button.
  4. Check the box next to Titan Text G1 - Lite (under the Amazon provider).
  5. Click Request model access at the bottom of the page.
  6. Wait for the Access status to change to Access granted.

📸 Screenshot: Model Access page showing "Access granted" next to Amazon Titan.

[!TIP] Amazon Titan models are typically granted instantly. Third-party models like Anthropic Claude may require additional use-case details to be submitted.

Step 2: Invoke a Foundation Model via CLI

Now that you have access, let's invoke the model to generate a response. We will pass a simple prompt asking the model to explain cloud computing.

bash
aws bedrock-runtime invoke-model \ --model-id amazon.titan-text-lite-v1 \ --body '{"inputText": "Explain the concept of Generative AI in one short sentence.", "textGenerationConfig": {"maxTokenCount": 50, "temperature": 0.5}}' \ --cli-binary-format raw-in-base64-out \ --accept "application/json" \ --content-type "application/json" \ output.txt
Console alternative
  1. In the Amazon Bedrock console, go to Playgrounds > Text.
  2. Click Select model and choose Amazon > Titan Text G1 - Lite.
  3. Type your prompt in the chat box.
  4. Click Run to see the generated response.

Step 3: Experiment with Inference Parameters

Generative AI models use parameters like temperature and topP to control the randomness and creativity of the output.

Loading Diagram...

Run the model again, but this time set the temperature to 0.0 for a highly deterministic response.

bash
aws bedrock-runtime invoke-model \ --model-id amazon.titan-text-lite-v1 \ --body '{"inputText": "Write a haiku about cloud computing.", "textGenerationConfig": {"temperature": 0.0}}' \ --cli-binary-format raw-in-base64-out \ output_deterministic.txt

Step 4: Consult Amazon Q Developer

Amazon Q Developer is your AI assistant for software development and AWS knowledge. Let's use it to understand the invoke-model command we just ran.

bash
# If you have the Amazon Q CLI installed: q "What does the --cli-binary-format raw-in-base64-out flag do in the AWS CLI?"
Console alternative
  1. Look for the Amazon Q icon on the right-hand sidebar of the AWS Management Console.
  2. Open the chat panel.
  3. Ask: "Why do I need to use --cli-binary-format raw-in-base64-out when calling Amazon Bedrock from the CLI?"
  4. Review Amazon Q's response, which will explain that it prevents the AWS CLI from interpreting the binary output incorrectly, treating the JSON payload correctly.

📸 Screenshot: Amazon Q chat panel with the response and source citations.

Checkpoints

Verify your progress by running the following checks:

Checkpoint 1: Read the Model Output

bash
cat output.txt

Expected Result: A JSON response containing a results array with the generated text explaining Generative AI.

Checkpoint 2: Verify Deterministic Output

bash
cat output_deterministic.txt

Expected Result: A JSON response containing a short 3-line poem (haiku) about cloud computing.

Clean-Up / Teardown

Because Amazon Bedrock models are serverless and charged per-token, you are not charged for idle time. However, it is good practice to clean up your local files.

bash
# Remove the generated output files rm output.txt output_deterministic.txt # Optional: Verify files are deleted ls -l output*.txt

[!WARNING] If you configured Provisioned Throughput for Amazon Bedrock (not covered in this basic lab), you must delete it in the console to avoid significant ongoing hourly charges.

Troubleshooting

Common ErrorCauseSolution
AccessDeniedExceptionYou did not request access to the Foundation Model.Go to Bedrock Console > Model Access and request access to Amazon Titan.
ValidationExceptionMalformed JSON in the --body parameter.Ensure you are using single quotes around the entire JSON body and double quotes for keys/values.
UnrecognizedClientExceptionYour AWS CLI is not configured with valid credentials.Run aws configure and input your Access Key and Secret Key.
Could not connect to the endpoint URLBedrock might not be supported in your default region.Append --region us-east-1 to your AWS CLI commands.

Concept Review

FeatureAmazon BedrockAmazon Q Developer
Primary Use CaseBuilding GenAI applications via APIsAssisting developers with coding and AWS architecture
InterfaceAPI, CLI, AWS Console PlaygroundsIDE Plugin, Terminal CLI, AWS Console Sidebar
CustomizationFine-tuning, RAG, Knowledge BasesOrganizational context, codebase indexing
Pricing ModelPay per input/output tokenFree tier available, Pro subscription per user

Stretch Challenge

Want to test your skills? Try using Amazon Q Developer to write a Python script (using boto3) that automates Step 2. Then, run the Python script to invoke the amazon.titan-text-lite-v1 model without using the AWS CLI directly.

Show solution
python
import boto3 import json client = boto3.client('bedrock-runtime', region_name='us-east-1') payload = { "inputText": "What are the benefits of AWS?", "textGenerationConfig": {"temperature": 0.7} } response = client.invoke_model( modelId='amazon.titan-text-lite-v1', contentType='application/json', accept='application/json', body=json.dumps(payload) ) response_body = json.loads(response['body'].read()) print(response_body['results'][0]['outputText'])
Curriculum Overview765 words

Curriculum Overview: Applying Natural Language Processing Services

Apply Natural Language Processing services

Read full article

Curriculum Overview: Apply Natural Language Processing Services

[!IMPORTANT] Natural Language Processing (NLP) bridges the gap between human communication and machine understanding. In the AWS ecosystem, pre-trained AI services allow you to integrate powerful NLP capabilities without requiring deep data science expertise.

Prerequisites

Before diving into this curriculum, learners should have a foundational understanding of the following concepts:

Prerequisite AreaRequired Knowledge
Cloud ComputingBasic familiarity with AWS global infrastructure, IAM, and API integration.
Data TypesUnderstanding the difference between structured data (databases) and unstructured data (text, emails, documents).
AI/ML FundamentalsBasic distinction between Artificial Intelligence, Machine Learning, Deep Learning, and Generative AI.

Module Breakdown

This curriculum is structured to take you from foundational text processing theories to deploying fully managed AWS NLP services.

ModuleTopicDifficultyEstimated TimeFocus Area
1NLP Foundations & Text PreprocessingBeginner2 HoursCleaning data, Lemmatization, Stemming, Stopwords
2Evolution of NLP ModelsIntermediate2 HoursBag-of-Words, TF-IDF, Word Embeddings, Transformers
3AWS Managed NLP ServicesIntermediate3 HoursComprehend, Lex, Polly, Translate, Transcribe, Kendra
4Real-World Architecture & IntegrationAdvanced3 HoursCombining services, API deployment, RAG patterns
Click to expand: Why focus heavily on Preprocessing?

To deal with the complexities of human language, NLP involves considerable text processing. Before an AI model can effectively analyze data, the dataset must be cleaned. Techniques like removing punctuation and stopwords significantly reduce the computational load and improve accuracy by filtering out "noise."

Learning Objectives per Module

Module 1: NLP Foundations & Text Preprocessing

  • Define Natural Language Processing and its role in modern AI applications.
  • Differentiate between Stemming and Lemmatization in reducing words to their root forms.
  • Apply lowercasing, stopword removal, and punctuation removal to unstructured text datasets.

[!TIP] Stemming vs. Lemmatization

  • Stemming chops off the ends of words (e.g., "running" \rightarrow "run"). It is fast but can be inaccurate.
  • Lemmatization transforms a word to its dictionary root or lemma considering context. It is slower but more accurate.

Module 2: Evolution of NLP Models

  • Trace the history of NLP from statistical methods to modern neural architectures.
  • Understand how text is converted to numerical formats (Word2Vec, GloVe).
  • Explain the role of Transformer architectures and self-attention mechanisms in Large Language Models (LLMs) like GPT and BERT.

vkingvman+vwomanvqueen\vec{v}_{\text{king}} - \vec{v}_{\text{man}} + \vec{v}_{\text{woman}} \approx \vec{v}_{\text{queen}} Equation: A conceptual representation of semantic word embeddings.

Module 3: AWS Managed NLP Services

  • Select the appropriate AWS service for specific text analysis tasks.
  • Configure Amazon Comprehend to extract entities, key phrases, and sentiment.
  • Design conversational interfaces using Amazon Lex.
  • Implement intelligent document search using Amazon Kendra.

Module 4: Real-World Architecture & Integration

  • Integrate multiple AWS AI services (e.g., Transcribe + Comprehend) into a cohesive pipeline.
  • Assess business value and determine when a managed AI service is preferable to training a custom model.

Visual Anchors

AWS NLP Service Selection Flowchart

Loading Diagram...

The Text Processing Pipeline

Compiling TikZ diagram…
Running TeX engine…
This may take a few seconds

Success Metrics

To know you have mastered this curriculum, you should be able to achieve the following success metrics:

  1. Architecture Design: Successfully draw an architecture diagram matching a business scenario to the correct AWS NLP services without referencing documentation.
  2. Vocabulary Mastery: Clearly articulate the difference between intelligent document processing (IDP) and natural language processing (NLP).
    • Check: IDP automates data extraction from business documents (Amazon Textract), while NLP handles broader text-processing and linguistic comprehension tasks (Amazon Comprehend).
  3. Exam Readiness: Consistently score 85%+ on practice questions related to the AWS Certified AI Practitioner (AIF-C01) NLP domain.
  4. Hands-on Validation: Deploy a basic Amazon Lex chatbot that successfully triggers an AWS Lambda function to return a specific intent response.

Real-World Application

Natural Language Processing is no longer confined to academic research; it is actively transforming industries. Modern applications, such as customer service chatbots, now provide real-time responses by deploying sophisticated Transformer models in online inference settings.

Career Impact: Software engineers and application developers—even those without deep ML expertise—can leverage AWS's intuitive APIs to bring powerful NLP capabilities to market. This reduces development time from months to days.

Common Industry Use Cases:

  • Healthcare: Using Amazon Comprehend Medical to extract medical ontologies and patient data from unstructured clinical notes.
  • Customer Service: Modernizing contact centers by chaining Amazon Transcribe (speech-to-text) with Amazon Comprehend (sentiment analysis) to evaluate caller frustration in real-time.
  • Global Commerce: Utilizing Amazon Translate and Amazon Polly to instantly localize product listings and provide multilingual accessibility features.
Curriculum Overview895 words

Curriculum Overview: Apply Natural Language Processing Services

Apply Natural Language Processing services

Read full article

Prerequisites

Before diving into the application of Natural Language Processing (NLP) services on AWS, learners must possess foundational knowledge in the following areas:

  • Cloud Fundamentals: Familiarity with basic AWS infrastructure, including IAM (Identity and Access Management) roles, permissions, and the AWS shared responsibility model.
  • Basic AI/ML Concepts: Understanding of the differences between artificial intelligence, machine learning, deep learning, and generative AI.
  • Data Literacy: Ability to differentiate between structured, unstructured, labeled, and unlabeled data. (NLP primarily deals with unstructured text data).
  • Foundational Generative AI: Basic comprehension of generative models, tokens, embeddings, and transformer architectures.

Module Breakdown

This curriculum is structured to take you from the raw fundamentals of text processing to deploying enterprise-grade, AI-powered NLP services using AWS.

ModuleTopicDifficultyEstimated TimeKey Focus
Module 1The NLP Preprocessing PipelineBeginner2 HoursCleaning dataset text for AI consumption (Lemmatization, Stemming, Stopwords).
Module 2Evolution of Text RepresentationsIntermediate2 HoursMoving from Bag-of-Words (BoW) to vector embeddings and Transformer models.
Module 3AWS Managed NLP ServicesBeginner/Intermediate3 HoursSelecting and applying Amazon Comprehend, Lex, Polly, Translate, and Transcribe.
Module 4Enterprise Search & LLMsAdvanced3 HoursUtilizing Amazon Kendra for intelligent search and Amazon Bedrock for generative NLP tasks.

The Learning Path

Loading Diagram...

Learning Objectives per Module

Module 1: The NLP Preprocessing Pipeline

  • Objective: Prepare unstructured text data for machine learning models using standard linguistic techniques.
  • Key Concept - Lemmatization: Transforming a word to its meaningful root (lemma) by removing affixes.
    • Real-World Example: A search engine converting the query "running shoes" to "run shoe" to match a broader set of relevant retail listings.
  • Key Concept - Stemming: Chopping off the ends of words without considering context.
    • Real-World Example: An automated spam filter reducing "runner", "runs", and "running" all to the crude root "run" to quickly flag suspicious patterns.
  • Key Concept - Stopword Removal: Filtering out words that add little semantic meaning (e.g., "the", "is", "at").
    • Real-World Example: Truncating a customer review from "The food is great" to "food great" to speed up database processing without losing the core sentiment.

Comparison: Stemming vs. Lemmatization

FeatureStemmingLemmatization
ApproachRule-based string truncationDictionary/Context-based root matching
SpeedFaster, highly efficientSlower, requires more compute
AccuracyLower (can result in non-words)Higher (preserves actual word meaning)
Best ForMassive, fast text classificationHigh-precision search engines and chatbots

Module 2: Evolution of Text Representations

  • Objective: Trace the historical progression of NLP and understand how modern AI interprets text.
  • Techniques: Understand Bag-of-Words (BoW) and Term Frequency-Inverse Document Frequency (TF-IDF).
  • Embeddings: Explain how models like Word2Vec and GloVe convert text into continuous mathematical vectors (vRnv \in \mathbb{R}^n) to capture semantic relationships.
  • Transformers: Describe how self-attention mechanisms in transformer architectures paved the way for Large Language Models (LLMs) like GPT and BERT.

Module 3: AWS Managed NLP Services

  • Objective: Choose and implement the correct purpose-built AWS AI service for a specific business problem.
  • Amazon Comprehend: Extract relationships, entities, and sentiment from unstructured text.
    • Real-World Example: Automatically tagging incoming support tickets as "Angry" or "Happy" to prioritize customer service responses.
  • Amazon Lex: Build conversational interfaces (chatbots) using voice and text.
    • Real-World Example: Powering the self-service chatbot on a bank's website to help users check their account balances.
  • Amazon Polly & Transcribe: Convert text-to-speech (Polly) and speech-to-text (Transcribe).
  • Amazon Translate: Implement highly accurate, neural machine translation across different languages.

Module 4: Enterprise Search & LLMs

  • Objective: Deploy advanced retrieval and generative solutions.
  • Amazon Kendra: Provide highly accurate, AI-powered enterprise search.
    • Real-World Example: An internal corporate portal where employees can type natural language questions (e.g., "What is the maternity leave policy?") and receive exact answers extracted from hundreds of HR PDFs.
  • Amazon Bedrock: Access and fine-tune foundation models to customize NLP applications using your organization's private data via Retrieval-Augmented Generation (RAG).

Success Metrics

To know you have mastered this curriculum, you should be able to:

  1. Map Use Cases to AWS Services: Given a business scenario, correctly select between Comprehend, Lex, Textract, and Kendra with 100% accuracy.
  2. Architect an NLP Pipeline: Successfully sketch a data flow from raw text collection cleaning(lowercasing,stopwordremoval)\rightarrow cleaning (lowercasing, stopword removal) \rightarrow inference via an AWS API.
  3. Differentiate AI Categories: Clearly explain the difference between Natural Language Processing (understanding text) and Intelligent Document Processing (automating data extraction from visual documents like PDFs).
  4. Evaluate Costs and Constraints: Assess the tradeoffs of using a simple pre-trained service (like Amazon Translate) versus fine-tuning a Large Language Model on Amazon Bedrock.

Real-World Application

Natural Language Processing is no longer an academic exercise; it is the backbone of modern enterprise automation and customer engagement.

Consider a globally distributed e-commerce company. They receive thousands of customer service calls and emails daily. By applying AWS NLP services, they can completely automate their pipeline:

Loading Diagram...

Career Impact

For a software engineer or data professional, mastering these NLP services means you can integrate state-of-the-art AI into applications without needing a PhD in machine learning. Whether you are modernizing a contact center, generating dynamic product recommendations, or building automated compliance checks, AWS NLP tools drastically reduce the time-to-market for cutting-edge features.

Curriculum Overview685 words

AWS Certified AI Practitioner (AIF-C01) Curriculum Overview

AWS Certified AI Practitioner (AIF-C01)

Read full article

AWS Certified AI Practitioner (AIF-C01) Curriculum Overview

This document provides a comprehensive roadmap for the AWS Certified AI Practitioner (AIF-C01) certification. This foundational-level credential validates your ability to recognize opportunities for AI/ML and implement them responsibly using AWS services.

## Prerequisites

Unlike Associate or Professional level certifications, the AIF-C01 is a Foundational exam.

  • Prior Experience: No technical background or prior AWS experience is required. It is designed for individuals from both technical and non-technical backgrounds.
  • Age Requirements: Candidates must be at least 13 years old (with parental consent for those 13–17).
  • Recommended Knowledge: A basic understanding of IT services and their cloud-based applications is helpful but not mandatory.

AWS Certification Path

Loading Diagram...

## Module Breakdown

The curriculum is structured into five core domains, moving from basic definitions to complex ethical and security frameworks.

DomainWeightingKey Focus Area
1. AI and ML Fundamentals~20-25%Core terminology, ML lifecycle, and supervised/unsupervised learning.
2. Fundamentals of GenAI~20-25%Large Language Models (LLMs), Transformers, and tokenization.
3. Applications of Foundation Models~15-20%Retrieval Augmented Generation (RAG), Prompt Engineering, and Model Tuning.
4. Guidelines for Responsible AI~15-20%Bias detection, fairness, transparency, and explainability.
5. Security & Governance~15-20%IAM, data privacy, and the AWS Shared Responsibility Model for AI.

## Learning Objectives per Module

Domain 1: Fundamentals of AI and Machine Learning

  • Terminology: Define AI, ML, Deep Learning, and Neural Networks.
  • Learning Types: Differentiate between Supervised (labeled data), Unsupervised (unlabeled data), and Reinforcement Learning (reward-based).
  • The ML Lifecycle: Understand the steps from data collection and EDA (Exploratory Data Analysis) to model monitoring and retraining.

Domain 2: Fundamentals of Generative AI

  • Core Concepts: Understand tokens, embeddings, and vectors.
  • Architecture: Identify the role of Transformers as the backbone of modern Generative AI.
  • Use Cases: Recognize applications in content creation, summarization, and code generation.

Domain 3: Applications of Foundation Models (FMs)

  • RAG (Retrieval Augmented Generation): Explain how to ground models in private data using vector databases like Amazon OpenSearch.
  • Prompt Engineering: Design effective prompts using context and logical reasoning steps.
  • Service Selection: Choose between Amazon Bedrock (serverless FMs) and Amazon SageMaker (custom ML builds).

Domain 4: Guidelines for Responsible AI

  • Bias & Fairness: Detect and mitigate bias using tools like SageMaker Clarify.
  • Transparency: Use SageMaker Model Cards for documenting model intent and performance.
  • Governance: Identify legal risks such as intellectual property infringement and hallucinations.

Domain 5: Security, Compliance, and Governance

  • Data Protection: Implement encryption at rest and in transit.
  • Infrastructure: Use AWS PrivateLink and IAM policies to restrict access to AI workloads.
  • Governance Tools: Leverage AWS Audit Manager and CloudTrail for compliance tracking.

## Success Metrics

To earn the certification, candidates must demonstrate proficiency through a proctored exam.

[!IMPORTANT] Passing Score: 700 / 1,000 (Scaled Score)

Exam Format

  • Questions: 65 total (50 scored, 15 unscored/experimental).
  • Question Types: Multiple choice, multiple response, matching, and ordering.
  • Duration: Typically 90–120 minutes.
Loading Diagram...

## Real-World Application

This certification translates theoretical knowledge into practical business value.

  • Business Leaders: Gain the vocabulary to lead AI initiatives and evaluate cost-benefit ratios (ROI).
  • Developers: Learn to integrate pre-trained models via Amazon Bedrock without needing a PhD in Data Science.
  • IT Professionals: Understand how to secure AI workloads and meet regulatory requirements (e.g., GDPR, HIPAA) within the cloud.

Example Use Case: Automated Customer Support

  1. Service: Use Amazon Lex for the chatbot interface.
  2. Intelligence: Use Amazon Bedrock to summarize previous customer interactions.
  3. Governance: Use Bedrock Guardrails to ensure the bot doesn't provide harmful or biased advice.
  4. Security: Use IAM to ensure the bot only accesses specific customer data.

[!TIP] Focus heavily on the "Responsibility" and "Security" domains (4 and 5), as these represent the "AWS way" of implementing AI, which is a major focus of the AIF-C01 exam.

More Study Notes (135)

Curriculum Overview: AWS Infrastructure for Generative AI Applications

AWS infrastructure and technologies for building GenAI applications

780 words

Hands-On Lab: Building GenAI Applications with Amazon Bedrock

AWS infrastructure and technologies for building GenAI applications

912 words

Hands-On Lab: Getting Started with AWS GenAI Infrastructure using Amazon Bedrock

AWS infrastructure and technologies for building GenAI applications

948 words

Curriculum Overview: Comparing AI, ML, Deep Learning, and GenAI

Compare AI, ML, Deep Learning, and GenAIDescribe the similarities and differences between AI, ML, GenAI, and deep learning

966 words

Curriculum Overview: Core Generative AI Concepts (AWS AIF-C01)

Core GenAI Concepts

782 words

Hands-On Lab: Core GenAI Concepts and Inference via Amazon Bedrock

Core GenAI Concepts

1,058 words

Hands-On Lab: Exploring Core GenAI Concepts with Amazon Bedrock

Core GenAI Concepts

1,058 words

Curriculum Overview: Amazon SageMaker's Role in the ML Lifecycle

Define Amazon SageMaker's role

820 words

Curriculum Overview: The Role of Amazon SageMaker in the ML Lifecycle

Define Amazon SageMaker's role

831 words

Curriculum Overview: Fundamentals of AI and Machine Learning Terminology

Define basic AI terms (for example, AI, ML, deep learning, neural networks, computer vision, natural language processing [NLP], model, algorithm, training and inferencing, bias, fairness, fit, large language models(LLMs))

863 words

Curriculum Overview: Foundational Generative AI Concepts

Define foundational GenAI concepts (for example, tokens, chunking, embeddings, vectors, prompt engineering, transformer-based LLMs, foundation models [FMs], multimodal models, diffusion models)

796 words

Curriculum Overview: Methods for Fine-Tuning Foundation Models

Define methods for fine-tuning an FM (for example, instruction tuning, adapting models for specific domains, transfer learning, continuous pre-training)

863 words

Curriculum Overview: Risks and Limitations of Prompt Engineering

Define potential risks and limitations of prompt engineering (for example, exposure, poisoning, hijacking, jailbreaking)

813 words

Curriculum Overview: Responsible Practices for AI Model Selection

Define responsible practices to select a model (for example, environmental considerations, sustainability)

923 words

Responsible AI: Practices for Sustainable Model Selection

Define responsible practices to select a model (for example, environmental considerations, sustainability)

923 words

Curriculum Overview: Retrieval-Augmented Generation (RAG) & Business Applications

Define Retrieval Augmented Generation (RAG) and describe its business applications (for example, Amazon Bedrock Knowledge Bases)

940 words

Curriculum Overview: Prompt Engineering Techniques

Define techniques for prompt engineering (for example, chain-of-thought, zero-shot, single-shot, few-shot, prompt templates)

870 words

Curriculum Overview: Concepts and Constructs of Prompt Engineering

Define the concepts and constructs of prompt engineering (for example, context, instruction, negative prompts, model latent space, prompt routing)

786 words

Amazon Bedrock Capabilities: Curriculum Overview

Describe Amazon Bedrock capabilities

863 words

Amazon Bedrock Capabilities & Foundation Models

Describe Amazon Bedrock capabilities

947 words

Secure Data Engineering for AI: Curriculum Overview

Describe best practices for secure data engineering (for example, assessing data quality, implementing privacy-enhancing technologies, data access control, data integrity)

923 words

Secure Data Engineering for AI: Curriculum Overview

Describe best practices for secure data engineering (for example, assessing data quality, implementing privacy-enhancing technologies, data access control, data integrity)

913 words

Curriculum Overview: Components of the Machine Learning Pipeline

Describe components of an ML pipeline (for example, data collection, exploratory data analysis [EDA], data pre-processing, feature engineering, model training, hyperparameter tuning, evaluation, deployment, monitoring)

851 words

Curriculum Overview: Cost Tradeoffs of AWS GenAI Services

Describe cost tradeoffs of AWS GenAI services (for example, responsiveness, availability, redundancy, performance, regional coverage, token-based pricing, provision throughput, custom models)

863 words

Curriculum Overview: AI Data Governance Strategies

Describe data governance strategies (for example, data lifecycles, logging, residency, monitoring, observation, retention)

765 words

Curriculum Overview: Data Governance Strategies for AI Systems

Describe data governance strategies (for example, data lifecycles, logging, residency, monitoring, observation, retention)

792 words

Curriculum Overview: Bias, Variance, and Responsible AI

Describe effects of bias and variance (for example, effects on demographic groups, inaccuracy, overfitting, underfitting)

895 words

Curriculum Overview: Bias, Variance, and Their Effects in Machine Learning

Describe effects of bias and variance (for example, effects on demographic groups, inaccuracy, overfitting, underfitting)

863 words

Curriculum Overview: Fundamental Concepts of MLOps

Describe fundamental concepts of ML operations (MLOps) (for example, experimentation, repeatable processes, scalable systems, managing technical debt, achieving production readiness, model monitoring, model re-training)

860 words

Curriculum Overview: Data Preparation for Fine-Tuning Foundation Models

Describe how to prepare data to fine-tune an FM (for example, data curation, governance, size, labeling, representativeness, reinforcement learning from human feedback [RLHF])

894 words

Preparing Data for Foundation Model Fine-Tuning: Curriculum Overview

Describe how to prepare data to fine-tune an FM (for example, data curation, governance, size, labeling, representativeness, reinforcement learning from human feedback [RLHF])

815 words

Methods to Use a Model in Production: Curriculum Overview

Describe methods to use a model in production (for example, managed API service, self-hosted API)

767 words

Curriculum Overview: Evaluating ML Models - Technical and Business Metrics

Describe model performance metrics (for example, accuracy, Area Under the Curve [AUC], F1 score) and business metrics (for example, cost per user, development costs, customer feedback, return on investment [ROI]) to evaluate ML models

873 words

Curriculum Overview: Human-Centered Design for Explainable AI

Describe principles of human-centered design for explainable AI

863 words

Curriculum Overview: Principles of Human-Centered Design for Explainable AI

Describe principles of human-centered design for explainable AI

863 words

Curriculum Overview: AI Governance Protocols and Strategies

Describe processes to follow governance protocols (for example, policies, review cadence, review strategies, governance frameworks such as the Generative AI Security Scoping Matrix, transparency standards, team training requirements)

811 words

Curriculum Overview: AI Governance Protocols & Security Frameworks

Describe processes to follow governance protocols (for example, policies, review cadence, review strategies, governance frameworks such as the Generative AI Security Scoping Matrix, transparency standards, team training requirements)

820 words

Curriculum Overview: Security and Privacy Considerations for AI Systems

Describe security and privacy considerations for AI systems (for example, application security, threat detection, vulnerability management, infrastructure protection, prompt injection, encryption at rest and in transit)

878 words

Curriculum Overview: Security and Privacy Considerations for AI Systems

Describe security and privacy considerations for AI systems (for example, application security, threat detection, vulnerability management, infrastructure protection, prompt injection, encryption at rest and in transit)

861 words

Curriculum Overview: Sources of ML Models and Customization Strategies

Describe sources of ML models (for example, open source pre-trained models, training custom models)

923 words

Curriculum Overview: Machine Learning Paradigms

Describe supervised learning, unsupervised learning, and reinforcement learning

862 words

Curriculum Overview: The Advantages of Generative AI

Describe the advantages of GenAI (for example, adaptability, responsiveness, simplicity)

673 words

AWS GenAI Advantages and Infrastructure: Curriculum Overview

Describe the advantages of using AWS GenAI services to build applications (for example, accessibility, lower barrier to entry, efficiency, cost-effectiveness, speed to market, ability to meet business objectives)

860 words

Curriculum Overview: AWS Infrastructure for GenAI Security and Compliance

Describe the benefits of AWS infrastructure for GenAI applications (for example, security, compliance, responsibility, safety)

811 words

Curriculum Overview: Source Citation and Data Origins

Describe the concept of source citation and documenting data origins (for example, data lineage, data cataloging, Amazon SageMaker Model Cards)

866 words

Curriculum Overview: Source Citation and Documenting Data Origins

Describe the concept of source citation and documenting data origins (for example, data lineage, data cataloging, Amazon SageMaker Model Cards)

878 words

Curriculum Overview: Transparent vs. Explainable AI Models

Describe the differences between models that are transparent and explainable and models that are not transparent and explainable

923 words

Curriculum Overview: Transparent vs. Non-Transparent AI Models

Describe the differences between models that are transparent and explainable and models that are not transparent and explainable

820 words

Curriculum Overview: Types of Data in AI Models

Describe the different types of data in AI models (for example, labeled and unlabeled, tabular, time-series, image, text, structured and unstructured)

846 words

Curriculum Overview: Mastering Inference Parameters (Temperature, Length, Top P & Top K)

Describe the effect of inference parameters on model responses (for example, temperature, input/output length)

925 words

Curriculum Overview: The Foundation Model Lifecycle

Describe the foundation model lifecycle (for example, data selection, model selection, pretraining, fine-tuning, evaluation, deployment, feedback)

863 words

Curriculum Overview: Foundation Model Training & Tuning

Describe the key elements of training an FM (for example, pre-training, fine-tuning, continuous pre-training, distillation)

925 words

Curriculum Overview: The Role of Agents in Multi-Step Tasks (Amazon Bedrock & Agentic AI)

Describe the role of agents in multi-step tasks (for example, Amazon Bedrock Agents, agentic AI, model context protocol)

822 words

Curriculum Overview: Detecting and Monitoring Bias in AWS AI Systems

Describe tools to detect and monitor bias, trustworthiness, and truthfulness (for example, analyzing label quality, human audits, subgroup analysis, Amazon SageMaker Clarify, SageMaker Model Monitor, Amazon Augmented AI [Amazon A2I])

820 words

Curriculum Overview: Detecting and Monitoring Bias & Trustworthiness in AI

Describe tools to detect and monitor bias, trustworthiness, and truthfulness (for example, analyzing label quality, human audits, subgroup analysis, Amazon SageMaker Clarify, SageMaker Model Monitor, Amazon Augmented AI [Amazon A2I])

917 words

Curriculum Overview: Tools for Transparent and Explainable AI

Describe tools to identify transparent and explainable models (for example, SageMaker Model Cards, open source models, data, licensing)

832 words

Curriculum Overview: Tools for Transparent and Explainable AI Models

Describe tools to identify transparent and explainable models (for example, SageMaker Model Cards, open source models, data, licensing)

890 words

Curriculum Overview: Types of AI Model Inferencing

Describe various types of inferencing (for example, batch, real-time)

765 words

Curriculum Overview: Design Considerations for Foundation Model Applications

Design considerations for applications that use foundation models (FMs)

845 words

Hands-On Lab: Design Considerations & Inference Parameters for Foundation Models

Design considerations for applications that use foundation models (FMs)

949 words

Hands-On Lab: Exploring Foundation Model Design Considerations with Amazon Bedrock

Design considerations for applications that use foundation models (FMs)

1,269 words

Curriculum Overview: Evaluating Foundation Model (FM) Performance

Determine approaches to evaluate FM performance (for example, human evaluation, benchmark datasets, Amazon Bedrock Model Evaluation)

820 words

Evaluating Foundation Model Performance: Curriculum Overview

Determine approaches to evaluate FM performance (for example, human evaluation, benchmark datasets, Amazon Bedrock Model Evaluation)

812 words

Curriculum Overview: Determining Business Value & Metrics for GenAI Applications

Determine business value and metrics for GenAI applications (for example, cross-domain performance, efficiency, conversion rate, average revenue per user, accuracy, customer lifetime value)

917 words

Curriculum Overview: Determining When AI/ML Solutions Are Not Appropriate

Determine when AI/ML solutions are not appropriate (for example, cost-benefit analyses, situations when a specific outcome is needed instead of a prediction)

948 words

Curriculum Overview: Aligning Foundation Models with Business Objectives

Determine whether a FM effectively meets business objectives (for example, productivity, user engagement, task engineering)

794 words

Curriculum Overview: Aligning Foundation Models with Business Objectives

Determine whether a FM effectively meets business objectives (for example, productivity, user engagement, task engineering)

929 words

Hands-On Lab: Effective Prompt Engineering Techniques with Amazon Bedrock

Effective prompt engineering techniques

1,056 words

Hands-On Lab: Mastering Effective Prompt Engineering Techniques

Effective prompt engineering techniques

1,215 words

Mastering Prompt Engineering: AWS Certified AI Practitioner Curriculum

Effective prompt engineering techniques

782 words

Curriculum Overview: Identifying Features of Responsible AI with AWS Tools

Explain how to use tools to identify features of responsible AI (for example, Amazon Bedrock Guardrails)

863 words

Curriculum Overview: Tools for Identifying Features of Responsible AI

Explain how to use tools to identify features of responsible AI (for example, Amazon Bedrock Guardrails)

917 words

AWS Managed AI/ML Services: Curriculum Overview

Explain the capabilities of AWS managed AI/ML services (for example, Amazon SageMaker AI, Amazon Transcribe, Amazon Translate, Amazon Comprehend, Amazon Lex, Amazon Polly)

820 words

Curriculum Overview: Cost Tradeoffs of Foundation Model Customization

Explain the cost tradeoffs of various approaches to FM customization (for example, pre-training, fine-tuning, in-context learning, RAG)

815 words

Curriculum Overview: Governance and Compliance for AI Systems

Governance and compliance regulations for AI systems

685 words

Hands-On Lab: Implementing Governance and Compliance for AI Workloads on AWS

Governance and compliance regulations for AI systems

1,058 words

Implementing AI Governance and Compliance Baselines on AWS

Governance and compliance regulations for AI systems

947 words

Curriculum Overview: Prompt Engineering Benefits and Best Practices

Identify and describe the benefits and best practices for prompt engineering (for example, response quality improvement, experimentation, guardrails, discovery, specificity and concision, using multiple comments)

874 words

Curriculum Overview: Evaluating FM Applications (RAG, Agents, & Workflows)

Identify approaches to evaluate the performance of applications built with FMs (for example, RAG, agents, workflows)

863 words

Curriculum Overview: Evaluating FM Applications (RAG, Agents, Workflows)

Identify approaches to evaluate the performance of applications built with FMs (for example, RAG, agents, workflows)

860 words

Curriculum Overview: AWS Governance and Regulation Compliance for AI

Identify AWS services and features to assist with governance and regulation compliance (for example, AWS Config, Amazon Inspector, AWS Audit Manager, AWS Artifact, AWS CloudTrail, AWS Trusted Advisor)

863 words

Curriculum Overview: Governance and Compliance for AI on AWS

Identify AWS services and features to assist with governance and regulation compliance (for example, AWS Config, Amazon Inspector, AWS Audit Manager, AWS Artifact, AWS CloudTrail, AWS Trusted Advisor)

822 words

Curriculum Overview: Developing GenAI Applications on AWS

Identify AWS services and features to develop GenAI applications (for example, Amazon SageMaker JumpStart, Amazon Bedrock PartyRock, Amazon Q, Amazon Bedrock Data Automation)

863 words

AWS AI Security Infrastructure: Curriculum Overview

Identify AWS services and features to secure AI systems (for example, IAM roles, policies, and permissions; encryption; Amazon Macie; AWS PrivateLink; AWS shared responsibility model)

917 words

Curriculum Overview: Securing AI Systems on AWS

Identify AWS services and features to secure AI systems (for example, IAM roles, policies, and permissions; encryption; Amazon Macie; AWS PrivateLink; AWS shared responsibility model)

875 words

Curriculum Overview: AWS Vector Database Services for Embeddings

Identify AWS services that help store embeddings within vector databases (for example, Amazon OpenSearch Service, Amazon Aurora, Amazon Neptune, Amazon RDS for PostgreSQL)

839 words

Curriculum Overview: Characteristics of Responsible AI Datasets

Identify characteristics of datasets (for example, inclusivity, diversity, curated data sources, balanced datasets)

863 words

Curriculum Overview: Dataset Characteristics & Responsible AI

Identify characteristics of datasets (for example, inclusivity, diversity, curated data sources, balanced datasets)

822 words

Curriculum Overview: Disadvantages and Risks of GenAI Solutions

Identify disadvantages of GenAI solutions (for example, hallucinations, interpretability, inaccuracy, nondeterminism)

873 words

Curriculum Overview: Real-World AI Applications & Cloud Implementations

Identify examples of real-world AI applications (for example, computer vision, NLP, speech recognition, recommendation systems, fraud detection, forecasting)

863 words

Curriculum Overview: Factors in Selecting GenAI Models

Identify factors to consider when selecting GenAI models (for example, model types, performance requirements, capabilities, constraints, compliance)

894 words

AWS Certified AI Practitioner: Features of Responsible AI - Curriculum Overview

Identify features of responsible AI (for example, bias, fairness, inclusivity, robustness, safety, veracity)

863 words

Curriculum Overview: Features of Responsible AI

Identify features of responsible AI (for example, bias, fairness, inclusivity, robustness, safety, veracity)

962 words

Curriculum Overview: Legal, Ethical, and Business Risks of Generative AI

Identify legal risks of working with GenAI (for example, intellectual property infringement claims, biased model outputs, loss of customer trust, end user risk, hallucinations)

810 words

Curriculum Overview: Legal Risks of Working with GenAI

Identify legal risks of working with GenAI (for example, intellectual property infringement claims, biased model outputs, loss of customer trust, end user risk, hallucinations)

863 words

Curriculum Overview: Generative AI Use Cases & Applications

Identify potential use cases for GenAI models (for example, image, video, and audio generation; summarization; AI assistants; translation; code generation; customer service agents; search; recommendation engines)

863 words

AWS ML Pipeline Stages & SageMaker Services Curriculum

Identify relevant AWS services and features for each stage of an ML pipeline (for example, SageMaker AI, SageMaker Data Wrangler, SageMaker Feature Store, SageMaker Model Monitor)

863 words

Curriculum Overview: AWS Services for the ML Pipeline

Identify relevant AWS services and features for each stage of an ML pipeline (for example, SageMaker AI, SageMaker Data Wrangler, SageMaker Feature Store, SageMaker Model Monitor)

685 words

Curriculum Overview: Foundation Model Evaluation Metrics

Identify relevant metrics to assess FM performance (for example, Recall-Oriented Understudy for Gisting Evaluation [ROUGE], Bilingual Evaluation Understudy [BLEU], BERTScore)

863 words

Curriculum Overview: Selection Criteria for Pre-Trained Foundation Models

Identify selection criteria to choose pre-trained models (for example, cost, modality, latency, multi-lingual, model size, model complexity, customization, input/output length, prompt caching)

863 words

Curriculum Overview: Tradeoffs Between AI Model Safety and Transparency

Identify tradeoffs between model safety and transparency (for example, measure interpretability and performance)

830 words

Curriculum Overview: Tradeoffs Between Model Safety and Transparency

Identify tradeoffs between model safety and transparency (for example, measure interpretability and performance)

863 words

Curriculum Overview: Identifying Use Cases for Amazon Q

Identify use cases for Amazon Q

830 words

Curriculum Overview: Identifying Use Cases for Amazon Q

Identify use cases for Amazon Q

786 words

Curriculum Overview: Evaluating Foundation Model (FM) Performance

Methods to evaluate foundation models (FM) performance

685 words

Hands-On Lab: Evaluating Foundation Models with Amazon Bedrock

Methods to evaluate foundation models (FM) performance

1,055 words

Hands-On Lab: Evaluating Foundation Models with Amazon Bedrock

Methods to evaluate foundation models (FM) performance

966 words

Curriculum Overview: Securing AI Systems (AWS AIF-C01)

Methods to secure AI systems

825 words

Hands-On Lab: Implementing Security, Governance, and Privacy for AI Workloads

Methods to secure AI systems

1,243 words

Hands-On Lab: Methods to Secure AI Systems on AWS

Methods to secure AI systems

1,058 words

Hands-On Lab: Navigating the AWS ML Development Lifecycle

ML development lifecycle

863 words

Hands-On Lab: Navigating the ML Development Lifecycle and Governance on AWS

ML development lifecycle

947 words

ML Development Lifecycle: Curriculum Overview

ML development lifecycle

780 words

Curriculum Overview: Practical Use Cases for AI

Practical use cases for AI

842 words

Hands-On Lab: Exploring Practical AI Use Cases on AWS

Practical use cases for AI

929 words

Hands-On Lab: Exploring Practical AI Use Cases with AWS Managed Services

Practical use cases for AI

1,056 words

Curriculum Overview: Applications and Value of AI/ML

Recognize applications where AI/ML can provide value (for example, assist human decision making, solution scalability, automation)

878 words

Curriculum Overview: Selecting AWS Services for Computer Vision

Select services for Computer Vision

863 words

Curriculum Overview: Selecting Computer Vision Services on AWS

Select services for Computer Vision

810 words

Curriculum Overview: Selecting the Appropriate ML Techniques

Select the appropriate ML techniques for specific use cases (for example, regression, classification, clustering)

943 words

Hands-On Lab: Exploring Specialized AWS AI Services

Specialized AI Services

925 words

Hands-On Lab: Exploring Specialized AWS AI Services

Specialized AI Services

1,145 words

Hands-On Lab: Exploring Specialized AWS AI Services (Comprehend & Polly)

Specialized AI Services

834 words

Specialized AI Services: Comprehensive AWS Curriculum Overview

Specialized AI Services

820 words

Curriculum Overview: Strategic Capabilities and Limitations of GenAI for Business

The capabilities and limitations of GenAI for solving business problems

820 words

Hands-On Lab: Capabilities and Limitations of GenAI for Business Solutions

The capabilities and limitations of GenAI for solving business problems

925 words

Hands-On Lab: Exploring GenAI Capabilities and Limitations via Amazon Bedrock

The capabilities and limitations of GenAI for solving business problems

1,056 words

Curriculum Overview: Developing Responsible AI Systems

The development of AI systems that are responsible

820 words

Hands-On Lab: Implementing Responsible AI with Amazon Bedrock Guardrails

The development of AI systems that are responsible

1,052 words

Hands-On Lab: Implementing Responsible AI with Amazon Bedrock Guardrails

The development of AI systems that are responsible

1,056 words

Curriculum Overview: Transparency and Explainability in AI Models

The importance of transparent and explainable models

685 words

Hands-On Lab: Implementing Transparent and Explainable AI Models with AWS

The importance of transparent and explainable models

947 words

Hands-On Lab: Implementing Transparent and Explainable Models in AWS

The importance of transparent and explainable models

948 words

Curriculum Overview: Training and Fine-Tuning Foundation Models (FMs)

the training and fine-tuning process for foundation models (FMs)

820 words

Hands-On Lab: Training and Fine-Tuning Foundation Models on AWS

the training and fine-tuning process for foundation models (FMs)

894 words

Ready to practice? Jump straight in — no sign-up needed.

Take practice tests, review flashcards, and read study notes right now.

Take a Practice Test

AWS Certified AI Practitioner (AIF-C01) Practice Questions

Try 15 sample questions from a bank of 353. Answers and detailed explanations included.

Q1medium

When evaluating Artificial Intelligence systems, which of the following best describes the distinction between transparency and explainability?

A.

Transparency involves disclosing the overall system design and data origins, whereas explainability focuses on the logic behind individual model predictions.

B.

Explainability involves disclosing the overall system design and data origins, whereas transparency focuses on the logic behind individual model predictions.

C.

Transparency and explainability are synonymous terms that both refer to the quantitative accuracy and F1F_1 score of the model during the validation phase.

D.

Transparency is a measure of the model's total training speed, while explainability is a measure of the total number of parameters (NN) within the neural network.

Show answer & explanation

Correct Answer: A

To distinguish between these terms, it is important to look at the scope of the disclosure:

  1. Transparency provides a holistic or 'system-wide' view. It focuses on the disclosure of the AI system's design, its data origins, and the overall objectives set by the developers for stakeholders to review.
  2. Explainability is more granular. It focuses on the specific reasoning or technical logic that justifies why a model arrived at a particular decision or prediction (e.g., why a specific loan application was denied).

Option B incorrectly swaps these definitions. Options C and D are incorrect because accuracy metrics, training speed, and parameter counts do not define the conceptual frameworks of transparency or explainability. Therefore, Option A is the correct distinction.

Q2medium

A financial services firm is building a customer support chatbot using Amazon Bedrock. The firm's compliance policy requires that the chatbot must not provide specific investment advice and must redact all Social Security Numbers (SSNs) from the interaction to prevent exposure of personally identifiable information (PII). If a policy violation is detected, the system must return a standardized refusal message. Which implementation strategy for Amazon Bedrock Guardrails best satisfies these requirements with the least operational overhead?

A.

Create a guardrail that defines 'Investment Advice' as a Denied Topic and enables Sensitive Information Filters for SSNs, applying the guardrail to both the user prompt and the model response.

B.

Enable Sensitive Information Filters with 'High' strength for all categories and apply it exclusively to the model's output to prevent the leakage of PII generated by the foundation model.

C.

Perform supervised fine-tuning on the foundation model using a safety-aligned dataset that includes examples of refusing to give investment advice and masking sensitive numeric patterns.

D.

Invoke the foundation model via an AWS Lambda function that uses custom regex logic to parse the text response for SSN patterns and keywords related to financial advice before delivery.

Show answer & explanation

Correct Answer: A

To meet the requirements effectively, Amazon Bedrock Guardrails should be implemented as the primary safety layer.

  1. Denied Topics: This feature allows developers to define prohibited subjects (e.g., 'Investment Advice'). Bedrock will automatically detect and block these topics based on the natural language definition provided.
  2. Sensitive Information Filters: This built-in capability uses machine learning and regular expressions to detect and redact PII like SSNs without requiring custom code or Lambda functions.
  3. Dual Application: Applying the guardrail to both the user prompt (input) and model response (output) ensures that malicious user requests are caught before reaching the model and that the model's generated output remains compliant.

Option B is insufficient because it ignores the input prompt. Option C (fine-tuning) is more complex and does not guarantee strict compliance like a guardrail. Option D introduces significant operational overhead by manually recreating features already natively available in Amazon Bedrock.

Q3medium

In a robust machine learning (ML) governance framework, which statement best explains the primary purpose and scope of documenting data origins and maintaining data lineage?

A.

To provide a comprehensive historical record of the data lifecycle—including sources, licensing, and transformations—to ensure transparency, traceability, and accountability.

B.

To automate the real-time horizontal scaling of inference endpoints based on incoming request volumes and latent traffic patterns.

C.

To optimize the mathematical convergence speed of gradient descent by selecting the most computationally efficient tensor formats for the underlying hardware.

D.

To monitor post-deployment performance metrics, such as accuracy and precision, to determine exactly when a model requires retraining.

Show answer & explanation

Correct Answer: A

Documenting data origins and data lineage serves as a foundational pillar of AI transparency and governance.

  1. Data Origins: This involves recording the specific 'how' and 'where' of data collection. It includes capturing source citations, licenses (e.g., Creative Commons or proprietary terms), and permissions to ensure legal and ethical compliance.
  2. Data Lineage: This provides a complete audit trail that maps how raw data was transformed during preprocessing (cleaning, curation, feature engineering) and which specific processes or versions of code shaped the final trained model.
  3. Governance Utility: These practices are critical for auditing systems, identifying the source of potential biases, and ensuring that every decision made by an AI model can be traced back to the specific data that informed its training.
  • Option B describes auto-scaling, which is an operational/infrastructure concern.
  • Option C describes algorithmic optimization, which focuses on training efficiency rather than documentation or governance.
  • Option D describes model monitoring (tracking data drift and performance), which happens after a model is in production, whereas lineage captures the historical development process.
Q4medium

Which of the following best describes the capabilities and application of Amazon SageMaker Clarify within the machine learning development lifecycle?

A.

It automatically cleans and rebalances datasets using synthetic data generation to remove bias without manual configuration or facet identification.

B.

It evaluates both pre-training datasets for imbalances and post-training models for disparate predictions based on user-specified sensitive facets like age or gender.

C.

Its bias detection capabilities are restricted solely to the real-time inference phase after a model has been deployed to a production endpoint.

D.

It requires developers to write custom mathematical scripts to implement fairness algorithms such as demographic parity and equal opportunity from scratch.

Show answer & explanation

Correct Answer: B

Amazon SageMaker Clarify is designed to detect bias at two critical stages of the ML lifecycle: pre-training (analyzing the raw data for imbalances across sensitive groups) and post-training (analyzing the model's predictions for disparate impact).

  1. Facets: Users define sensitive 'facets' (e.g., race, gender, age) that Clarify uses to calculate science-based fairness metrics.
  2. Explainability: It also provides feature attribution using methods like SHAP (Shapley Additive Explanations) to show how different inputs contribute to a model's decision.
  3. Reports: The tool generates visual reports to assist in transparency and compliance.

Option A is incorrect because Clarify identifies bias but does not automatically 'clean' or rebalance data without user intervention. Option C is incorrect because it works during training and evaluation, not just real-time inference. Option D is incorrect because Clarify provides built-in algorithms for fairness metrics rather than requiring manual scripting.

The correct answer is B.

Q5medium

A company is preparing for a SOC 2 audit and needs to fulfill two requirements as part of the AWS Shared Responsibility Model. First, they must provide their auditors with AWS's own ISO and SOC reports to prove the security of the underlying infrastructure. Second, they need to automate the collection of evidence from their own AWS account to show that their resource configurations meet specific compliance controls. Which combination of AWS services should the company use to meet these requirements?

A.

Use AWS Artifact to download AWS compliance reports and AWS Audit Manager to automate evidence collection for their own resources.

B.

Use AWS Audit Manager to download AWS compliance reports and AWS Artifact to monitor resource configurations.

C.

Use Amazon Inspector to review AWS's third-party certifications and AWS Config to generate the final audit-ready SOC 2 report.

D.

Use AWS Trusted Advisor to access AWS's internal audit results and AWS Artifact to automate the evidence collection process.

Show answer & explanation

Correct Answer: A

To solve this scenario, we must distinguish between services that provide AWS's internal documentation and services that audit the customer's environment.

  1. AWS Artifact is the correct service for the first requirement. It is a self-service portal that provides on-demand access to AWS's security and compliance reports (such as SOC, ISO, and PCI reports) and select online agreements. This allows the customer to prove the security of the underlying AWS infrastructure to their auditors.

  2. AWS Audit Manager is the correct service for the second requirement. It helps automate the collection of evidence for audits by mapping AWS usage and configurations (often sourced from AWS Config) to specific compliance frameworks like SOC 2, HIPAA, or PCI DSS. It generates audit-ready reports based on the evidence collected from the customer's account.

  • Distractor B is incorrect because it swaps the functions of Audit Manager and Artifact.
  • Distractor C is incorrect because Amazon Inspector is used for vulnerability scanning and network exposure analysis, not for downloading AWS's compliance certifications.
  • Distractor D is incorrect because Trusted Advisor provides best practice recommendations (cost, performance, security), not formal AWS compliance reports, and Artifact does not automate evidence collection for customer-side resources.

Therefore, the correct combination is AWS Artifact for AWS's reports and AWS Audit Manager for the customer's own evidence collection.

Q6medium

A development team is tasked with modernizing a legacy enterprise application currently running on Java 8 to a more secure and performant Java 17 environment. Simultaneously, they need to migrate a separate .NET application from a Windows-based architecture to a Linux-based environment to optimize infrastructure costs. Which of the following best describes how Amazon Q Developer for Code Transformation can be applied to assist this team?

A.

It analyzes the existing source code, generates a transformation plan, and automatically performs the code and configuration changes required to upgrade the Java version and port the .NET application to Linux.

B.

It scans the repository to identify deprecated APIs and provides a list of modernization steps, but the development team must manually rewrite the code and update build configuration files.

C.

It acts primarily as a documentation assistant that provides comparative analysis of Java 17 features, as it is unable to modify compiled back-end languages or perform cross-platform migrations.

D.

It requires the development team to first upload a manually generated target architecture map and a complete dependency graph, as the tool cannot autonomously analyze source code to determine modernization steps.

Show answer & explanation

Correct Answer: A

Amazon Q Developer for Code Transformation is designed to automate the heavy lifting of software modernization.

  1. Analysis and Planning: The service analyzes the existing codebase (such as Java 8 or 11) and generates a detailed transformation plan that identifies the necessary changes for the upgrade to a newer version like Java 17.
  2. Automatic Execution: Unlike simple scanners, it automatically performs the code transformations, updates dependencies, and modifies configuration files.
  3. Platform Migration: Its capabilities extend beyond simple version upgrades to include platform porting, such as moving .NET applications from Windows to Linux environments.
  4. Efficiency: By automating these tasks, it significantly reduces manual effort—Amazon has used this internally to migrate thousands of applications, saving an estimated thousands of years of manual developer work.

Therefore, Option A is the correct description of its end-to-end automated capabilities.

Q7medium

A data scientist is preparing a dataset for a K-Nearest Neighbors (KNN) model. The dataset contains numerical features with significantly different scales (e.g., 'Age' in years and 'Annual Income' in dollars) and several extreme outliers. Which statement correctly explains the selection and impact of feature scaling techniques for this scenario?

A.

Standardization is generally preferred over normalization because it is more robust to outliers, preventing them from compressing the majority of the data into a very narrow range.

B.

Normalization is the superior choice because it strictly bounds all feature values to the range [0,1][0, 1], which effectively eliminates the influence of outliers on distance-based calculations.

C.

Standardization should be used to transform skewed or bimodal distributions into perfectly Gaussian distributions to meet the underlying statistical assumptions of the KNN algorithm.

D.

Feature scaling is a mandatory pre-processing step for all machine learning models, including tree-based algorithms like Random Forest, to ensure the optimizer can find optimal split points.

Show answer & explanation

Correct Answer: A

In machine learning pre-processing, feature scaling is critical for distance-based algorithms like K-Nearest Neighbors (KNN).

  1. Standardization (Z-score Scaling): This technique rescales features to have a mean (Δ)of0andastandarddeviation(σ)of1usingtheformulaz=xμσΔ) of 0 and a standard deviation (σ) of 1 using the formula z = \frac{x - \mu}{\sigma}. Because it uses the mean and standard deviation, it is more robust to outliers. It does not bound the data to a specific range, meaning outliers will still exist as large z-scores but won't 'crush' the rest of the data.

  2. Normalization (Min-Max Scaling): This technique rescales features to a fixed range, typically [0,1],usingtheformulaxnorm=xxminxmaxxmin.Ifthedatasetcontainsextremeoutliers(xmax[0, 1], using the formula x_{norm} = \frac{x - x_{min}}{x_{max} - x_{min}}. If the dataset contains extreme outliers (x_{max} or xminx_{min}), the majority of the 'normal' data points will be compressed into a tiny interval (e.g., $0.0001 to $0.01), making it difficult for the model to learn patterns.

  3. Model Invariance: While scaling is vital for KNN and Gradient Descent, it is important to note that tree-based algorithms (like Random Forest) are scale-invariant and typically do not require these steps.

Therefore, Standardization is generally preferred when outliers are present.

Q8medium

An organization is migrating its data warehouse to a cloud environment and needs to implement a robust access control strategy. The security team insists on following the Principle of Least Privilege (PoLP) while also requiring dynamic, context-aware access for sensitive datasets (e.g., access based on project tags and time-of-day). Which of the following approaches best explains the appropriate application of these data access control principles?

A.

Grant administrative privileges to all developers by default to ensure they have the necessary access to troubleshoot real-time production issues without delay.

B.

Implement Attribute-Based Access Control (ABAC), as it allows for granular and dynamic permission logic using attributes like user roles, resource tags, and environmental conditions.

C.

Use Role-Based Access Control (RBAC) exclusively, as it is the only method that can handle fine-grained, context-aware access for project-specific tags and time-of-day constraints.

D.

Store all encryption keys and access credentials within the application source code to simplify the authentication process for automated data pipelines.

Show answer & explanation

Correct Answer: B

To follow the Principle of Least Privilege (PoLP) and meet the requirement for context-aware access, Attribute-Based Access Control (ABAC) is the most suitable choice.

  1. PoLP requires that users and systems are granted only the minimum permissions necessary to perform their specific tasks, which reduces the overall attack surface.
  2. ABAC uses a flexible policy model that evaluates attributes (e.g., user department, resource tags like Project: Alpha, and environment factors like Time: 0900-1700) to make access decisions. This provides more granularity than Role-Based Access Control (RBAC), which typically assigns permissions based on broad job functions without considering real-time context.
  3. Option A violates PoLP by over-provisioning permissions. Option C is incorrect because RBAC is generally less suited for dynamic, context-aware conditions compared to ABAC. Option D is a critical security failure (hardcoding credentials).

Therefore, Option B is the correct best practice.

Q9medium

A data engineer is designing a GraphRAG (Retrieval-Augmented Generation) system that requires the ability to perform similarity-based searches on graph nodes using vector embeddings while maintaining the context of graph relationships. Which Amazon Neptune feature or component provides an integrated engine that natively supports vector search capabilities, such as Hierarchical Navigable Small World (HNSW) indexing, to query these embeddings?

A.

Amazon Neptune Analytics

B.

The Neptune pgvector extension for graph schemas

C.

Neptune ML Tabular S3 Integration

D.

Gremlin Euclidean Distance Manual Predicates

Show answer & explanation

Correct Answer: A

The correct answer is Amazon Neptune Analytics. This is an integrated analytics engine for Amazon Neptune that supports vector search capabilities, allowing users to perform similarity searches using optimized indexes like HNSW alongside traditional graph queries. This is essential for building relationship-aware LLM applications (GraphRAG).

  • Amazon Neptune Analytics specifically handles the storage and similarity-based querying of vectors within the graph context.
  • pgvector is a vector search extension for PostgreSQL (used in Amazon Aurora and RDS), not for the Neptune graph database.
  • Neptune ML uses SageMaker to train Graph Neural Networks (GNNs) that generate embeddings by considering both graph topology and properties, but the integrated vector search indexing for querying those embeddings is provided by Neptune Analytics.
  • Manual distance calculations in Gremlin (Option D) do not support optimized approximate nearest neighbor (ANN) indexing required for high-performance similarity search.
Q10medium

An organization processes thousands of varied document formats daily, including invoices, legal contracts, and handwritten forms. Currently, manual data entry is leading to a high rate of errors and increasing operational costs. Which AWS service and approach should the organization identify as the most effective for automating this repetitive task while minimizing manual configuration?

A.

Implement Amazon Textract as part of an Intelligent Document Processing (IDP) workflow to automatically extract text and structured data without manual templates.

B.

Deploy a standard template-based OCR solution to ensure that new document formats are automatically recognized and processed without human intervention.

C.

Utilize Amazon Rekognition to extract and categorize complex financial data and tabular structures from PDF spreadsheets.

D.

Develop a custom neural network using Amazon SageMaker for every unique document layout to achieve the highest possible extraction accuracy.

Show answer & explanation

Correct Answer: A

To automate the extraction of structured data from varied document formats effectively, Amazon Textract is the most appropriate service. Unlike traditional Optical Character Recognition (OCR), which often relies on fixed templates for specific layouts, Amazon Textract uses machine learning to understand the structure of documents (like forms and tables) and can extract data regardless of the layout changes.

  1. Amazon Textract identifies text, handwriting, and data from tables/forms automatically.
  2. Intelligent Document Processing (IDP) refers to the broader pattern of using AI services to handle the end-to-end lifecycle of unstructured data.
  3. Option B is incorrect because traditional template-based OCR requires manual configuration for every new format.
  4. Option C is incorrect because Amazon Rekognition is primarily for image and video analysis (object detection, facial recognition), not for document-based data extraction.
  5. Option D is inefficient because building a custom model for every layout is time-consuming and negates the benefits of using a managed AI service designed for this specific task.
Q11medium

A developer is evaluating an Amazon Bedrock agent designed to handle multi-step customer service workflows. Although the agent eventually provides a correct final resolution, the developer suspects the agent is inefficient, potentially entering redundant reasoning loops or failing to pass necessary parameters to internal APIs in the correct format. Which analytical approach is most effective for diagnosing these specific failures in the agent's orchestration and tool invocation logic?

A.

Analyze the Model Trace to inspect the sequence of Thought, Action, and Observation steps and calculate Tool Usage Accuracy for API calls.

B.

Calculate the Faithfulness and Answer Relevance metrics to ensure the generated responses are grounded in the retrieved knowledge base.

C.

Evaluate the final response using ROUGE-L and BLEU scores to determine the semantic overlap with a human-verified reference dataset.

D.

Monitor Response Latency and Context Relevance to determine if delays are caused by slow retrieval from the vector database.

Show answer & explanation

Correct Answer: A

To diagnose failures in multi-step reasoning and tool usage, a developer must evaluate the intermediate steps of the agent's workflow. 1. Model Trace in Amazon Bedrock is the primary diagnostic tool for this purpose; it breaks down the orchestration process into Thought (the reasoning logic), Action (the specific tool or API call), and Observation (the result returned from the tool). This visibility allows developers to pinpoint where the logic enters an infinite loop or stops prematurely. 2. Tool Usage Accuracy is the specific metric used to verify that the agent correctly mapped user intent to the appropriate API within an Action Group and provided the required parameters in the valid format. 3. Other metrics such as Faithfulness (which measures RAG grounding) or ROUGE/BLEU (which measure text similarity) do not provide insight into the logic of the tool calls or the multi-step execution path. Correct Answer: A

Q12medium

When comparing structured and unstructured data for a machine learning project on AWS, which of the following best describes their differences in terms of storage, schema, and analytical processing?

A.

Structured data is highly organized in tabular formats with a predefined schema and is typically stored in Amazon RDS, whereas unstructured data lacks a fixed format and often requires deep learning architectures for feature extraction.

B.

Unstructured data is easier to process for initial model training because it represents raw information without the overhead of a rigid schema, while structured data requires significant embedding techniques like those used in NLP before it can be used.

C.

Amazon S3 is the primary AWS service for managing structured tabular data with strict schemas, while Amazon RDS is designed to store high-volume unstructured files such as video and audio to optimize deep learning performance.

D.

The primary distinction between structured and unstructured data is the total volume; datasets exceeding 1 Terabyte are classified as unstructured regardless of whether they follow a fixed tabular row-and-column format.

Show answer & explanation

Correct Answer: A

To compare these two data types accurately, we must look at their structure, storage, and how models interact with them:

  1. Structure and Schema: Structured data follows a strict, predefined schema (Schema-on-Write). It is organized into rows and columns, making it ideal for relational databases like Amazon RDS or data warehouses like Amazon Redshift. Unstructured data (images, video, raw text) lacks a predefined model and is often stored in data lakes like Amazon S3.
  2. Machine Learning Requirements: Because structured data is already in a numeric/tabular format, it is well-suited for classic ML algorithms like Random Forests or Logistic Regression. Unstructured data is high-dimensional and complex, typically requiring deep learning (e.g., Convolutional Neural Networks for images) to extract meaningful features.
  3. Preprocessing: Structured data allows for direct statistical manipulation. Unstructured data requires heavy preprocessing, such as converting pixels to tensors or text to embeddings, before it can be effectively used in most models.

Option B is incorrect because unstructured data is significantly harder (not easier) to process due to its lack of organization. Option C incorrectly swaps the primary use cases for RDS and S3. Option D is incorrect because the distinction is based on the format and organization of the data, not its size/volume.

Correct Answer: A

Q13medium

Which of the following best explains the mechanism and primary characteristic of an indirect prompt injection attack against a Large Language Model (LLM) application?

A.

An attacker places hidden instructions in an external data source, such as a website or document, which the LLM then processes and follows when a legitimate user asks the model to retrieve that information.

B.

An attacker gains direct access to the LLM's primary console and manually inputs commands like "Ignore all previous instructions" to override safety guardrails.

C.

An attacker modifies the model's foundational training data or fine-tuning weights to ensure the model exhibits specific biased behaviors during inference.

D.

An attacker exploits vulnerabilities in the physical hardware hosting the LLM to modify the model's memory states and bypass API-level authentication.

Show answer & explanation

Correct Answer: A

An indirect prompt injection occurs when an LLM retrieves data from an external, untrusted source (like a webpage, email, or PDF) that contains hidden malicious instructions.

  1. Mechanism: Unlike direct injection where the attacker interacts with the model directly, this attack leverages tools like Retrieval-Augmented Generation (RAG) or web-browsing plugins.
  2. The Victim: The victim is often an innocent user who asks the LLM to summarize or process the tainted external content.
  3. Execution: The LLM treats the embedded malicious instructions as part of its prompt and executes them—for example, it might be instructed to exfiltrate user data to an attacker-controlled server.

Option B describes direct prompt injection (or jailbreaking). Option C describes training data poisoning. Option D describes a hardware-level security breach, which is not a prompt injection. Option A is the correct explanation of the indirect injection pathway.

Q14medium

A financial institution is developing an AI model to automate loan approval decisions. Due to strict consumer protection regulations, the company must be able to provide clear, human-understandable explanations for any credit rejection to satisfy regulatory audits. The data science team is comparing a high-performing Deep Neural Network with a simpler Logistic Regression model. Based on the tradeoffs between model performance, safety, and transparency, which approach is most appropriate for this regulated use case?

A.

Choosing the Logistic Regression model to prioritize transparency and interpretability, even if it results in slightly lower predictive performance.

B.

Choosing the Deep Neural Network because its high parameter count allows for a more detailed and transparent explanation of complex edge cases.

C.

Implementing the Deep Neural Network and adding a differential privacy layer, as this technique increases both the transparency of the logic and the model's accuracy.

D.

Prioritizing the performance of the Deep Neural Network, as higher accuracy inherently ensures that the model's decisions are safer and easier to explain to regulators.

Show answer & explanation

Correct Answer: A

In highly regulated industries like finance, there is a fundamental performance-interpretability tradeoff. While complex models like Deep Neural Networks (DNNs) often achieve higher accuracy (performance), they function as "black boxes" with low transparency. Conversely, simpler models like Logistic Regression are highly transparent and interpretable, allowing auditors to see exactly how input features (like income or credit score) affect the outcome. For a loan approval system where the ability to explain decisions is a legal requirement, the institution must often accept the tradeoff of lower performance in exchange for higher transparency. Distractors B and D are incorrect because complexity usually decreases transparency, and accuracy does not equate to explainability. Distractor C is incorrect because differential privacy is a safety/privacy technique that typically introduces noise, which can actually reduce both accuracy and transparency.

Q15medium

An engineer is designing an autonomous agent that must perform a multi-step task: first, it must retrieve a specific configuration file from a local directory, and second, it must execute a Python script to validate that configuration. If the engineer uses the Model Context Protocol (MCP) to facilitate this workflow, which of the following best describes how the MCP architecture handles these two steps?

A.

The configuration file is accessed as an MCP Resource (read-only), while the Python script execution is performed via an MCP Tool (executable action).

B.

MCP automatically increases the LLM's internal context window to $1,000,000 tokens so that all local files and scripts are stored in the model's weights.

C.

The protocol acts as a static middleware that translates the user prompt into a hard-coded API script, bypassing the model's reasoning capabilities.

D.

MCP serves as a security-only layer that encrypts the JSON-RPC traffic between the client and server without exposing file contents to the model.

Show answer & explanation

Correct Answer: A

The Model Context Protocol (MCP) is an open standard designed to connect AI applications (clients) to external data and functions (servers).

  1. Step 1 (Retrieval): In MCP, data sources like local files, databases, or API responses are classified as Resources. These are typically read-only contexts provided to the model to inform its reasoning.
  2. Step 2 (Execution): Actions that change state or perform computations, such as running a Python script or modifying a file, are classified as Tools.
  3. Multi-step Workflow: By using a standardized protocol (often over JSON-RPC), the model can dynamically decide to 'read' a resource and then 'call' a tool based on the information it just retrieved.

Distractor Rationale: Option B incorrectly describes MCP as a model training/context window technique. Option C incorrectly suggests it is static middleware that bypasses reasoning. Option D incorrectly limits MCP's scope to simple encryption rather than functional integration.

The correct answer is A.

These are 15 of 353 questions available. Take a practice test →

AWS Certified AI Practitioner (AIF-C01) Flashcards

340 flashcards for spaced-repetition study. Showing 30 sample cards below.

Advantages of AWS Generative AI Services(5 cards shown)

Question

Lower Barrier to Entry

Answer

AWS democratizes access to advanced AI by providing managed services that don't require deep data science expertise.

[!NOTE] This allows developers, business analysts, and smaller teams to experiment with and deploy Foundation Models (FMs) without building complex systems from scratch.

Key Drivers:

  • Simpliied interfaces
  • Pre-configured tools
  • Automated workflows

Question

Speed to Market (Efficiency)

Answer

The ability to deploy generative AI applications quickly by leveraging pre-built infrastructure and managed APIs.

Loading Diagram...

Advantages:

  • Reduced development time
  • No physical hardware maintenance
  • Focus on innovation rather than infrastructure management.

Question

Cost-Effectiveness

Answer

AWS provides flexible pricing models that align expenses with actual usage, rather than requiring large upfront capital investments.

FeatureBenefit
Serverless InferenceNo idle resource costs
Token-based PricingPay only for processed data
Provisioned ThroughputConsistent performance for high-demand apps

[!TIP] Optimizing responsiveness and availability through regional redundancy is a key trade-off for managing total cost of ownership (TCO).

Question

Accessibility (Single API Access)

Answer

The capability of services like Amazon Bedrock to provide access to a wide variety of foundation models (FMs) through a unified interface.

Why it matters:

  • Flexibility: Easily swap models to compare performance.
  • Simplicity: One API for text generation, image creation, and summarization.
  • Scalability: Built-in AWS infrastructure ensures performance standards are met as demand grows.

Question

Operational Reliability & Security

Answer

Building on AWS ensures that GenAI applications inherit enterprise-grade security, compliance, and safety features.

[!WARNING] Data privacy is a core concern; AWS ensures customer data is not used to train foundation models for other organizations.

Core Benefits:

  • Built-in Security: IAM roles, encryption, and PrivateLink.
  • Compliance: Meets global regulatory standards.
  • Operational Resilience: High availability and automated scaling.

Advantages of Generative AI(5 cards shown)

Question

Adaptability

Answer

The ability of a Foundation Model (FM) to span many different domains and tasks using a single model architecture.

[!NOTE] For businesses, this means instead of relying on multiple niche applications, one general-purpose model can be adapted for diverse needs like coding, writing, and analysis.

Example: A single model being used for both technical documentation and creative marketing copy.

Question

Responsiveness

Answer

The capacity of Generative AI to generate human-like responses in near real-time.

Key Characteristics:

  • Chatbots: Provides immediate interaction for customer service.
  • Low Latency: Critical for applications like live streaming or gaming.
  • User Engagement: High responsiveness prevents user drop-off.

[!TIP] Think of responsiveness as the 'speed of conversation.'

Question

Simplicity

Answer

The ability to generate high-quality, complex content using only a natural language prompt.

Benefits:

  • Lower Barrier to Entry: Non-technical users can produce professional results.
  • Fast Prototyping: Ideas can be visualized or drafted in seconds.
  • Ease of Use: Removes the need for complex syntax or programming to achieve specific outputs.

Question

Efficiency (Automation)

Answer

The use of GenAI to automate tedious and repetitive activities, significantly reducing the time and manual effort required for tasks like summarization.

FeatureManual ProcessGenAI Process
SpeedMinutes to HoursSeconds
ScalabilityLimited by headcountVirtually unlimited
AccuracyProne to human fatigueConsistent across large volumes

[!TIP] Automation frees up human workers to focus on higher-value strategic tasks.

Question

Personalization

Answer

The ability to tailor responses and content to specific user preferences and historical data interactions.

Mechanisms:

  • Contextual Awareness: The model remembers prior parts of the conversation.
  • Tailored Output: Adjusts tone, style, and complexity based on the audience.
  • Automation: Preferences can be applied automatically based on user profiles.
Loading Diagram...

AI and Machine Learning Fundamentals(5 cards shown)

Question

Artificial Intelligence (AI) vs. Machine Learning (ML) vs. Deep Learning (DL)

Answer

These terms represent a hierarchical relationship where each is a subset of the previous:

  • AI: The broad field of creating systems that mimic human intelligence.
  • ML: A subset of AI where computers learn from data without being explicitly programmed.
  • DL: A subset of ML that uses multi-layered neural networks to process large datasets and detect complex patterns.
Loading Diagram...

Question

Neural Networks

Answer

A computational model inspired by the structure of the human brain. It consists of interconnected nodes (neurons) organized into layers:

  1. Input Layer: Receives the raw data.
  2. Hidden Layers: One or more layers that perform mathematical operations to extract features and learn patterns.
  3. Output Layer: Provides the final prediction or classification.

[!NOTE] The "Deep" in Deep Learning refers to having many hidden layers within the neural network.

Question

What is the difference between an Algorithm and a Model?

Answer

FeatureAlgorithmModel
DefinitionA mathematical procedure or set of rules.The specific program created after training.
AnalogyThe recipe for a cake.The actual cake produced.
ExampleLinear Regression, Random Forest.A file that predicts house prices based on ZIP code.

An algorithm is applied to data to produce a model.

Question

Training vs. Inferencing

Answer

  • Training: The iterative process where an algorithm analyzes a dataset to learn patterns and adjust internal parameters (weights) to minimize error.
  • Inferencing: The phase where the trained model is used to make predictions on new, unseen data.

[!TIP] Training is computationally expensive and happens once (or periodically); Inferencing must often happen in real-time or batch as users interact with the system.

Question

Bias and Fairness in AI

Answer

  • Bias: Systematic error in a model's output, often stemming from unrepresentative training data, which leads to unfair advantages or prejudices against specific groups.
  • Fairness: The practice of ensuring AI systems provide impartial and equitable results, regardless of attributes like race, gender, or age.

[!WARNING] High bias can also refer to underfitting, where a model is too simple to capture the underlying patterns in the data.

AI Governance Protocols and Frameworks(5 cards shown)

Question

AI Governance

Answer

The policies, procedures, and oversight structures established to guide the ethical and compliant development and operation of AI systems.

Key Components:

  • Roles & Responsibilities: Clearly defined ownership of AI assets.
  • Risk Mitigation: Protocols for identifying and reducing bias or privacy violations.
  • Accountability: Frameworks for auditing and explaining AI decisions.

[!NOTE] Effective governance is dynamic, requiring ongoing monitoring rather than a one-time setup.

Question

Generative AI Security Scoping Matrix

Answer

A structured framework used to assess risks and implement security measures based on the specific deployment model of a generative AI application.

ScopeDescriptionPrimary Concern
Scope 1Consumer Apps (e.g., ChatGPT)Data input/output privacy
Scope 2Enterprise Apps (Third-party)Vendor contracts and SLAs
Scope 3-5Custom/Self-hostedInfrastructure & Model security
Loading Diagram...

Question

Amazon SageMaker Model Cards

Answer

A transparency standard providing a centralized, standardized way to document the "full story" of a machine learning model.

What they document:

  • Data Lineage: Origins and licenses of training data.
  • Intended Use: What the model should (and shouldn't) be used for.
  • Risk Profile: Known biases, limitations, or quality issues.
  • Performance: Benchmarks and evaluation results.

[!TIP] Use Model Cards to build trust with stakeholders and simplify the audit process.

Question

Review Cadence and Strategy

Answer

The scheduled frequency (cadence) and methodological approach (strategy) for evaluating AI systems to ensure they remain safe and compliant.

Common Strategies:

  • Stress Testing: Testing the system under extreme or adversarial conditions.
  • Human Audits: Manual review of model outputs by subject matter experts.
  • Subgroup Analysis: Checking performance across different demographic groups to detect bias.

[!WARNING] Without a regular review cadence, models may suffer from 'drift,' where performance degrades or bias increases over time as real-world data changes.

Question

Cross-functional Governance Teams

Answer

A team training and organizational requirement involving diverse stakeholders to manage the complex lifecycle of AI systems.

Stakeholders Included:

  • Legal/Compliance: Ensuring adherence to laws like GDPR or HIPAA.
  • Ethicists: Evaluating social impact and fairness.
  • Technologists: Managing data security and infrastructure.
  • Business Leaders: Aligning AI outputs with organizational goals.

Goal: Proactively identify emerging ethical dilemmas that a single department might overlook.

AI, ML, Deep Learning, and Generative AI Fundamentals(5 cards shown)

Question

Artificial Intelligence (AI)

Answer

The broadest field in computer science focused on creating systems capable of performing tasks that typically require human intelligence.

Key Characteristics:

  • Simulates human-like cognitive functions (problem-solving, reasoning).
  • Can be based on logic/rules (GOFAI) or data-driven approaches.

[!NOTE] AWS defines AI as technology with human-like problem-solving capabilities, such as recognizing images or writing poems.

Question

Machine Learning (ML)

Answer

A subset of AI that focuses on algorithms that learn patterns from data to make predictions or decisions without being explicitly programmed for the specific task.

Core Principle: As the system is exposed to more data, it iteratively improves its performance.

Common Use Cases:

  • Fraud detection
  • Recommendation engines
  • Predictive analytics (e.g., predicting house prices)

[!TIP] If AI is the goal, ML is the most common "engine" used to reach it today.

Question

Deep Learning (DL)

Answer

A specialized subset of Machine Learning that utilizes multi-layered Artificial Neural Networks to learn complex patterns in large datasets.

Distinguishing Features:

  • End-to-end learning: Often eliminates the need for manual feature engineering.
  • Neural Networks: Modeled loosely after the human brain with input, hidden, and output layers.
  • High-Dimensional Data: Excels at processing images, speech, and natural language.
Loading Diagram...

Question

Generative AI (GenAI)

Answer

A subset of Deep Learning that uses Foundation Models to create entirely new content rather than just classifying or predicting existing data.

Output Types:

  • Text (e.g., LLMs like GPT or Claude)
  • Images (e.g., Diffusion models)
  • Audio, Video, and Software Code

[!WARNING] "Generative" is the keyword—unlike traditional ML which might tell you if a photo contains a cat, GenAI creates a new photo of a cat.

Question

The AI Hierarchy (Nested Relationship)

Answer

The relationship between these fields is best visualized as a series of nested circles, where each subsequent field is a specialized version of the one before it.

LevelConceptPrimary Focus
OuterAIHuman-like intelligence/problem-solving
InnerMLLearning from data without explicit rules
DeeperDLUsing multi-layered neural networks
CoreGenAIGenerating new content from foundation models
Loading Diagram...

Amazon Bedrock Capabilities(5 cards shown)

Question

Amazon Bedrock

Answer

A fully managed service that offers a choice of high-performing foundation models (FMs) from leading AI companies (like AI21 Labs, Anthropic, Cohere, Meta, Mistral AI, Stability AI, and Amazon) via a single API.

[!NOTE] It is a serverless service, meaning you don't have to manage any underlying infrastructure to deploy and scale your AI applications.

Key Capabilities:

  • Text generation and image synthesis
  • Data analysis and summarization
  • Unified API for multi-model experimentation

Question

Knowledge Bases for Amazon Bedrock

Answer

A feature that automates the end-to-end Retrieval-Augmented Generation (RAG) workflow, including data ingestion, retrieval, prompt augmentation, and citations.

FeatureDescription
Data IngestionAutomatically parses and chunks unstructured data.
GroundingEnsures model responses are based on trusted, proprietary data.
Hallucination ReductionMinimizes incorrect outputs by providing context-aware facts.
Loading Diagram...

Question

Agents for Amazon Bedrock

Answer

Managed components that can execute multi-step tasks by interpreting user requests, breaking them down into logical steps, and interacting with company systems through API calls.

[!TIP] Think of Agents as the "hands" of the model—they don't just talk; they do (e.g., booking a flight or checking inventory).

Components:

  • Instructions: Natural language guidance for the agent.
  • Action Groups: API schemas and Lambda functions to execute logic.
  • Traceability: Visible reasoning steps to debug how the agent reached an answer.

Question

Amazon Bedrock Guardrails

Answer

A security and safety feature that allows developers to implement customized policies to ensure responsible AI outputs and protect against risks.

Capabilities:

  • Content Filtering: Blocks hate speech, insults, and sexual content.
  • Sensitive Information Filtering: Masks or blocks PII (Personally Identifiable Information).
  • Word Filters: Blocks specific custom keywords (e.g., competitor names).
  • Contextual Grounding Checks: Detects hallucinations by checking if the response is supported by the source data.

Question

Model Customization (Fine-tuning & Pre-training)

Answer

Techniques used to adapt foundation models to specific domains or organizational styles using private datasets.

  • Fine-tuning: Adapts a pre-trained model using task-specific labeled data to improve performance on specific functions.
  • Continued Pre-training: Further trains a model on large volumes of unlabeled, domain-specific data (e.g., medical journals or legal codes) to increase its specialized knowledge.

[!WARNING] Customization creates a private copy of the model; your data is never used to train the base models provided by AWS or third-party providers.

Showing 30 of 340 flashcards. Study all flashcards →

Ready to ace AWS Certified AI Practitioner (AIF-C01)?

Access all 353 practice questions, 5 timed mock exams, study notes, and flashcards — no sign-up required.

Start Studying — Free