AWS GenAI Service Selector
AIP-C01 · Which service for which scenario
Decision Reference · v2
Exam pattern: ~30% of AIP-C01 questions are service selection scenarios. Read the stem carefully — the trigger words tell you exactly which service. The key axis: managed simplicity vs custom control. All flows in the diagram are fully connected. NLP sub-services sit one level below the NLP branch node.
What is the primary need? Start here for every AIP-C01 scenario question Build GenAI app No-code chatbot NLP / Docs / Audio Custom ML / GPU Build a custom GenAI application? FM invocation · RAG · Agents · fine-tuning · Guardrails No-code enterprise chatbot? Zero custom code · connect company docs · ready in days NLP · Document processing · Audio? Classify · Extract text/tables · Transcribe · Search Train / Deploy custom ML on GPU? Fine-tuned LLMs · DJL serving · MLOps · Model Registry Amazon Bedrock FM APIs · RAG · Agents · Guardrails · Fine-tune invoke Claude · Llama · Titan · Cohere · Mistral Amazon Q Business Managed RAG chatbot · 40+ connectors · Zero code SharePoint · Confluence · S3 · Salesforce Amazon SageMaker LMI · DJL · GPU optimization · Model Registry tensor_parallel_degree · async endpoint · LoRA ↓ see services below Knowledge Bases RAG · Vectors Agents ReAct · Tools RoC · Multi-agent Guardrails PII · Topics Safety · Grounding Prompt Mgmt Templates · Batch Comprehend Sentiment · Entities PII · Language detect Textract PDFs · Forms · Tables Scanned documents Kendra Enterprise search Keyword + Semantic Rekognition Image · Video · Faces Content moderation LMI + DJL tensor_parallel continuous batch Model Registry Approve → Deploy LoRA versioning Async Endpoint Long jobs > 60s Serverless / RT when pipeline needed Multi-Step Orchestration? Retry · Parallel branches · Error handling AWS Step Functions State machine · Retry · Circuit breaker · Parallel Glue Data Quality Validate before ingest SM Data Wrangler · Lineage Audio pipeline: Amazon Transcribe → transcript → embed → RAG Knowledge Base Data lineage: Glue Data Catalog tracks source attribution · SM Data Wrangler preps fine-tune data Responsible AI: SageMaker Model Cards → programmatic model documentation Org safety: SCP + bedrock:GuardrailIdentifier enforces Guardrail across all AWS accounts
Quick-Reference Comparison
ServicePrimary Use CaseCustom Code?FM Included?Key Exam Trigger Words
Amazon BedrockFM API, RAG, Agents, Guardrails, fine-tuning✓ Yes✓ Yes"integrate FM", "custom RAG", "invoke Claude/Llama", "Agents"
Amazon Q BusinessNo-code enterprise chatbot over company docs✗ No✓ Yes"no custom code", "connect SharePoint/S3", "employees ask questions", "days not months"
Amazon KendraEnterprise intelligent search (keyword + semantic)~ Minimal✗ No"enterprise search", "relevance ranking", "search not chat"
Amazon SageMakerTrain/deploy custom ML, LMI/DJL GPU inference✓ Yes✗ No"fine-tuned model GPU", "tensor_parallel_degree", "DJL", "MLOps", "Model Registry"
Amazon ComprehendNLP: sentiment, entities, PII, language detection✗ No✗ No"sentiment analysis", "extract entities", "detect PII in text"
Amazon TextractExtract text, tables, forms from PDFs/images✗ No✗ No"scanned docs", "extract form fields", "PDF tables", "invoice processing"
Amazon RekognitionImage/video labels, faces, content moderation✗ No✗ No"image classification", "face detection", "content moderation", "video labels"
AWS Step FunctionsMulti-step workflow orchestration with retry/error handling✓ Yes✗ No"orchestrate pipeline", "retry logic", "parallel branches", "circuit breaker"
Glue Data QualityValidate datasets before FM fine-tuning or RAG ingestion✗ No✗ No"data validation", "quality rules", "completeness", "validate before ingestion"
Amazon TranscribeConvert audio/speech to text for multimodal FM pipelines✗ No✗ No"call recordings", "audio data", "voice to text", "multimodal audio"
Service Deep-Dive Cards
Amazon Bedrock
Core Exam Service
Fully managed API to invoke 20+ FMs. Includes Knowledge Bases (RAG), Agents (multi-step reasoning), Guardrails (safety), Prompt Management, Model Evaluation, and Batch Inference.
Use when scenario mentions:
Building a custom GenAI application
RAG pipeline with control over chunking / retrieval
Bedrock Agents for autonomous multi-step tasks
Fine-tuning, distillation, or prompt management
Safety controls via Guardrails
Custom codeFM includedAgents + RAG
Amazon Q Business
No-Code RAG
Fully managed enterprise AI assistant. Connect to 40+ data sources (S3, SharePoint, Confluence, Salesforce). Employees ask questions in natural language — Q retrieves and answers from company docs.
Use when:
"Zero custom code" or "operational in days"
Internal employee chatbot over company documents
Connect SharePoint, Confluence, Salesforce, S3
No ML engineers available
No custom codeFM included40+ connectors
Amazon Kendra
Enterprise Search
Managed intelligent search combining keyword (BM25) and semantic search. Returns ranked results, not conversational answers. Often combined with Bedrock for a generation layer on top.
Use when:
Search product (results list), not a chatbot
Hybrid keyword + semantic search needed
Enterprise doc search with relevance tuning
Often feeds into a Bedrock FM for answer generation
Search onlyNo generation40+ connectors
Amazon SageMaker
ML Platform + LMI
Full ML lifecycle AND primary service for deploying fine-tuned LLMs on GPU via Large Model Inference (LMI) containers with DJL Serving. Required when Bedrock doesn't give enough infrastructure control.
Use when:
Deploying fine-tuned LLMs on GPU (LMI + DJL)
Optimizing GPU: tensor parallelism, continuous batching
Long FM completions (>60s) → Async Inference endpoint
Model versioning + approval → Model Registry
Training from scratch or LoRA with full control
Custom codeNo FM includedLMI · DJL · GPU
DJL Serving (SageMaker LMI)
GPU Optimization
Open-source serving framework inside SageMaker LMI containers. Key insight: KV-cache ∝ max_sequence_length. Reducing it frees GPU memory for more concurrent requests.
Critical exam parameters:
tensor_parallel_degree=N → shards across N GPUs. 8 GPUs ÷ 4 = 2 replicas
max_sequence_length → controls KV-cache. ↓ length = ↑ batch size
max_rolling_batch_size → concurrent requests per instance
Continuous batching fills GPU slots dynamically
KV-Cache ∝ seq_lenMulti-replica
Amazon Comprehend
NLP Service
Managed NLP for text analysis: sentiment, key phrases, named entities, language detection, PII detection, and custom classification. Pre/post-processing step alongside FM calls.
Use when:
Sentiment analysis on customer reviews/feedback
Extract named entities from text at scale
Detect and classify PII in text documents
Route tickets by topic before sending to FM
No generationText analysisPre-processing
Amazon Textract
Document Extraction
Extracts text, tables, forms, key-value pairs from scanned documents and PDFs. Essential pre-processing before feeding document content into a Bedrock RAG pipeline.
Use when:
Scanned PDFs, invoices, medical forms, contracts
Extract structured tables from documents
Pre-process docs before RAG ingestion
"Document extraction" or "OCR" in scenario
Ingestion pre-stepTables + formsNo generation
Amazon Rekognition
Computer Vision
Managed computer vision: object/scene labels, facial analysis, celebrity recognition, content moderation, text-in-image detection, and activity detection in video.
Use when:
Image or video analysis in the pipeline
Content moderation for user-uploaded media
Multimodal: image → labels → FM prompt
"Face detection" or "image classification"
Vision onlyNo text generation
AWS Step Functions
Workflow Orchestration
Visual workflow orchestration with state machines. Coordinates Lambda, Bedrock, SageMaker, and other services with built-in retry, error handling, parallel branches, and circuit breaker patterns.
Use when:
Multi-step AI pipeline needing retry + error handling
Parallel processing branches required
Coordinate Textract → Comprehend → Bedrock pipeline
"Orchestrate", "circuit breaker", "error handling between steps"
Pipeline onlyRetry built-inNo FM
AWS Glue Data Quality
Data Validation
Defines and runs data quality rules (completeness, uniqueness, validity) against datasets before FM fine-tuning or RAG ingestion. Prevents bad data from degrading model or retrieval quality.
Use when:
Validating documents before RAG KB ingestion
Checking training data quality before fine-tuning
"Data validation", "quality rules", "completeness check"
Pre-ingestionNo FM
Amazon Transcribe
Speech-to-Text
Managed speech-to-text service converting audio recordings to text. Standard pre-processing step for audio data entering FM pipelines — enables voice/call data to be embedded, indexed, or analyzed.
Use when:
Audio recordings or call data needs FM analysis
Multimodal pipeline includes voice/audio data
"Call recordings", "audio to text", "voice data"
Audio pre-stepNo generation