AIF-C01: Exam Tips & Strategy
Strategic guidance for exam preparation and taking the AIF-C01 AWS Certified AI Practitioner exam.
← Back to Overview | ← Objectives | ← Study Notes
⚠️ Exam Traps & Gotchas
Common mistakes and tricky areas that often appear on the exam.
Trap 1: Confusing ML Types
What it looks like: "A company wants to group customers with similar buying patterns. Which ML type?"
Why it's tricky: All three ML types can seem applicable
Remember:
- Supervised: Has labels (spam detection, price prediction)
- Unsupervised: No labels, find patterns (clustering, segmentation)
- Reinforcement: Learn through rewards (games, robotics)
This scenario → Unsupervised (clustering, no labels)
Trap 2: Hallucination vs. Bias
What it looks like: "A model generates factually incorrect information. What is this?"
Why it's tricky: Bias and hallucination both involve incorrect outputs
Remember:
- Hallucination: Model invents false information presented as fact
- Bias: Model systematically favors certain outcomes due to training data
Key Difference: Hallucinations are random false facts; bias is systematic
Trap 3: RAG vs. Fine-Tuning
What it looks like: "A company needs an LLM to answer questions about their internal documentation. What should they do first?"
Why it's wrong: Jumping to fine-tuning is expensive and unnecessary
Remember:
- Start with prompt engineering
- Try RAG (provide documents as context)
- Fine-tune only if above don't work
This scenario → RAG (provide internal docs as context)
Trap 4: Context Window Confusion
What it looks like: "Claude has a 200K token context window. Can it process 200K tokens of output?"
Why it's wrong: Context window includes both input AND output
Remember:
- Context window = prompt + response
- If prompt uses 150K tokens, only 50K left for response
Trap 5: Precision vs. Recall
What it looks like: "A medical diagnosis system should prioritize which metric?"
Why it's tricky: Both sound important for healthcare
Remember:
- Precision: "Of predicted positives, how many are correct?" (avoid false positives)
- Recall: "Of actual positives, how many did we catch?" (avoid false negatives)
Medical diagnosis → Recall (catch all diseases, even if some false positives) Spam filter → Precision (don't mark real emails as spam)
Trap 6: Amazon Bedrock vs. SageMaker
What it looks like: "Which service for accessing pre-trained foundation models via API?"
Why it's tricky: Both are ML services
Remember:
- Amazon Bedrock: Access foundation models (Claude, Titan, etc.) via API
- SageMaker: Build, train, and deploy custom ML models
For FMs → Bedrock For custom models → SageMaker
📚 Study Strategy
High-Priority Topics (Appear Most Often)
1. Generative AI Fundamentals (24%)
- ✅ Hallucinations and how to mitigate
- ✅ Context windows and tokens
- ✅ Foundation models vs. LLMs
- ✅ Transformers and attention mechanisms
- ✅ Embeddings
2. Foundation Model Applications (28%)
- ✅ RAG architecture and components
- ✅ Prompt engineering techniques
- ✅ When to use RAG vs. fine-tuning
- ✅ Model selection criteria
- ✅ Amazon Bedrock models and capabilities
3. ML Fundamentals (20%)
- ✅ Supervised vs. unsupervised vs. reinforcement learning
- ✅ Overfitting vs. underfitting
- ✅ Precision vs. recall
- ✅ ML lifecycle phases
4. Responsible AI (14%)
- ✅ Bias detection and mitigation
- ✅ Explainability importance
- ✅ SageMaker Clarify
- ✅ Human review with Amazon A2I
5. Security and Governance (14%)
- ✅ Encryption at rest and in transit
- ✅ IAM for access control
- ✅ Model governance and versioning
- ✅ Compliance requirements (HIPAA, GDPR)
What NOT to Over-Study
Don't Waste Time On
- Deep math - No calculus or linear algebra required
- Coding implementation - Conceptual understanding, not code
- Training algorithms - Focus on use cases, not backpropagation
- Specific pricing - Know models, not exact costs
- Advanced ML theory - This is a practitioner exam, not data scientist
⏱️ Time Management
Exam Format
- 120 minutes for 85 questions
- About 1.4 minutes per question
- Mix of multiple choice and multiple response
Strategy
First Pass (80 minutes):
- Answer questions you know
- Flag uncertain questions
- Don't spend more than 2 minutes per question
Review Pass (35 minutes):
- Revisit flagged questions
- Eliminate wrong answers
- Make educated guesses
Final Check (5 minutes):
- Ensure all answered
- Quick double-check of marked questions
No Negative Marking
Always answer every question. No penalty for wrong answers!
🎯 Decision Tables
Which ML Type?
| Scenario | ML Type |
|---|---|
| Have labeled data, predict outcomes | Supervised Learning |
| No labels, find patterns/groups | Unsupervised Learning |
| Learn through trial and error | Reinforcement Learning |
| Classify images | Supervised (Classification) |
| Predict stock price | Supervised (Regression) |
| Customer segmentation | Unsupervised (Clustering) |
| Game AI | Reinforcement Learning |
Which Approach for LLM Applications?
| Requirement | Approach |
|---|---|
| Adjust model behavior | Prompt Engineering |
| Need current/private data | RAG |
| Domain-specific language | Fine-Tuning |
| Reduce hallucinations | RAG + Guardrails |
| Cost-effective first step | Prompt Engineering |
Which AWS AI Service?
| Use Case | Service |
|---|---|
| Access GPT/Claude/Llama | Amazon Bedrock |
| Business Q&A assistant | Amazon Q |
| Code suggestions | Amazon CodeWhisperer |
| Custom ML models | Amazon SageMaker |
| Analyze images/videos | Amazon Rekognition |
| Extract text from PDFs | Amazon Textract |
| Sentiment analysis | Amazon Comprehend |
| Translation | Amazon Translate |
| Speech-to-text | Amazon Transcribe |
| Text-to-speech | Amazon Polly |
| Build chatbots | Amazon Lex |
| Product recommendations | Amazon Personalize |
| Detect fraud | Amazon Fraud Detector |
| Vector search for RAG | Amazon OpenSearch Service |
| Human review workflows | Amazon A2I |
| Detect model bias | SageMaker Clarify |
📝 Exam Day Tips
Key Concepts to Memorize
AI/ML Hierarchy
AI (broadest)
└─ ML (learn from data)
└─ DL (neural networks)Foundation Model Providers on Bedrock
- Anthropic: Claude (long context, reasoning)
- Amazon: Titan (text, embeddings, images)
- AI21 Labs: Jurassic (multilingual)
- Cohere: Command (text generation)
- Meta: Llama 2 (open-source)
- Stability AI: Stable Diffusion (images)
RAG Components
- Document ingestion and chunking
- Embedding generation
- Vector database storage
- Semantic search/retrieval
- Context-augmented generation
Responsible AI Pillars
- Fairness
- Explainability
- Privacy
- Safety
- Security
- Transparency
Keywords to Watch For
"Most cost-effective":
- Prompt engineering (first)
- Use smaller models
- RAG before fine-tuning
"Reduce hallucinations":
- Use RAG
- Provide factual context
- Implement guardrails
"Real-time":
- Smaller models for lower latency
- Consider caching
- Not fine-tuning (too slow to iterate)
"Private/proprietary data":
- RAG with Amazon Bedrock Knowledge Bases
- SageMaker for custom models
- Not public foundation models alone
"Bias detection":
- Amazon SageMaker Clarify
- Test across demographic groups
- Model cards
"Human review":
- Amazon Augmented AI (A2I)
- Low-confidence predictions
- High-stakes decisions
"Long documents":
- Claude (200K context)
- Document chunking for RAG
"Medical/Healthcare":
- HIPAA compliance
- High recall (catch all cases)
- Explainability required
🧠 Memory Aids
Remember Generative AI Limitations (HCBCT)
- Hallucinations
- Context window limits
- Bias in training data
- Cost (computational)
- Training cutoff date
Remember ML Lifecycle (BDFTEDM)
- Business problem
- Data collection
- Feature engineering
- Training
- Evaluation
- Deployment
- Monitoring
Remember Prompt Engineering Types (ZFC)
- Zero-shot (no examples)
- Few-shot (with examples)
- Chain-of-thought (show reasoning)
🎓 Final Checklist
Two weeks before:
- [ ] Complete all domain study notes
- [ ] Understand ML types thoroughly
- [ ] Know all Amazon Bedrock models
- [ ] Understand RAG architecture
- [ ] Know responsible AI principles
- [ ] Review all AWS AI services
One week before:
- [ ] Take practice exams
- [ ] Review decision tables
- [ ] Focus on weak areas
- [ ] Understand precision vs. recall
- [ ] Know when to use each approach (prompt eng, RAG, fine-tuning)
Day before:
- [ ] Light review of notes
- [ ] Review keywords and decision tables
- [ ] Get good sleep
- [ ] Prepare exam details and ID
💡 Question Patterns
Pattern 1: "Which ML type?"
Look for:
- Labels → Supervised
- No labels → Unsupervised
- Rewards → Reinforcement
Pattern 2: "Reduce hallucinations"
Answer:
- RAG (most common)
- Provide factual context
- Not just "use bigger model"
Pattern 3: "Most cost-effective for LLM application"
Answer:
- Prompt engineering (try first)
- RAG (if need data)
- Fine-tuning (last resort)
Pattern 4: "Responsible AI concern"
Look for:
- Bias → SageMaker Clarify
- Explainability → Model cards, SHAP
- Human review → Amazon A2I
- Privacy → Encryption, IAM
Pattern 5: "Which Bedrock model?"
Consider:
- Long context → Claude
- Cost-effective → Titan
- Images → Stable Diffusion
- Embeddings → Titan Embeddings
🚀 You're Ready When...
- ✅ You can explain supervised vs. unsupervised vs. reinforcement learning
- ✅ You understand what hallucinations are and how to mitigate them
- ✅ You know when to use RAG vs. fine-tuning
- ✅ You can describe RAG architecture
- ✅ You know all Amazon Bedrock model providers
- ✅ You understand prompt engineering techniques
- ✅ You can explain precision vs. recall and when each matters
- ✅ You know responsible AI principles
- ✅ You can match AWS AI services to use cases
- ✅ You score 80%+ on practice exams
Good luck with your AI Practitioner exam! 🚀