AIF-C01: Cheatsheet โ
โ Overview ยท โ Exam Guide
Exam Day Reference
Print this or review 5 minutes before the exam.
AWS AI Services Quick Lookup โ
| Need | Service |
|---|---|
| Foundation models | Amazon Bedrock |
| Business assistant | Amazon Q |
| Code suggestions | Amazon CodeWhisperer |
| Full ML platform | Amazon SageMaker |
| Image/video analysis | Amazon Rekognition |
| Extract text | Amazon Textract |
| Sentiment analysis | Amazon Comprehend |
| Translation | Amazon Translate |
| Speech-to-text | Amazon Transcribe |
| Text-to-speech | Amazon Polly |
| Chatbot | Amazon Lex |
| Recommendations | Amazon Personalize |
| Fraud detection | Amazon Fraud Detector |
| Search | Amazon Kendra |
| Human review | Amazon A2I |
Bedrock Models โ
| Model | Provider | Best For |
|---|---|---|
| Claude | Anthropic | Long docs (200K), reasoning, code |
| Titan Text | Amazon | Cost-effective, simple tasks |
| Titan Embeddings | Amazon | RAG, semantic search |
| Titan Image | Amazon | Image generation |
| Jurassic | AI21 Labs | Multilingual text |
| Command | Cohere | Text generation |
| Llama 2 | Meta | Open-source, customizable |
| Stable Diffusion | Stability AI | Image generation |
ML Metrics (PARC) โ
Precision = TP / (TP + FP)
- Use when: False positives costly (spam filter)
Accuracy = (TP + TN) / Total
- Use when: Balanced classes
Recall = TP / (TP + FN)
- Use when: False negatives costly (medical diagnosis)
Confusion Matrix
- TP, TN, FP, FN
Model Fit Issues โ
| Issue | Performance | Solution |
|---|---|---|
| Underfitting | Poor on train AND test | More complex model |
| Good Fit | Good on both | โ Goal |
| Overfitting | Great on train, poor on test | More data, regularization |
RAG Workflow (5 Steps) โ
- User asks question
- Convert question to embedding (vector)
- Search vector database for similar content
- Retrieve relevant documents
- LLM generates answer with context
AWS Vector DB: Amazon OpenSearch Service
Prompt Engineering Ladder โ
Zero-Shot: No examples
- "Translate to French: Hello"
Few-Shot: 3-5 examples
- Show input/output pairs
Chain-of-Thought: Show reasoning
- "Show your work"
Customization Approaches โ
| Approach | When | Cost | Effort |
|---|---|---|---|
| Prompt Engineering | First choice | Low | Low |
| RAG | Current/private data | Medium | Medium |
| Fine-Tuning | Domain-specific | High | High |
Order: Always try Prompt โ RAG โ Fine-Tuning
Responsible AI (FEPST) โ
- Fairness โ Avoid bias, treat all equitably
- Explainability โ Understand decisions
- Privacy โ Protect data (encryption, IAM)
- Safety โ Prevent harmful outputs
- Transparency โ Document capabilities/limits
Responsible AI Tools โ
| Need | AWS Tool |
|---|---|
| Detect bias | SageMaker Clarify |
| Monitor drift | SageMaker Model Monitor |
| Human review | Amazon A2I |
Security Checklist โ
- โ Encrypt at rest (AWS KMS)
- โ Encrypt in transit (TLS/SSL)
- โ IAM policies for access
- โ Input validation
- โ Monitor for prompt injection
- โ Bedrock data stays private (not used for training)
Compliance โ
| Regulation | What It Means |
|---|---|
| HIPAA | Healthcare data (use HIPAA-eligible services) |
| GDPR | EU data privacy (right to explanation) |
| SOC 2 | Security controls |
| PCI DSS | Payment card data |
Learning Types โ
| Type | Data | Example |
|---|---|---|
| Supervised | Labeled (X, y) | Spam detection |
| Unsupervised | Unlabeled | Customer segmentation |
| Reinforcement | Rewards | Game AI |
The ML Pipeline โ
- Business Goal โก๏ธ 2. Data Prep โก๏ธ 3. Model Training โก๏ธ 4. Evaluation โก๏ธ 5. Deployment (Inference) โก๏ธ 6. Monitoring
Inference Types โ
- Real-time: Low latency, immediate (e.g., fraud check at checkout).
- Batch: High volume, delayed (e.g., monthly product recommendations).
GenAI Limitations โ
- โ Hallucinations (generates false info)
- โ Training cutoff (no current events)
- โ Context limits (can't process very long docs)
- โ Bias (reflects training data)
- โ No real-time data (use RAG)
Mitigation: RAG, guardrails, human review
โ Overview ยท โ Exam Guide
Last Updated: 2026-02-05