Domain 4: Guidelines for Responsible AI (14%) โ
โ Domain 3 ยท Next Domain โ
4.1: Responsible AI Principles โ
Responsible AI Principles
1 / 3
โ
What is Fairness in AI?
(Click to reveal)๐ก
Avoid bias and discrimination
Ensure AI treats all groups equitably
Test across demographic groups, use diverse training data.
Ensure AI treats all groups equitably
Test across demographic groups, use diverse training data.
1. Fairness โ
Avoid bias and discrimination
Types of Bias:
- Training Data Bias: Unrepresentative data
- Selection Bias: Biased sampling
- Algorithmic Bias: Model amplifies existing bias
Mitigation:
- Diverse training data
- Test across demographic groups
- Regular bias audits
- Use SageMaker Clarify
2. Explainability โ
Understand how models make decisions
Why It Matters:
- Build trust
- Debug errors
- Meet regulatory requirements
- Identify bias
Techniques:
- Feature importance
- SHAP values
- Attention visualization
- Model cards
3. Privacy โ
Protect sensitive data
Best Practices:
- Data encryption (at rest, in transit)
- Access controls (IAM)
- Data anonymization
- Differential privacy
- Secure data deletion
4. Safety โ
Prevent harmful outputs
Risks:
- Toxic content generation
- Misinformation
- Dangerous instructions
- Manipulation
Guardrails:
- Content filtering
- Human review (Amazon A2I)
- Output validation
- Rate limiting
5. Transparency โ
Document capabilities and limitations
Model Cards Should Include:
- Intended use cases
- Training data description
- Known limitations
- Performance metrics
- Bias testing results
4.2: AWS Tools for Responsible AI โ
Amazon SageMaker Clarify โ
Detect bias in data and models
Responsible AI Tools
1 / 3
โ
What does SageMaker Clarify do?
(Click to reveal)๐ก
Detect bias in data and models
Pre-training and post-training bias detection
Feature importance and explainability reports.
Pre-training and post-training bias detection
Feature importance and explainability reports.
Features:
- Pre-training bias detection
- Post-training bias detection
- Feature importance
- Explainability reports
Bias Metrics:
- Class imbalance
- Disparate impact
- Difference in positive proportions
Amazon SageMaker Model Monitor โ
Continuous monitoring
Monitors:
- Data quality
- Model quality (drift)
- Bias drift
- Feature attribution drift
Alerts:
- CloudWatch alarms
- SNS notifications
Amazon Augmented AI (A2I) โ
Human review workflows
Use Cases:
- Review low-confidence predictions
- Audit high-stakes decisions
- Compliance requirements
- Continuous improvement