Skip to content

Domain 1: Fundamentals of AI and ML (20%) โ€‹

โ† Overview ยท Next Domain โ†’


1.1: AI, ML, and DL Hierarchy โ€‹

โ”Œโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”
โ”‚    Artificial Intelligence (AI)     โ”‚  Broadest: Machines that mimic human intelligence
โ”‚  โ”Œโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”  โ”‚
โ”‚  โ”‚   Machine Learning (ML)       โ”‚  โ”‚  Subset: Learn from data without explicit programming
โ”‚  โ”‚  โ”Œโ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”  โ”‚  โ”‚
โ”‚  โ”‚  โ”‚  Deep Learning (DL)     โ”‚  โ”‚  โ”‚  Subset: Neural networks with multiple layers
โ”‚  โ”‚  โ””โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”˜  โ”‚  โ”‚
โ”‚  โ””โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”˜  โ”‚
โ””โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”€โ”˜

Why This Matters

Exam questions test whether you understand the hierarchy: All DL is ML, all ML is AI, but not all AI is ML.


Types of Machine Learning โ€‹

ML Learning Types

1 / 3
โ“

What type of ML uses labeled data?

(Click to reveal)
๐Ÿ’ก
Supervised Learning
Uses features (X) and labels (y) to learn a mapping. Examples: spam detection, house price prediction.

1. Supervised Learning โ€‹

Definition: Learn from labeled data

Components:

  • Features (X): Input variables
  • Labels (y): Known output/target
  • Model: Learns mapping from X to y

Types:

  • Classification: Predict categories (spam/not spam)
  • Regression: Predict continuous values (house price)

Examples:

  • Email spam detection
  • Image classification (cat vs dog)
  • House price prediction
  • Customer churn prediction

2. Unsupervised Learning โ€‹

Definition: Learn patterns from unlabeled data

Types:

  • Clustering: Group similar items
  • Dimensionality Reduction: Reduce features while preserving information
  • Anomaly Detection: Find outliers

Examples:

  • Customer segmentation
  • Anomaly detection in network traffic
  • Product recommendations

3. Reinforcement Learning โ€‹

Definition: Learn through trial and error with rewards

Components:

  • Agent: Learns and makes decisions
  • Environment: What agent interacts with
  • Actions: What agent can do
  • Rewards: Feedback for actions

Examples:

  • Game playing (AlphaGo)
  • Robotics
  • Autonomous vehicles
  • Resource optimization
Learning TypeDataGoalExample
SupervisedLabeledPredict labelsSpam detection
UnsupervisedUnlabeledFind patternsCustomer groups
ReinforcementRewards/penaltiesMaximize rewardsGame AI

ML Model Performance Metrics โ€‹

Classification Metrics โ€‹

Confusion Matrix:

                  Predicted
              Positive  Negative
Actual  Pos      TP        FN
        Neg      FP        TN

Classification Metrics

1 / 3
โ“

When should you prioritize Precision?

(Click to reveal)
๐Ÿ’ก
When false positives are costly
Precision = TP / (TP + FP)
Example: Spam filter (do not mark important emails as spam)

Key Metrics:

  • Accuracy = (TP + TN) / Total

    • Good when classes are balanced
  • Precision = TP / (TP + FP)

    • "Of all predicted positives, how many are actually positive?"
    • Use when false positives are costly
  • Recall = TP / (TP + FN)

    • "Of all actual positives, how many did we catch?"
    • Use when false negatives are costly

Exam Scenario

Medical diagnosis: Prefer high recall (catch all diseases, even with false positives) Spam filter: Prefer high precision (don't mark important emails as spam)

Regression Metrics โ€‹

  • MAE (Mean Absolute Error): Average absolute difference
  • MSE (Mean Squared Error): Average squared difference
  • RMSE (Root Mean Squared Error): Square root of MSE

Overfitting vs. Underfitting โ€‹

Model Fit Issues

1 / 3
โ“

What is Underfitting?

(Click to reveal)
๐Ÿ’ก
Model too simple
Poor on BOTH training and test data.
Solution: More complex model, more features.
IssueDescriptionPerformanceSolution
UnderfittingModel too simplePoor on training AND testMore complex model, more features
Good FitJust rightGood on bothโœ… This is the goal
OverfittingMemorized training dataGreat on training, poor on testMore data, regularization, simpler model

Visual:

Underfitting:  ๐Ÿ™ Training: 70%  ๐Ÿ™ Test: 65%
Perfect Fit:   โœ… Training: 92%  โœ… Test: 90%
Overfitting:   โœ… Training: 99%  ๐Ÿ™ Test: 72%

ML Development Lifecycle โ€‹

  1. Business Problem Definition

    • What are we trying to predict/optimize?
    • What's the success metric?
  2. Data Collection and Preparation

    • Gather data
    • Clean data (handle missing values, outliers)
    • Split data (train/validation/test)
  3. Feature Engineering

    • Create useful features
    • Transform variables
    • Encode categorical variables
  4. Model Training

    • Choose algorithm
    • Train on training data
    • Tune hyperparameters
  5. Model Evaluation

    • Test on validation data
    • Check metrics (accuracy, precision, recall)
    • Compare multiple models
  6. Model Deployment

    • Deploy to production
    • Create API endpoint
    • Monitor performance
  7. Model Monitoring and Maintenance

    • Track model drift
    • Retrain when needed
    • Update as data changes

Exam Focus

The exam tests whether you understand that ML is iterative, not one-and-done. Models need monitoring and retraining.


โ† Overview ยท Next Domain โ†’

Happy Studying! ๐Ÿš€ โ€ข Privacy-friendly analytics โ€” no cookies, no personal data
Privacy Policy โ€ข AI Disclaimer โ€ข Report an issue