AI Curriculum Notes for Class 10 Facilitator Handbook (CBSE + Intel)
Unit 1: Revisiting AI Project Cycle & Ethical Frameworks for AI
- Context: AI Curriculum for Class 10 Facilitator Handbook curated with support from Intel; CBSE collaboration since 2019.
- Purpose: Equip educators and students with AI concepts, project workflow, ethical reasoning, and real‑world relevance for an AI‑driven world.
- Acknowledgements: Patrons and content curation team listed (CBSE, Intel, government officials, educators).
1.1 AI Project Cycle
- The AI Project Cycle is a cyclical process to complete an AI project with 6 stages:
- 1) Problem Scoping: define the goal and scope of the AI project.
- 2) Data Acquisition: collect data from reliable sources; data are the base for the project.
- 3) Data Exploration/Visualization: interpret patterns using graphs, databases, maps, etc. to understand relations.
- 4) Modelling: decide on a suitable model(s) after researching online and evaluating alternatives.
- 5) Testing/Evaluation: test models on newly fetched data to assess performance.
- 6) Deployment: integrate the model into real-world environments to deliver value.
- Practical example: Greeting card creation to illustrate planning; parallels to AI project planning (idea → materials → execution → iteration → delivery).
- Key takeaway: Planning and iterative refinement are essential; data quality and model selection determine success.
1.2 Introduction to AI Domains
- Three broad AI domains introduced:
- Statistical Data: AI domain focused on data systems, data collection, storage, and extraction of insights; examples include price comparison websites (PriceGrabber, PriceRunner, Junglee, Shopzilla, DealTime).
- Computer Vision (CV): Enables machines to acquire, analyze, and act on visual data (images, videos, infrared sensors); aims to translate visual data into computer‑readable descriptions; examples include agricultural monitoring, surveillance, etc.
- Natural Language Processing (NLP): Interaction between computers and humans using natural language; aims to read, decipher, understand, and extract meaning from human language; examples include email filters and machine translation.
- CV applications: Facial recognition, face filters, image search, retail analytics, self‑driving cars, medical imaging, Google Translate app.
- NLP applications: Email filters, machine translation, sentiment analysis, text classification, keyword extraction.
- Importance of domain understanding: Contextualizing AI concepts to real‑world domains enhances learning and social impact.
1.3 Ethical Frameworks for AI
- What is a framework? A framework provides a step‑by‑step, organized approach to solving problems; common language for collaboration; helps ensure consistency.
- What is an ethical framework? A systematic approach to navigate moral dilemmas by considering principles and stakeholder impacts; helps ensure choices do not cause unintended harm.
- Why do we need Ethical Frameworks for AI?
- To mitigate bias and ensure morally acceptable recommendations from AI systems.
- To anticipate and avoid unintended outcomes before deployment.
- Types of ethical frameworks:
- Sector-based frameworks: Tailored to specific sectors (e.g., Bioethics in healthcare). Other sectors include finance, education, governance, etc.
- Value-based frameworks: Grounded in fundamental ethical principles (e.g., rights-based, utility-based, virtue-based).
- Value-based frameworks (3 families):
- Rights-based: Emphasize protection of human rights and dignity; autonomy and freedoms; avoid discrimination.
- Utility-based (Utilitarian): Maximize overall utility/benefit; weigh benefits vs. risks to maximize societal good.
- Virtue-based: Focus on character and intentions of individuals/organizations; alignment with virtuous traits (honesty, compassion, integrity).
- Note: Value-based frameworks assess moral worth of actions and guide ethical behavior across contexts.
- Bioethics (a prominent sector‑based framework in healthcare):
- Core principles: Respect for Autonomy; Do No Harm (non‑maleficence); Beneficence (maximize benefit); Justice (fair distribution of benefits/burdens).
- Non‑maleficence: Avoid causing harm; minimize harm where possible.
- Maleficence: Avoid intentional harm.
- Beneficence: Promote well‑being and positive outcomes for stakeholders.
- Case study: Applying bioethics to AI in healthcare
- Example problem: An AI algorithm trained on healthcare expense data (US‑centric) may misrepresent illness severity across regions, leading to biased resource allocation.
- Application of bioethics:
- Respect for Autonomy: Provide transparency, reproducibility, and accessible model information to patients.
- Do No Harm: Distribute benefits and harms justly; minimize regional biases in data and outcomes.
- Maximum Benefit: Use unbiased, clinically relevant data; ensure medical practice standards guide AI deployment.
- Justice: Consider social determinants of health; address structural biases affecting outcomes across populations.
- Activities to reveal ethical biases:
- My Goodness game: interactive exercise exploring donation decisions to surface personal biases and decision‑making factors.
- Factors that can influence decisions: identity/location of recipients, bias toward relatives, information disclosure, cultural and religious values, intuition.
- Ethical frameworks in practice:
- Students learn to outline the AI project cycle, identify AI domains, classify ethical frameworks, explore bioethics principles, and practice applying frameworks to AI solutions.
Unit 2: Advanced Concepts of Modeling in AI
- Goal: Introduce modeling concepts, distinguishing rule‑based vs learning‑based AI, and the major learning paradigms.
2.1 Revisiting AI, ML, DL
- Definitions and relationships:
- AI (Artificial Intelligence): Umbrella term for techniques enabling computers to mimic human intelligence.
- ML (Machine Learning): Subset of AI; systems improve with experience/data; learns from data to improve performance.
- DL (Deep Learning): Subset of ML; uses large volumes of data; multiple layers of neural networks to learn representations.
- Relationship: DL is a subset of ML, which is a subset of AI (AI ⊇ ML ⊇ DL).
- Conceptual visualization: AI includes a broad set of techniques; ML is about learning from data; DL is a specific family of learning methods using deep neural networks.
2.2 Modelling: Rule‑Based vs Learning‑Based; ML Model Categories
- Two broad modelling approaches:
- Rule‑Based: Developer defines rules and data interactions; model behavior is fixed after training; limited ability to adapt to new data without reprogramming.
- Learning‑Based: Machine learns rules from data; adapts to changing data patterns; examples include spam filtering where rules evolve with data.
- Categories of learning‑based models (ML/DL):
- Supervised Learning: Trained on labeled data; tasks include classification and regression.
- Unsupervised Learning: Learns structure from unlabeled data; tasks include clustering and association.
- Reinforcement Learning: Learns via rewards/punishments to maximize cumulative reward; interacts with environment.
- Neural networks (NN): Core mechanism in many learning systems; comprised of layers and nodes; automatic feature extraction; capable of handling large datasets (e.g., images).
- Subtopics to know:
- Supervised Learning: Classification vs Regression; labeled data; examples like coin weight classification (weights as features; currency as labels).
- Unsupervised Learning: Clustering and Association; no labels; discover structure/patterns; examples include customer segmentation and market basket analysis.
- Reinforcement Learning: Trial‑and‑error with reward signal; used for tasks like parking a car or humanoid walking.
- Practical note: Real‑world AI often combines approaches to handle data gaps and changing environments.
2.3 Neural Networks (NNs)
- What is a neural network? A system inspired by the brain, consisting of layers of nodes (neurons) arranged as input, hidden, and output layers.
- How it works:
- Data flows from input to output through hidden layers where each node applies a function to weighted inputs plus a bias.
- Activation functions introduce nonlinearity to enable complex mappings.
- Training adjusts weights (via error signal) to minimize loss; learning continues through backpropagation across multiple epochs.
- Depth and capacity: More hidden layers (deep networks) can model more complex functions; the number of nodes/layers depends on problem complexity.
- Real‑world NN applications: Facial recognition, chatbots, price prediction, etc.
- Quick perceptron note (illustrative): A simple decision unit with weights, bias, and threshold; output often uses a step function to map to a binary decision.
Unit 3: Evaluating Models
- Purpose: Understand model performance using standard metrics; establish a feedback loop to improve accuracy and reliability.
3.1 Importance of Model Evaluation
- Evaluation is the process of using metrics to quantify model performance.
- Why evaluation matters: prevents deployment of flawed models; provides a basis for comparison and improvement.
3.2 Train–Test Split
- Concept: Divide data into training and testing subsets to estimate performance on unseen data.
- Rationale: Training data are used to learn; testing data evaluate predictive ability on new inputs; avoid overfitting.
- General rule: Use train‑test split when there is a sufficiently large dataset.
3.3 Accuracy and Error
- Accuracy: Proportion of correct predictions across all predictions.
- Error: Proportion (or rate) of incorrect predictions; used to quantify miss rate.
- Relationship: Accuracy = 1 − Error.
- Real‑world nuance: In imbalanced datasets, accuracy alone can be misleading; other metrics may be preferred.
3.4 Evaluation Metrics for Classification
- Classification: Assigning items to discrete classes; metrics include:
- Confusion matrix: 2x2 table for binary cases; entries include True Positive (TP), True Negative (TN), False Positive (FP), False Negative (FN).
- Accuracy = (TP + TN) / (TP + TN + FP + FN)
- Precision = TP / (TP + FP)
- Recall (Sensitivity) = TP / (TP + FN)
- F1 Score = 2 * (Precision * Recall) / (Precision + Recall)
- When to use which metric:
- Precision emphasizes reducing FP (e.g., high‑risk medical tests, satellite launch decisions).
- Recall emphasizes reducing FN (e.g., disease detection to avoid missing cases).
- F1 Score balances precision and recall, especially in imbalanced datasets.
3.5 Ethical Concerns in Model Evaluation
- Bias: Evaluation data can reflect societal biases; must ensure fair assessment across groups.
- Transparency: Clear reporting of model behavior, limitations, and risk disclosures.
- Accountability: Define responsibility for outcomes and potential harms.
Unit 4: Statistical Data & No‑Code AI for Statistics
- No‑Code AI for Statistical Data: No‑code vs low‑code approaches; aim to enable non‑programmers to build AI models without writing code.
- No‑Code vs Low‑Code: No‑code tools emphasize drag‑and‑drop GUIs; low‑code offers some scripting; both aim to reduce traditional coding requirements.
- No‑Code tools highlighted in this unit include Orange Data Mining, Lobe, Teachable Machine, Google Cloud AutoML, Azure ML, etc.
- Orange Data Mining (ODMap): A no‑code data mining tool with a GUI for loading data, exploring, preprocessing, modeling, evaluating, and deploying models; supports both supervised and unsupervised workflows via widgets.
- Palmer Penguins case study (OD Map): Demonstrates data exploration, modeling, and evaluation using a real dataset to predict penguin species; illustrates mapping of the AI project cycle to a concrete data science problem.
- Statistical concepts covered:
- Descriptive statistics: Mean, Median, Mode.
- Distributions: Normal distribution and other distributions; understanding skew and central tendency.
- Probability and variance; standard deviation.
- Outliers and data quality considerations.
- Excel for statistical analysis: Introduction to using Excel’s Analysis ToolPak for simple linear regression (speed vs distance) with steps to load add‑ins, run regression, and interpret outputs including regression equation and R².
Unit 5: Computer Vision
- 5.1 Introduction to CV: CV enables machines to see and interpret visual data; CV is a branch of AI focusing on image/video understanding.
- 5.2 CV Applications: Face recognition, face filters, image search, CV in retail, self‑driving cars, medical imaging, Translate by image.
- 5.3 CV Tasks:
- Classification: Assign an image to one of several categories.
- Classification + Localisation: Locate and classify a single object in an image.
- Object Detection: Identify and locate multiple instances of objects within an image.
- Image Segmentation / Instance Segmentation: Pixel‑level labeling of objects.
- Image basics: Pixels, resolution, grayscale vs RGB, image channels.
- 5.4 No‑Code AI tools for CV:
- Lobe: Auto‑ML for image classification; drag‑and‑drop, no coding required.
- Teachable Machine: Web tool by Google for image, audio, and pose classification; runs on TensorFlow.js.
- Orange Data Mining: Includes CV‑oriented widgets and use case walkthroughs (e.g., coral bleaching, smart sorters).
- 5.5 Image Features: Features such as corners, edges, and blobs; corners are generally robust and good features to extract.
- 5.6 Convolution: Core operator in CV; convolution multiplies image with a kernel to produce a feature map; used for edge detection, blurring, sharpening, etc.
- 5.7 Convolutional Neural Network (CNN): A deep learning model that uses convolutional layers, ReLU activation, pooling layers, and fully connected layers to perform image understanding.
- 5.7.1 Convolution Layer: Extracts high‑level features (edges, textures) and produces a feature map.
- 5.7.2 ReLU: Introduces non‑linearity by setting negative values to zero; enhances feature detection.
- 5.7.3 Pooling Layer: Reduces spatial dimensions; max pooling vs average pooling; increases invariance to small shifts.
- 5.7.4 Fully Connected Layer: Maps features to class probabilities and outputs final prediction.
- 5.8 Python libraries: TensorFlow, Keras, OpenCV; used for implementing CV workflows in code.
Unit 6: Natural Language Processing (NLP)
- 6.1 Introduction to NLP: NLP focuses on enabling computers to understand and process human language; features of natural language include ambiguity, context dependence, and dynamic evolution.
- 6.2 Applications of NLP: Autogenerated captions, voice assistants, language translation, sentiment analysis, text classification, keyword extraction.
- 6.3 Stages of NLP:
- Lexical Analysis: Tokenization and lexicon handling; breaking text into words/tokens.
- Syntactic Analysis (Parsing): Grammar checking and relationships among words; syntactic structure.
- Semantic Analysis: Meaning and validity of statements; semantic plausibility checks.
- Discourse Integration: Coherence across sentences; narrative flow.
- Pragmatic Analysis: Practical usage and real‑world intent; contextual relevance.
- 6.4 Chatbots: Scripted vs smart bots; chatbots simulate conversation; examples of chatbots and experiences with them.
- 6.5 Text Processing: Text normalization, Bag of Words, TFIDF; importance of preprocessing and feature extraction for ML models.
- 6.6 NLP: Use case walkthroughs; no‑code tools in NLP (Orange Data Mining) highlighted; emphasis on sentiment analysis and other NLP tasks.
- No‑Code NLP tools: Orange Data Mining, NLP workflows; focus on sentiment analysis and keyword extraction.
Practical Concepts and Formulas
- TFIDF in text processing (concept and formula):
- Let N be the total number of documents; DF(W) is the document frequency (how many documents contain word W).
- IDF(W) = \log_{10}\left( \frac{N}{DF(W)} \right).
- For a document d, TF(W, d) is the term frequency of word W in document d (how many times W occurs in d).
- TFIDF(W, d) = TF(W, d) * IDF(W).
- TFIDF values are used to identify important words in a document relative to the corpus; high TFIDF indicates a word that is frequent in a document but not common across all documents.
- Perceptron (illustrative): Output ≈ step(\sumi wi x_i + b); a simple binary classifier; foundational concept for neural nets.
- Convolutional Neural Network (CNN) architecture (high‑level):
- Convolution Layer → ReLU → Pooling Layer → Fully Connected Layer → Output.
- Feature maps (activation maps) produced by convolution; ReLU introduces nonlinearity; pooling reduces dimensionality and increases translation invariance; fully connected layers perform final classification.
Connections to Real World and Foundational Principles
- Ethical considerations: Throughout Unit 1, ethical frameworks underpin design decisions, data collection, model evaluation, and deployment, ensuring fairness, transparency, and accountability.
- Cross‑unit connections:
- Unit 1 sets the stage for applying ethical frameworks across all AI projects.
- Unit 2 clarifies the modeling choices that drive Unit 3’s evaluation metrics.
- Unit 4 introduces statistics and no‑code tools that support data exploration used in CV and NLP tasks (Unit 5 and Unit 6).
- Unit 5 and Unit 6 provide domain‑specific methods (CV and NLP) that rely on data representations learned in earlier units (features, vectors, embeddings, etc.).
Test Yourself (Representative Questions and Concepts)
- Example question types from the transcript focus on problem scoping, domain categorization, evaluation metrics, CV/NLP concepts, and ethical frameworks. Sample categories include:
- Problem statement purpose in AI Project Cycle.
- Data domain categorization based on data fed to models.
- Statistical data handling and uses of no‑code tools.
- CV tasks (classification, localization, object detection, segmentation).
- NLP stages (lexical, syntactic, semantic, discourse integration, pragmatic).
- Ethical frameworks (rights, utility, virtue, bioethics) and their application to AI projects.
Case Studies and Scenarios
- Healthcare AI case study (bioethics): address bias and fairness; apply four bioethics principles to ensure equity in patient care and resource allocation.
- Palmer Penguins case (Orange ODMap): a practical example of mapping the AI project cycle to a real dataset and evaluating model performance.
- My Goodness bias activity: a practical activity to reflect on bias in charitable decisions and how personal values influence AI design.
Key Takeaways for Exam Preparation
- Know the 6 stages of the AI Project Cycle and their purposes.
- Be able to distinguish AI domains: Statistical Data, Computer Vision, NLP; know typical examples and applications.
- Understand rule‑based vs learning‑based modelling, and the three ML paradigms: supervised, unsupervised, reinforcement learning.
- Be able to explain basic NN concepts (layers, activations, backpropagation) and the core CNN components (convolution, ReLU, pooling, fully connected).
- Master evaluation metrics for classification (accuracy, precision, recall, F1, confusion matrix) and know when to use each in different data contexts (balanced vs imbalanced datasets).
- Understand no‑code vs low‑code AI tools, and their role in democratizing AI without heavy coding, with examples like Orange Data Mining, Lobe, Teachable Machine, and AutoML platforms.
- Grasp TFIDF for NLP, including the definitions of TF, IDF, and TFIDF; understand why high TFIDF words are informative for document classification and information retrieval.
- Appreciate ethical frameworks as guiding principles in AI design and deployment, especially piecewise application to sector‑specific problems (healthcare, finance, governance).
Notation and Formulas (LaTeX)
TFIDF in document d for word W:
ext{TFIDF}(W, d) = ext{TF}(W, d) imes ext{IDF}(W)Inverse Document Frequency:
ext{IDF}(W) =
\log_{10}iggl( \frac{N}{DF(W)} \biggr)
where N is the total number of documents and DF(W) is the number of documents containing W.Confusion Matrix (binary): TP, TN, FP, FN; Accuracy:
ext{Accuracy} = \frac{TP + TN}{TP + TN + FP + FN}Precision, Recall, F1:
ext{Precision} = \frac{TP}{TP + FP}, \quad ext{Recall} = \frac{TP}{TP + FN}, \quad ext{F1} = 2 \times rac{\text{Precision} \times \text{Recall}}{\text{Precision} + \text{Recall}}Conceptual equation for a simple perceptron (binary decision):
y = \begin{cases}1, & \sumi wi x_i + b > 0 \ 0, & \text{otherwise} \end{cases}CNN workflow (high level):
Input Image → Convolution Layer (feature maps) → ReLU → Pooling → Fully Connected Layer → Output (class probabilities)
Connections to Foundational Principles and Real‑World Relevance
- Ethical framing informs all stages of AI development, from data collection to deployment.
- No‑Code tools democratize access to AI, enabling non‑technical stakeholders to prototype and deploy solutions, but require awareness of limitations (flexibility, potential biases, security concerns).
- CV and NLP are critical for visual and language information processing in modern applications (autonomous vehicles, healthcare imaging, sentiment analysis, translation).
- Statistical thinking (descriptive stats, distributions, regression) underpins data exploration and model evaluation across domains.
Quick Reference Glossary
AI Project Cycle: Problem Scoping, Data Acquisition, Data Exploration, Modelling, Testing, Deployment.
Ethical Frameworks: Sector‑based (Bioethics) and Value‑based (Rights, Utility, Virtue).
No‑Code / Low‑Code: Platforms enabling AI development with minimal or no coding.
TFIDF: Term Frequency–Inverse Document Frequency; a weighting scheme for text representation.
Convolutional Neural Network (CNN): Deep learning model for image data using convolutional layers.
Lexical Analysis / Syntactic Analysis / Semantic Analysis / Discourse / Pragmatics: Stages of NLP processing.
For each unit, be prepared to discuss definitions, examples, ethical considerations, and how to apply these concepts to real situations (classroom activities and case studies).