Seven Patterns of AI (Video) - Fill-in-the-Blank Flashcards

When to Use AI & Cognitive Technology

  • AI & Cognitive Technologies are not always the right solution. Many problems are better solved with rules-based or traditional programming, especially if:

    • The problem is repetitive, deterministic, or requires high accuracy.

    • Inputs and outputs are always the same and predictable.

    • Automation (not intelligence) is the main goal.

  • Key questions to ask before using AI:

    • Are the inputs/outputs always the same? ext{(Yes/No)}

    • Is the process highly repetitive and predictable? ext{(Yes/No)}

    • Is high accuracy required, with no tolerance for ambiguity or error? ext{(Yes/No)}

    • Is there enough data to train an AI system? ext{(Yes/No)}

    • Would a human be more reliable, cost-effective, or provide a better experience? ext{(Yes/No)}

  • AI is best for problems with:

    • High variability in data, process, or flow. ext{(Yes)}

    • Ambiguity or tolerance for probabilistic (not guaranteed) results. ext{(Yes)}

    • Situations requiring cognitive intelligence (perception, prediction, planning). ext{(Yes)}

    • If 100% accuracy is required, or if there is no data, AI/ML is not suitable. ext{(No)}

    • Sometimes, hiring a person is more effective than implementing AI. ext{(Yes/No)}

  • A note: There is lots of value in AI & Cognitive Tech without being vague, but choose solution type carefully.

The Seven Patterns of AI

  • The patterns group AI applications into related areas. There are 7 patterns:

    • Recognition Pattern

    • Conversational (Human Interaction) Pattern

    • Predictive Analytics & Decisions

    • Goal-Driven Systems (Optimization & Autonomy)

    • Autonomous Systems

    • Patterns & Anomalies

    • Hyperpersonalization

  • Objective: Provide a categorization system to group different applications of AI into like application areas.

Recognition Pattern

  • Purpose: Make sense of unstructured data (images, audio, video, handwriting, faces, gestures).

  • Technologies: Computer vision, deep learning (e.g., CNNs), OCR.

  • Example use cases: Facial recognition, sound recognition, item & object detection, handwriting/text recognition, gesture/motion analysis.

  • Related concepts: Object recognition, classification, segmentation, pose estimation, event detection, scene reconstruction, image indexing, motion estimation, 3D modeling, image restoration.

  • Real-world uses:

    • Security and surveillance (facial recognition)

    • Retail and shopping analytics (gesture analysis)

    • Medical imaging and robotics (pose estimation, surgical assistance)

    • Entertainment and content analysis (music recognition, video tagging)

The Conversational Pattern (Conversation & Human Interaction)

  • Purpose: Machines and humans interact using natural language across voice, text, written and image forms.

  • Objective: Machines interact with humans the way humans interact with each other.

  • Technologies: NLP, NLU, NLG, STT (speech-to-text), TTS (text-to-speech), ASR (Automatic Speech Recognition).

  • Applications:

    • Chatbots (text or voice)

    • Voice assistants (e.g., Siri, Alexa)

    • Machine translation

    • Sentiment / mood / intent analysis

    • Content summarization and other NLP tasks

  • Important terms:

    • NLP = Natural Language Processing: getting machines to understand and communicate in human language.

    • NLU = Natural Language Understanding: understanding intent, entities, and context beyond words.

    • NLG = Natural Language Generation: generating human-language text or speech from data.

    • ASR = Automatic Speech Recognition: converting speech to text (often interchangeable with STT).

    • STT = Speech-to-Text; TTS = Text-to-Speech.

  • Content generation and translation involve a spectrum of NLP components (NLU/NLG) and are not limited to text-only capabilities.

Predictive Analytics & Decisions

  • Purpose: Use past or current data to help make better decisions and forecast outcomes.

  • Objective: Help humans make better decisions.

  • Technologies: Machine learning models for classification, regression, time-series analysis; pattern recognition; anomaly detection.

  • Applications:

    • Forecasting (inventory, weather, financial markets)

    • Risk/fraud detection and anomaly detection in transactions

    • Recommendation engines

    • Predictive maintenance

    • Marketing optimization and campaign planning

  • Key concepts:

    • Decision support: Presenting data and alternatives to help humans choose.

    • Predictive analytics: Forecasting trends and outcomes from historical data.

  • Real-world uses:

    • Retail forecasting, finance risk assessment, manufacturing maintenance, marketing optimization.

Goal-Driven Systems (Autonomy & Optimization)

  • Purpose: Find the optimal solution to a problem through trial and error and reinforcement learning.

  • Objective: Find the best outcome by exploring many scenarios and iterations.

  • Technologies: Reinforcement learning, simulation, optimization, planning, perception, prediction.

  • Applications:

    • Scenario simulation and scenario planning

    • Game playing and strategy optimization

    • Resource optimization (money, equipment, time, other resources)

    • Robo-advising and real-time bidding for ads

  • Key concept: Machine learning can discover hidden rules and strategic moves that lead to optimal results.

  • Real-world examples:

    • Deep reinforcement learning systems that learn to play complex games and optimize real-time decisions.

Autonomous Systems

  • Definition: Systems (physical or digital) that accomplish tasks and achieve goals with minimal or no human involvement.

  • Objective: Minimize human labor and intervention.

  • Example use cases:

    • Autonomous vehicles and drones

    • Autonomous robots and software systems

    • Autonomous business processes (software bots making independent decisions)

  • Important distinctions:

    • Automation vs. Intelligence: Automation repeats tasks; intelligence implies perception, prediction, planning and handling variability without human intervention.

    • The question to ask: Is there any ML in the system? Can the system improve over time and handle next steps without exceptions? If not, it’s not truly intelligent.

  • Note on levels and terms:

    • Automation vs. Autonomy: automation is often rule-based and can be unattended or attended; autonomy encompasses perception, prediction, planning.

    • Cobots (collaborative robots) are designed to work with humans in close proximity to augment human capabilities.

Patterns & Anomalies (Pattern Recognition & Anomaly Detection)

  • Purpose: Detect patterns and outliers in large data sets; identify data points that don’t fit patterns.

  • Technologies: Pattern recognition, anomaly detection, classification.

  • Applications:

    • Fraud detection and risk assessment in finance and commerce

    • Automatic error detection and correction (data quality, medical prescriptions, etc.)

    • Intelligent monitoring of systems (IT, cyber-physical systems)

    • HR/recruiting candidate screening and profiling

    • Content moderation and safety analytics

  • Key concepts:

    • Classification: Group data into categories

    • Anomaly detection: Identify data points deviating from the norm

  • Real-world example: Walmart’s discovery that strawberry Pop-Tarts sales spike before hurricanes (pattern discovery in time-series data).

Hyperpersonalization

  • Purpose: Treat each individual as an individual using adaptive models that personalize experiences over time.

  • Technologies: User profiling, adaptive learning, recommendation systems, perceptual ML pipelines.

  • Applications:

    • Personalized content delivery

    • Personalized recommendations and product suggestions

    • Behavioral profiling

    • Personalized medicine, finance, education

  • Key terms:

    • Personalization: Customizing offerings for user groups

    • Hyper-personalization: Customizing for each individual

    • Recommendation system: Suggests products, content, or actions based on user profile and behavior

  • Real-world uses: Social media feeds, online shopping, healthcare, finance, education

  • Summary: Personalization-driven experiences intensify engagement and relevance by leveraging individual-level data and adaptive models.

Robotic Process Automation (RPA) and Intelligent Process Automation (IPA)

  • RPA purpose: Automate repetitive software tasks at the UI level; handle back-office and front-office workflows.

  • Types:

    • Attended Bots: Assist humans in real-time; interact with users to speed up tasks.

    • Unattended Bots: Run in the background; operate autonomously on schedules or rules.

  • Technologies: Screen recording, scripting, low-code/no-code platforms, rules-based automation.

  • Applications: Data entry, invoice processing, data transfers between systems, customer support workflows, document processing.

  • IPA (Intelligent Process Automation): Add AI to RPA to handle variability, unstructured data, exceptions; helps automate perception, prediction, and planning within processes.

  • Levels of intelligent automation (illustrative ladder): Level 0 (basic automation) to Level 3+ (fully autonomous optimization), with increasing capability to handle data and exceptions.

  • Platform concepts:

    • Low-code/No-code: Enables citizen developers to build automation without heavy coding.

    • Attended vs. Unattended: Distinct modes of operation and collaboration with humans.

Robotics, Automation, and the 4 D’s / 3K’s of Robotics

  • Robotics: Engineering discipline to design, build, operate, and apply robots

  • 4 D’s / 3K’s (why automation is needed in manufacturing): Dull, Dangerous, Dirty, Demeaning (and sometimes Dehumanizing)

  • Cobots (Collaborative Robots): Robots designed to work alongside humans to augment capabilities; not simply isolating humans from machines.

  • Note: Robotics and automation are not inherently AI; many robots are programmed or use simple detection systems without AI.

Levels of Automation & Autonomy (Examples)

  • Autonomy levels (example: autonomous vehicles): Level 0 to Level 5, where Level 0 = no autonomy and Level 5 = fully driverless with no human involvement; Levels 1–4 represent increasing autonomous capabilities and decreasing human intervention.

    • Level 0: No autonomous features

    • Level 1: One autonomous function (e.g., automatic braking)

    • Level 2: Two or more automated functions; human remains in control

    • Level 3: Capable of dynamic driving but requires human intervention

    • Level 4: Driverless in controlled environments

    • Level 5: Fully autonomous across all environments

Autonomous Retail & Examples

  • Autonomous Retail: Removing the human from the loop in retail contexts

  • Examples:

    • Amazon Go

    • LoweBot (intelligent store assistant, 2016)

    • Walmart shelf-scanning bots (2017) — note: faced challenges and failures

    • Ongoing deployment of store bots across various chains

  • Future possibility: Autonomous bot baristas and related experiments

Goal-Driven Systems: Reinforcement Learning & Game Playing

  • Objective: Find the optimal solution to a problem through trial and error using reinforcement learning in real-world settings

  • Real-world uses:

    • Scenario simulation

    • Game playing (board games, video games)

    • Resource optimization (money, equipment, time, other resources)

    • Iterative problem solving and robo-advising

    • Bidding and real-time auctions

  • Notable examples in AI history:

    • DeepMind and reinforcement learning breakthroughs

    • AlphaGo and AlphaZero achievements in Go and other games

DeepMind, AlphaGo & AlphaZero

  • AlphaGo: AI designed to play Go using reinforcement learning and deep learning; defeated top human player Lee Sedol in 2016.

  • AlphaZero: Built on the success of AlphaGo; trained purely by self-play; surpassed AlphaGo in 24 hours and beat the earlier system at superhuman levels.

  • Significance: Demonstrated that self-play can yield superhuman performance in complex tasks, and that a single algorithmic framework can master multiple domains.

Combining Patterns for Applications

  • Many real-world applications combine multiple AI patterns to achieve outcomes.

  • Example: Assistant Enabled Commerce

    • Pattern(s) involved: Hyperpersonalization (personalized data-driven tailoring), Conversation (dialog between user and bot), Pattern recognition (analyzing buying behavior) to deliver tailored product suggestions and guided conversations.

Quiz Highlights and Key Takeaways

  • Automation vs. Intelligence: Automation repeats tasks; intelligence perceives, predicts, and plans to handle variability.

  • The seven patterns are a useful framework for planning AI projects and understanding data needs and algorithms.

  • Recognize that some problems are better solved by traditional programming, rules-based systems, or human labor rather than AI.

  • AI benefits from data: Without data or reliable training data, ML/AI solutions may underperform.

Natural Language Processing: NLP, NLU, NLG, ASR, STT, and TTS

  • NLP (Natural Language Processing): Getting machines to understand and communicate using human language; umbrella term.

  • NLU (Natural Language Understanding): Adds deeper understanding of intent, entities, and semantics; part of NLP.

  • NLG (Natural Language Generation): Generating human-like text from data; a complement to NLP, adds semantic understanding to speech/text generation.

  • ASR (Automatic Speech Recognition): Converts spoken language to text; often used in STT contexts and chatbot/voice assistant pipelines; often treated as a component of NLU.

  • STT (Speech-to-Text) and TTS (Text-to-Speech): STT converts audio to text; TTS converts text to spoken audio.

  • Content summarization & analysis relies on NLU to extract key information and produce concise semantic summaries.

ImageNet, Computer Vision, and Vision Challenges

  • ImageNet: Created in 2006 by Fei-Fei Li; a large free repository with over 14{,}000{,}000 labeled images organized by WordNet hierarchy for training and benchmarking computer vision systems.

  • ImageNet Large Scale Visual Recognition Challenge (ILSVRC) started in 2010 to benchmark object detection/classification performance.

  • Concerns: Data quality and bias — estimates have suggested that up to 5\% of ImageNet data may be mislabeled or biased, highlighting training data quality importance.

  • Computer Vision components include: image acquisition, processing, analysis, object detection, recognition, segmentation, scene reconstruction, event detection, motion estimation, 3D modeling, image restoration, and indexing.

  • Applications: Facial recognition, object detection, gesture analysis, medical imaging, document processing, robotics, entertainment, and more.

Handwriting/Text Recognition, Gesture Recognition & Motion Analysis

  • Handwriting/Text Recognition: Recognizing handwritten content beyond OCR; can train on images of unstructured data; used in check imaging, intelligent document processing, forms extraction.

  • Gesture Recognition & Motion Analysis: Interpreting human gestures as commands; use cases in gaming, retail experiences, virtual try-ons, and surgical assistance.

  • Motion analysis supports pose estimation, 3D understanding, and tool/gesture-based interfaces.

Sound & Audio Recognition

  • Sound/Audio Recognition: Recognizing sounds, music, instruments, and language components; identifying songs, patterns, and environmental sounds.

  • Applications include music recognition, language identification, animal sounds, and cross-language recognition tasks.

Computer Vision: Core Concepts & Capabilities

  • Computer Vision enables machines to interpret and understand visual inputs as part of the Recognition Pattern.

  • Core components include image acquisition, processing, analysis, object detection, recognition, segmentation, scene understanding, motion estimation, 3D reconstruction, and more.

The Recognition Pattern in Depth: Real-World Uses

  • Facial recognition in security and consumer devices

  • Object recognition and image classification in surveillance, retail, and manufacturing

  • Handwriting/Text recognition in document processing

  • Gesture recognition in customer analytics and interactive experiences

  • Sound recognition in media and safety applications

  • Document processing and intelligent document processing (IDP) when combining handwriting, text extraction, and table recognition

The Conversation Pattern: Real-World Interfaces

  • Conversational interfaces include chatbots and voice assistants that converse with users in natural language

  • They rely on a combination of NLP, NLU, NLG, STT, and TTS to provide engaging interactions

Practical Examples Across Patterns

  • Chatbots and voice assistants for customer service

  • Content generation and natural language generation in consumer platforms

  • Sentiment/mood/intent analysis in social listening and feedback systems

  • Machine translation bridging languages across contexts

  • Content summarization and semantic analysis for long-form documents

  • Content recommendations and personalization driven by hyperpersonalization patterns

Content Summarization & Analysis

  • Purpose: Use AI/ML to produce concise overviews of bodies of text or other content

  • Approach: Use NLU to extract key information and generate semantic summaries

Quiz Highlights (Key Takeaways)

  • Which one is NOT a pattern of AI? Automation is not a pattern; it is a form of automation, not AI per se.

  • The first real conversational chatbot timeline and the number of AI patterns (7) are factual checkpoints to remember.

Key Definitions and Concepts (glossary)

  • Automation: Repetitive task execution, not necessarily intelligent.

  • Autonomous Systems: Minimize human involvement; require perception, prediction, and planning.

  • Computer Vision: Techniques to interpret visual data using ML/DL.

  • NLP: Teaching computers to understand and generate human language.

  • NLU: Understanding intent, context, and entities in language data.

  • NLG: Generating readable human language from data.

  • Sentiment Analysis: Classifying text by mood, intent, or opinion.

  • RPA: Software bots automating business processes via UI-level automation.

  • Attended vs. Unattended Bots: Collaboration vs. autonomous background operation.

  • Low-code/No-code: Platforms enabling automation with minimal or no programming.

  • Cobot: Collaborative robot; works alongside humans.

  • IPA: Intelligent Process Automation; RPA enhanced with AI for greater variability handling and autonomy.

  • Level-based autonomy (vehicles example): Levels 0–5 define increasing degrees of self-operation without human input.

References for further study

  • ImageNet and ILSVRC background discussions

  • DeepMind AlphaGo & AlphaZero milestones

  • RPA/IPA strategies in modern enterprises

  • Distinctions among NLP, NLU, NLG, ASR, STT, and TTS

  • The differences between automation, autonomous systems, and AI intelligence