AI Systems Landscape

Recommendation / Retrieval AI — Interactive Architecture Chart

A comprehensive interactive exploration of Recommendation AI — the multi-stage retrieval pipeline, 8-layer stack, collaborative filtering, embeddings, vector search, benchmarks, market data, and more.

~52 min read · Interactive Reference

Hameem M Mahdi, B.S.C.S., M.S.E., Ph.D. · 2026

Senior Principal Applied Scientist | Private Equity Leader | AI Innovative Solutions

📄 Forthcoming Paper

Multi-Stage Retrieval Pipeline

Modern recommendation systems use a multi-stage funnel that progressively narrows the candidate set — from millions of items to a handful of personalised results served in real time.

Candidate Generation Millions → Thousands Scoring / Ranking Thousands → Hundreds Re-Ranking Hundreds → Tens Business Rules / Filtering Policy & Constraint Application Serving Personalised Results

Click a stage above to learn more

Each stage in the retrieval funnel progressively filters and refines candidates to deliver personalised recommendations with low latency.

Did You Know?

1

Netflix's recommendation engine drives 80% of content watched on the platform.

2

Amazon attributes 35% of its revenue directly to its recommendation algorithm.

3

Spotify's Discover Weekly playlist, powered by collaborative filtering, reaches 100+ million users weekly.

Knowledge Check

Test your understanding — select the best answer for each question.

Q1. What is collaborative filtering?

Q2. What does RAG stand for?

Q3. What is the "cold start" problem in recommendation systems?

8-Layer Recommendation Stack

A full-stack view of modern recommendation systems — from raw data ingestion through to live A/B experimentation. Click any layer to expand details.

8
A/B Testing & Experimentation
Online experimentation platforms for measuring recommendation quality. Includes interleaving experiments, multi-armed bandits for explore/exploit trade-offs, and statistical significance testing for model comparisons.
7
Model Serving
Real-time inference infrastructure serving recommendations under strict latency SLAs (<50 ms). Includes feature store lookups, model caching, GPU/CPU serving clusters, and dynamic batching for throughput optimisation.
6
Re-Ranking & Policies
Post-scoring re-ranking for diversity, freshness, and business rule enforcement. Applies fairness constraints, content deduplication, publisher diversity quotas, and regulatory compliance filters before final slate assembly.
5
Scoring Models
Deep ranking networks that score candidate relevance. Architectures include two-tower models, cross-attention networks, DLRM (Deep Learning Recommendation Model), and DCN (Deep & Cross Network) for capturing complex feature interactions.
4
Candidate Retrieval
Approximate Nearest Neighbour (ANN) search over embedding spaces, collaborative filtering recall sets, and content-based retrieval. Fetches thousands of candidates from billions of items using HNSW, IVF, or ScaNN indices.
3
Feature Engineering
User features (demographics, history), item features (metadata, embeddings), interaction features (click/skip/dwell), and contextual signals (time, device, location). Feature stores ensure consistent online/offline feature computation.
2
Data Pipelines
Ingestion of clickstreams, impressions, purchases, session logs, and explicit ratings. ETL/ELT pipelines process terabytes of behavioural data daily. Streaming pipelines (Kafka, Flink) enable near-real-time feature updates.
1
User & Item Catalogues
Foundational data layer: user profiles (preferences, demographics, subscription tier), product/content metadata (title, category, tags, descriptions), and content taxonomy/ontology for structured item organisation.

Recommendation Sub-Types

Ten major families of recommendation and retrieval systems — from classical collaborative filtering to modern retrieval-augmented generation.

Memory-Based

Collaborative Filtering (User-Based)

Similar users like similar items. Uses user-user similarity matrices to find neighbours with overlapping preferences. Foundational approach behind early Netflix recommendations. Struggles with cold-start and sparsity at scale.

Item Similarity

Collaborative Filtering (Item-Based)

Computes item-item similarity from co-occurrence in user histories. Powers Amazon's "Customers who bought this also bought" feature. More stable than user-based CF since item relationships change less frequently.

Latent Factors

Matrix Factorisation

Decomposes the sparse user-item interaction matrix into low-rank latent factor matrices via SVD or ALS. Won the Netflix Prize ($1M). Captures latent taste dimensions — e.g., preference for art-house vs. action films.

Feature Matching

Content-Based Filtering

Matches item feature vectors (TF-IDF, embeddings, metadata) to learned user profiles. No dependency on other users' data — works for new users if item features are rich. Common in news and document recommendation.

Constraint-Based

Knowledge-Based

Uses explicit domain knowledge and constraints for high-value, infrequent purchases (cars, real estate, financial products). Case-based reasoning matches past solutions. No cold-start problem since recommendations are constraint-driven.

Ensemble

Hybrid Recommenders

Combines collaborative filtering, content-based, and knowledge-based signals. Strategies include weighted blending, switching (choose method by context), cascading (coarse → fine), and stacking (meta-learner over base models).

Neural

Deep Learning Recommenders

Two-tower models, DLRM, DCN, and autoencoders that learn complex non-linear feature interactions from massive datasets. Handle sparse categorical features via learned embedding tables. Power modern production systems at scale.

Temporal

Sequential / Session-Based

Models user interaction sequences with GRU4Rec, SASRec, and BERT4Rec. Captures temporal dynamics — what you clicked 5 minutes ago matters more than last month. Critical for e-commerce sessions and music playlists.

Dialogue

Conversational Recommenders

Dialogue-driven preference elicitation: the system asks clarifying questions to narrow preferences ("Do you prefer sci-fi or drama?"). Reduces cold-start and improves user satisfaction through interactive refinement of recommendations.

LLM + Retrieval

Retrieval-Augmented Generation (RAG)

Combines large language models with vector retrieval for knowledge-grounded generation. Retrieved documents are injected as context to reduce hallucination. Powers enterprise search, customer support, and knowledge management systems.

Core Architectures

Seven foundational model architectures that power modern recommendation and retrieval systems at scale.

Dual Encoder

Two-Tower Model

Separate user and item encoder towers produce embeddings; relevance scored via dot product or cosine similarity. Enables pre-computation of item embeddings for sub-millisecond ANN retrieval. Used in YouTube DNN, Google, and Spotify.

Meta

DLRM (Deep Learning Recommendation Model)

Meta's architecture combining sparse categorical features (via embedding tables) with dense numerical features through bottom MLPs and feature interaction layers. Handles click-through rate prediction at trillion-scale interactions.

Google

DCN (Deep & Cross Network)

Explicit feature cross layers that learn bounded-degree interactions alongside deep layers for implicit patterns. Efficiently captures high-order feature crosses without exponential parameter growth. Used in Google's ad ranking systems.

Hybrid

Wide & Deep

Memorisation (wide linear model with cross-product features) + generalisation (deep neural network). Originally deployed for Google Play app recommendations. Balances learning specific feature co-occurrences with broad generalisable patterns.

Attention

Transformer-based Sequential

SASRec and BERT4Rec apply self-attention over user interaction histories. Captures long-range dependencies and position-aware item relationships. Outperforms RNN-based sequential models on most benchmarks for next-item prediction.

Graph

Graph Neural Networks

User-item bipartite graphs with message passing for embedding learning. PinSage (Pinterest) scales to billions of nodes via random-walk sampling. Captures social signals, co-purchase patterns, and multi-hop relational information.

Vector Search

Approximate Nearest Neighbour (ANN)

HNSW, IVF, and ScaNN algorithms for billion-scale vector similarity search. Trade small recall loss for orders-of-magnitude speed gains. Foundation of the retrieval stage — enabling sub-10ms candidate generation from massive catalogues.

Tools & Platforms

Key tools, services, and frameworks powering recommendation and retrieval systems in production — from managed cloud services to open-source libraries.

ToolProviderFocus
Amazon PersonalizeAWSManaged recommender service; real-time personalisation
Google Recommendations AIGoogleRetail-focused; managed; Discovery AI
Merlin / NVTabularNVIDIAGPU-accelerated RecSys training + feature engineering
FAISSMetaBillion-scale ANN vector search; GPU-optimised
PineconePineconeManaged vector database; serverless; hybrid search
WeaviateWeaviateOpen-source vector DB; hybrid search; modules
MilvusZillizOpen-source vector DB; distributed; cloud-native
QdrantQdrantRust-based vector DB; filtering + payload search
AlgoliaAlgoliaSearch-as-a-service; instant search; ranking rules
ElasticsearchElasticFull-text + vector search; kNN; hybrid retrieval
LensKitOpen-sourcePython RecSys toolkit; evaluation; reproducible research
SurpriseOpen-sourcePython CF library; SVD, KNN, baselines
RecBoleOpen-sourceUnified RecSys framework; 70+ models; benchmarking
LlamaIndexLlamaIndexRAG framework; data connectors; retrieval pipelines

Use Cases

How recommendation and retrieval AI is deployed across industries — from e-commerce to enterprise knowledge management.

E-Commerce Product Recommendations
  • Powers "You may also like" and "Frequently bought together" on Amazon, Shopify, and Alibaba
  • Drives an estimated 35% of Amazon's total revenue through recommendation engines
  • Uses hybrid collaborative filtering + deep ranking for personalised product feeds
  • Real-time re-ranking based on session context, cart contents, and purchase history
Video / Music Streaming
  • Netflix personalises artwork, row ordering, and content selection for 250M+ subscribers
  • YouTube's two-tower DNN generates candidates from billions of videos in milliseconds
  • Spotify Discover Weekly uses collaborative filtering + audio embeddings for music discovery
  • Watch/listen history and implicit signals (skip, replay, completion) drive personalised feeds
Social Media Feeds
  • TikTok's For You page uses a multi-stage funnel with heavy sequential modelling
  • Instagram Explore surfaces content from unfollowed creators based on engagement patterns
  • Twitter/X timeline ranking blends social graph signals with content relevance scoring
  • Engagement optimisation balanced with content diversity and safety policies
Job / Talent Matching
  • LinkedIn matches job seekers and recruiters using two-sided recommendation models
  • Skill-role compatibility scoring with dense embeddings of job descriptions and resumes
  • Indeed and Glassdoor personalise job suggestions based on search history and application patterns
  • Two-tower models encode candidates and roles separately for scalable matching
News & Content Personalisation
  • Google News and Apple News blend freshness, relevance, and diversity signals
  • Real-time re-ranking to surface breaking stories and reduce stale content
  • Content-based filtering with NLP embeddings of article text and user reading history
  • Editorial diversity quotas to mitigate filter bubbles and ensure topic breadth
Enterprise Search / RAG
  • Internal knowledge retrieval powered by Glean, Coveo, and custom RAG pipelines
  • LLM-augmented search: retrieve relevant documents then generate grounded answers
  • Vector databases index internal wikis, Slack messages, tickets, and documentation
  • Hybrid search combining lexical (BM25) and semantic (dense embedding) retrieval

Benchmarks

Performance benchmarks for recommendation quality and vector search efficiency across standard datasets and systems.

RecSys Benchmarks (NDCG@10)

ANN Search (Recall@10 at QPS)

Market Data

Market sizing and growth projections for the recommendation, personalisation, and retrieval AI ecosystem.

Market Segments ($B)

Market Growth 2024 → 2030 (CAGR ~20%)

Risks & Challenges

Key risks and ethical concerns in deploying recommendation and retrieval AI systems at scale.

Filter Bubbles

Over-personalisation narrows user exposure to a shrinking set of topics and viewpoints, creating echo chambers that reinforce existing beliefs and reduce serendipitous discovery.

Cold Start

New users and items have no interaction history, leading to poor initial recommendations. Workarounds include content-based fallbacks, knowledge-based methods, and active preference elicitation.

Popularity Bias

Systems disproportionately favour popular items with more interaction data, suppressing long-tail items and niche creators. Calibration and diversity-aware re-ranking help counteract this bias.

Privacy & Tracking

Behavioural data collection (clicks, dwell time, purchase history) raises consent and surveillance concerns. GDPR, CCPA, and emerging AI regulations demand transparency and user control over personal data usage.

Manipulation & Gaming

Sellers, creators, and bad actors game recommendation algorithms for visibility through fake reviews, click farms, and engagement manipulation — degrading recommendation quality for all users.

Fairness & Diversity

Systematic under-recommendation of minority content and creators. Feedback loops amplify historical biases in training data. Fairness-aware algorithms and auditing frameworks are critical safeguards.

Glossary

Key terms in recommendation and retrieval AI — search to filter.

ANNApproximate Nearest Neighbour; fast similarity search that trades small recall loss for massive speed gains over exact search.
Collaborative FilteringRecommending items based on the preferences of similar users (user-based or item-based).
Content-Based FilteringRecommending items similar to those a user has liked, based on item features.
Hybrid RecommenderCombining collaborative and content-based approaches for improved recommendation accuracy.
Matrix FactorisationDecomposing the user-item interaction matrix into latent factor representations.
EmbeddingDense vector representation of users, items, or queries in a continuous semantic space.
RAGRetrieval-Augmented Generation — retrieving relevant documents to ground and enhance generative model outputs.
Vector SearchFinding nearest neighbours in high-dimensional embedding spaces using approximate algorithms (HNSW, IVF).
Two-Tower ModelArchitecture with separate encoders for queries and items, compared via inner product or cosine similarity.
Cold Start ProblemDifficulty recommending for new users or items with insufficient interaction history.
Click-Through RateFraction of users who click on a recommended item — primary online metric for recommender systems.
A/B TestingControlled experiment comparing two variants to measure the impact of changes on user behaviour.
PersonalisationTailoring content, products, or experiences to individual user preferences and behaviour.
HNSWHierarchical Navigable Small World — graph-based algorithm for fast approximate nearest neighbour search.
Re-RankingSecond-stage scoring of retrieved candidates incorporating business rules, diversity, or fairness constraints.
Session-Based RecommendationRecommending based on the current browsing session without long-term user profiles.
Click-Through Rate (CTR)Fraction of impressions resulting in clicks; primary online metric for ranking model evaluation.
Collaborative FilteringRecommendations derived from collective user behaviour patterns — users who agreed in the past will agree in the future.
Cold StartThe challenge of recommending for new users or items with no interaction history.
Content-Based FilteringRecommendations based on matching item features (metadata, embeddings) to a user's preference profile.
EmbeddingDense vector representation of users or items in a continuous latent space, enabling similarity computation via dot product or cosine distance.
HNSWHierarchical Navigable Small World graph — a state-of-the-art algorithm for approximate nearest neighbour search with high recall and speed.
Hybrid RecommenderA system combining multiple recommendation strategies (CF, content-based, knowledge-based) for improved accuracy and coverage.
Implicit FeedbackInferred user signals (clicks, views, dwell time, skips) as opposed to explicit ratings. More abundant but noisier data source.
Matrix FactorisationDecomposing the sparse user-item interaction matrix into low-rank latent factor matrices to predict missing entries.
NDCGNormalised Discounted Cumulative Gain; a ranking quality metric that accounts for position-dependent relevance — items ranked higher contribute more.
PersonalisationTailoring content, results, and experiences to individual user preferences, context, and behaviour history.
RAGRetrieval-Augmented Generation; grounding large language model outputs with retrieved documents to reduce hallucination and improve factuality.
Two-Tower ModelDual-encoder architecture with separate user and item towers producing embeddings for scalable dot-product retrieval.
Vector DatabaseSpecialised database optimised for storing, indexing, and querying high-dimensional embedding vectors at scale.

Visual Infographics

Animation infographics for Recommendation / Retrieval AI — overview and full technology stack.

Regulation

Detailed reference content for regulation.

Regulation & Governance

Key Regulatory Frameworks

Regulation Jurisdiction Relevance to Recommendation / Retrieval AI
EU Digital Services Act (DSA) EU Mandates transparency of recommender systems; requires non-profiling-based recommendation option
EU AI Act EU Recommender systems may be classified as limited or high risk depending on deployment context
GDPR EU Consent for profiling; right to explanation; data minimisation for personalisation
California CPRA US (CA) Consumer right to opt out of profiling and automated decision-making
UK Online Safety Act UK Platforms must address algorithmic amplification of harmful content
FTC Section 5 US Unfair or deceptive algorithmic practices; ad targeting discrimination
China Algorithmic Recommendation Regulations China Requires algorithm registration; user opt-out; transparency of recommendation logic

Transparency Requirements

Requirement Description
Algorithmic Transparency Explain the main parameters and criteria used by recommender systems (DSA Art. 27)
Non-Profiling Alternative Offer a recommendation option not based on user profiling (DSA Art. 38)
Audit Access Provide researcher and regulator access to recommendation system data (DSA Art. 40)
User Notification Inform users when content is recommended vs. organically surfaced
Ad Library & Transparency Maintain public archives of targeted advertising and recommendation criteria

Deep Dives

Detailed reference content for deep dives.

Neural Information Retrieval — Deep Dive

The Evolution from Lexical to Semantic Search

Generation Era Approach Key Technology Limitation
1st 1990s Boolean keyword matching Inverted indices No ranking; exact match only
2nd 2000s Statistical term weighting TF-IDF, BM25 Lexical gap — misses synonyms and paraphrases
3rd 2018+ Dense neural retrieval BERT, DPR, ColBERT Computationally expensive; requires training data
4th 2024+ Generative retrieval & RAG Differentiable search indices, LLM + retrieval Active research area; architectures still evolving

Key Dense Retrieval Models

Model Architecture Key Innovation
DPR Dual BERT encoders First effective dense retriever; outperformed BM25 on open-domain QA
ColBERT Late interaction dual encoder Token-level interaction for fine-grained matching; fast with pre-computation
Contriever Contrastive unsupervised BERT No labelled data needed for training; strong zero-shot retrieval
E5 Unified text embedding model Instruction-tuned for diverse retrieval tasks
BGE / GTE BERT-based general embeddings Open-source; competitive with proprietary embedding models
OpenAI Embeddings text-embedding-3-large High-quality proprietary embeddings; 3072 dimensions
Cohere Embed v3 Multi-stage trained embedding Supports 100+ languages; compression-friendly
Google Gecko Distilled from large LM Compact embedding model; efficient for on-device retrieval

Vector Databases & ANN Search

System Type Key Features
Pinecone Managed vector DB Fully managed; real-time indexing; metadata filtering
Weaviate Open-source vector DB Hybrid search (vector + keyword); multi-modal; GraphQL API
Qdrant Open-source vector DB Rust-based; fast and memory-efficient; filtering during search
Milvus / Zilliz Open-source vector DB Large-scale; distributed architecture; GPU-accelerated
Chroma Lightweight vector DB Developer-friendly; embedded or client-server; popular for RAG
pgvector PostgreSQL extension Vector search inside existing Postgres infrastructure
Elasticsearch / ESRE Hybrid search engine BM25 + dense vector search; enterprise standard
Google Vertex AI Search Managed search + RAG Grounding + retrieval + ranking in one managed service
FAISS (Meta) ANN library Industry-standard ANN search library; GPU-optimised; billions of vectors

Retrieval-Augmented Generation (RAG) — Deep Dive

RAG bridges Recommendation/Retrieval AI and Generative AI — using retrieval to ground generative models in real, verifiable information.

RAG Architecture

+------------------------------------------------------------------------+
| RAG PIPELINE |
| |
| USER QUERY --> RETRIEVER --> TOP-K DOCUMENTS --> LLM GENERATOR |
| (dense / (relevant (generates answer |
| sparse / context from grounded in |
| hybrid) knowledge base) retrieved docs) |
+------------------------------------------------------------------------+

RAG Components

Component Role Key Technologies
Document Ingestion Chunk, embed, and index source documents LangChain, LlamaIndex, Unstructured
Embedding Model Convert text chunks into dense vectors OpenAI Embeddings, Cohere Embed, E5, BGE
Vector Store Store and retrieve embeddings by similarity Pinecone, Weaviate, Qdrant, Chroma, pgvector
Retriever Find the most relevant chunks for a query Dense, sparse, or hybrid retrieval
Re-Ranker Re-score retrieved chunks for fine-grained relevance before passing to the LLM Cross-encoders (Cohere Rerank, BGE Reranker)
Generator (LLM) Synthesise an answer from the retrieved context GPT-4, Claude, Gemini, Llama, Mistral
Grounding / Citation Map generated claims back to source documents for verifiability Source attribution layers, inline citations

Advanced RAG Patterns

Pattern Description
Naive RAG Simple retrieve-then-generate; single retrieval pass
Advanced RAG Query rewriting, multi-step retrieval, re-ranking, chunk optimisation
Modular RAG Composable pipeline with pluggable retriever, reranker, and generator components
Corrective RAG (CRAG) Evaluates retrieved documents for relevance; triggers web search if quality is low
Self-RAG LLM decides when to retrieve, what to retrieve, and whether retrieved docs are useful
Graph RAG Combines knowledge graph traversal with vector retrieval for structured + unstructured data
Agentic RAG Agent loop that iteratively queries, evaluates, and refines retrieval
Multi-Modal RAG Retrieves across text, images, tables, and other modalities

Real-Time Personalisation & Session-Based Recommendation

Session-Based Models

Model / Approach Architecture Key Innovation
GRU4Rec GRU (Recurrent Neural Network) First neural session-based recommender; models click sequences
SASRec Self-Attention (Transformer) Applies self-attention to user action sequences; captures long-range deps
BERT4Rec Masked Transformer Bidirectional self-attention for sequential recommendation
Transformers4Rec (NVIDIA) Modular Transformer framework Production-ready; supports multiple architectures and feature types
Recbole Unified framework 90+ recommendation algorithms in a standardised framework

Contextual Bandits for Exploration

Aspect Detail
Core Mechanism Model recommendation as an explore/exploit trade-off; learn from partial feedback
Why It Matters Overcomes popularity bias; discovers niche content that greedy ranking would never surface
Key Algorithms LinUCB, Thompson Sampling, epsilon-greedy, neural contextual bandits
Real-World Usage Spotify Discover Weekly, news personalisation, ad selection, homepage curation
Connection to RL Contextual bandits are a simplified (single-step) form of reinforcement learning

Overview

Detailed reference content for overview.

Definition & Core Concept

Recommendation and Retrieval AI is the branch of artificial intelligence focused on systems that find, rank, and present the most relevant items from large collections — matching users to products, content, documents, or search results based on preferences, behaviour, and context. It is arguably the most widely deployed form of AI in production today, powering the core experience of Google Search, Netflix, Amazon, Spotify, YouTube, TikTok, LinkedIn, and virtually every digital platform.

Retrieval and recommendation are two sides of the same coin. Retrieval AI focuses on finding relevant items in response to a query (search). Recommendation AI focuses on proactively surfacing items a user is likely to want, often without an explicit query. Modern systems blur this boundary: a Netflix homepage is recommendation without a query; a YouTube search is retrieval with personalisation; and RAG (retrieval-augmented generation) is retrieval embedded inside generative AI.

The defining characteristic is selection from an existing corpus — the system does not create new content (Generative AI), predict a numeric outcome (Predictive AI), or reason about goals (Agentic AI). It selects, ranks, and presents what already exists.

Dimension Detail
Core Capability Retrieves and ranks — surfaces the most relevant items from large catalogues for a given user or query
How It Works Collaborative filtering, content-based filtering, embedding-based retrieval, two-tower models, learning-to-rank
What It Produces Ranked lists of items, personalised feeds, search results, content recommendations, document retrievals
Key Differentiator Selects from what exists — it does not generate new content, predict a label, or pursue autonomous goals

Recommendation / Retrieval AI vs. Other AI Types

AI Type What It Does Example
Recommendation / Retrieval AI Surfaces relevant items from large catalogues based on user signals and queries Netflix suggestions, Google Search, Spotify Discover Weekly
Agentic AI Pursues goals autonomously using tools, memory, and planning Research agent, coding agent, autonomous workflow
Analytical AI Extracts insights and explanations from existing data Dashboard, root-cause analysis, anomaly detection
Autonomous AI (Non-Agentic) Operates independently within fixed boundaries without human input Autopilot, auto-scaling, algorithmic trading
Bayesian / Probabilistic AI Reasons under uncertainty using probability distributions Clinical trial analysis, A/B testing, risk modelling
Cognitive / Neuro-Symbolic AI Combines neural learning with symbolic reasoning LLM + knowledge graph, physics-informed neural net
Conversational AI Manages multi-turn dialogue between humans and machines Customer service chatbot, voice assistant
Evolutionary / Genetic AI Optimises solutions through population-based search inspired by natural selection Neural architecture search, logistics scheduling
Explainable AI (XAI) Makes AI decisions understandable to humans SHAP explanations, LIME, Grad-CAM
Generative AI Creates new original content from learned distributions Write an essay, generate an image, synthesise a video
Multimodal Perception AI Fuses vision, language, audio, and other modalities GPT-4o processing image + text, AV sensor fusion
Optimisation / Operations Research AI Finds optimal solutions to constrained mathematical problems Vehicle routing, supply chain planning, scheduling
Physical / Embodied AI Acts in the physical world through sensors and actuators Autonomous vehicle, robot arm, drone
Predictive / Discriminative AI Classifies or forecasts from historical patterns Fraud score, churn probability, demand forecast
Privacy-Preserving AI Trains and runs AI without exposing raw data Federated hospital models, differential privacy
Reactive AI Responds to current input with no learning or memory Chess engine, rule-based spam filter
Reinforcement Learning AI Learns optimal behaviour from reward signals via trial and error AlphaGo, robotic locomotion, RLHF
Scientific / Simulation AI Solves scientific problems and models physical systems AlphaFold, climate simulation, molecular dynamics
Symbolic / Rule-Based AI Reasons over explicit rules and knowledge to derive conclusions Medical expert system, legal reasoning engine

Key Distinction from Predictive AI: Predictive AI assigns a label, score, or forecast to an individual input. Recommendation AI selects and ranks items from a collection for a user — the output is a ranked list, not a single prediction.

Key Distinction from Generative AI: Generative AI creates new content. Recommendation AI selects from existing content. RAG bridges both by retrieving existing documents and feeding them to a generative model.