AI Systems Landscape

Predictive / Discriminative AI — Interactive Architecture Chart

A comprehensive interactive exploration of Predictive AI — the ML pipeline, 8-layer stack, classification/regression/forecasting, gradient boosting, benchmarks, market data, and more.

~63 min read · Interactive Reference

Hameem M Mahdi, B.S.C.S., M.S.E., Ph.D. · 2026

Senior Principal Applied Scientist | Private Equity Leader | AI Innovative Solutions

📄 Forthcoming Paper

ML Pipeline

The end-to-end machine learning pipeline for predictive / discriminative AI — from problem definition through monitoring and retraining.

1. Problem Definition

Define target, success metrics, constraints

2. Data Collection

Gather training data from DBs, APIs, logs

3. Data Cleaning / EDA

Handle missing values, outliers, distributions

4. Feature Engineering

Transform, encode, create informative features

5. Model Selection

Choose algorithm family, baseline models

6. Training

Fit model, tune hyperparameters, validate

7. Evaluation

Cross-val, hold-out, precision/recall, AUC

8. Deployment

REST/gRPC API, batch scoring, edge deploy

9. Monitoring / Retrain

Drift detection, alerts, scheduled retraining

Did You Know?

1

Random forests remain the most-used ML algorithm in Kaggle competitions, despite deep learning advances.

2

XGBoost's gradient boosting approach won more Kaggle competitions than any other single algorithm between 2015-2020.

3

Logistic regression, invented in 1958, still powers billions of real-time ad-targeting decisions daily.

Knowledge Check

Test your understanding — select the best answer for each question.

Q1. What does a discriminative model learn?

Q2. Which algorithm uses an ensemble of decision trees with boosting?

Q3. What metric measures the area under the ROC curve?

The Predictive AI Stack — 8 Layers

Click any layer to expand details about the components and technologies at each level of the predictive ML stack.

8Monitoring & Drift Detection
Feature drift detection, prediction drift alerts, data quality monitoring, model performance dashboards, automated retraining triggers.
7Model Serving
REST/gRPC APIs, batch scoring pipelines, edge inference, model versioning, A/B serving, canary deployments, latency optimization.
6Evaluation & Validation
K-fold cross-validation, stratified hold-out sets, A/B testing, champion-challenger evaluation, fairness audits, calibration checks.
5Model Training
Gradient descent optimization, hyperparameter tuning (grid, random, Bayesian), AutoML pipelines, distributed training, early stopping.
4Feature Engineering
Numerical transformations, categorical encoding, text embeddings, feature crosses, feature stores for online/offline serving, automated feature generation.
3Data Processing
ETL/ELT pipelines, data cleaning, missing value imputation, normalisation, standardisation, outlier handling, data versioning.
2Data Storage
Data lakes (S3, GCS), data warehouses (Snowflake, BigQuery), feature stores (Feast, Tecton), vector databases, Delta Lake / Iceberg.
1Data Sources
Transactional databases, application logs, event streams (Kafka), REST/GraphQL APIs, IoT sensors, third-party data providers.

Sub-Types of Predictive / Discriminative AI

Fifteen distinct prediction paradigms — from binary classification to zero-shot learning.

Classification

Binary Classification

Spam/not-spam, fraud/legit, churn/retain. Core methods: logistic regression, SVM, gradient boosting.

Classification

Multi-Class Classification

Image recognition, named entity recognition. Core methods: softmax, random forest, deep neural networks.

Classification

Multi-Label Classification

Document tagging, scene annotation. Multiple labels per instance; binary relevance, classifier chains.

Regression

Regression

House prices, demand forecasting, salary prediction. Linear, polynomial, neural regression models.

Forecasting

Time-Series Forecasting

Sales, weather, energy demand. ARIMA, Prophet, N-BEATS, temporal fusion transformers.

Detection

Anomaly Detection

Fraud, intrusion, defect detection. Isolation forest, autoencoders, one-class SVM.

Ranking

Ranking

Search results, recommendations. LambdaMART, RankNet, listwise learning-to-rank.

Survival

Survival Analysis

Churn prediction, reliability engineering. Cox proportional hazards, Kaplan-Meier estimator.

Vision

Object Detection

Bounding box localisation. YOLO, Faster R-CNN, DETR, anchor-free detectors.

Vision

Image Classification

Scene & object categorisation. CNN, Vision Transformer (ViT), EfficientNet.

Sequence

Sequence Labelling

NER, POS tagging, slot filling. CRF, BiLSTM-CRF, token-level transformers.

Structured

Structured Prediction

Parsing, semantic segmentation. Energy-based models, graph neural networks.

Calibration

Calibrated Prediction

Risk scoring with reliable probabilities. Platt scaling, isotonic regression, temperature scaling.

Ordinal

Ordinal Regression

Ratings, severity levels, ordered categories. Proportional odds model, threshold-based approaches.

Few-Shot

Zero-Shot / Few-Shot Classification

Generalise with minimal labels. CLIP, prompt-tuning, prototypical networks, meta-learning.

Core Architectures

The fundamental model families powering predictive AI — from interpretable baselines to automated ensembles.

Ensemble

Gradient Boosting

XGBoost, LightGBM, CatBoost. Sequential weak learners correcting predecessors. Tabular SOTA; dominant in competitions and production.

Ensemble

Random Forest

Bagging of decision trees with feature randomisation. Robust, minimal tuning, strong baseline for classification and regression.

Deep Learning

Deep Neural Networks

MLP, CNN, Transformer architectures. Learn representations end-to-end; excel on vision, sequence, and large-scale tabular data.

Kernel

Support Vector Machines

Kernel trick for non-linear boundaries. Max-margin classifier; effective on small-to-medium datasets with careful tuning.

Linear

Logistic / Linear Regression

Simple, interpretable, and fast inference. Excellent baseline; widely used in regulated industries for explainability.

Instance

k-Nearest Neighbours

Instance-based lazy learning; no explicit training phase. Predictions via distance metrics; works well with low-dimensional data.

Probabilistic

Naive Bayes

Probabilistic classifier with independence assumption. Fast text classification; strong baseline for NLP tasks.

AutoML

AutoML Ensembles

Auto-sklearn, H2O AutoML, AutoGluon. Automated model selection, hyperparameter tuning, and stacking for optimal performance.

Tools & Frameworks

The leading libraries, platforms, and services powering the predictive AI ecosystem.

ToolProviderFocus
scikit-learnOpen-sourceSwiss-army knife ML; preprocessing, models, evaluation
XGBoostDMLCGradient boosting; Kaggle champion; tabular data
LightGBMMicrosoftFast gradient boosting; histogram-based; large datasets
CatBoostYandexGradient boosting with native categorical support
PyTorchMetaDeep learning framework; dynamic graphs; research
TensorFlow / KerasGoogleProduction DL framework; TF Serving, TFLite
Hugging FaceHugging FaceTransformers hub; fine-tuning; model sharing
Amazon SageMakerAWSEnd-to-end ML platform; training, deployment, monitoring
Vertex AIGoogleGCP ML platform; AutoML, pipelines, model registry
Azure MLMicrosoftEnterprise ML platform; Designer, SDK, endpoints
MLflowDatabricksExperiment tracking, model registry, deployment
FeastTecton / LFOpen-source feature store; online/offline serving
Weights & BiasesW&BExperiment tracking, hyperparameter sweeps, artifacts
Evidently AIEvidentlyML monitoring; data drift, model quality dashboards

Use Cases by Industry

How predictive / discriminative AI creates value across major industry verticals.

Financial Services
Credit scoring and underwriting, real-time fraud detection, algorithmic trading signal generation, insurance pricing and claims prediction, anti-money laundering risk scoring.
Healthcare
Disease risk prediction, hospital readmission scoring, medical imaging diagnosis (X-ray, MRI, pathology), drug response prediction, clinical trial patient matching.
Retail & E-Commerce
Demand forecasting and inventory optimisation, customer churn prediction, dynamic pricing engines, market basket analysis, next-best-offer recommendations.
Manufacturing
Predictive maintenance for equipment failure, quality control and defect detection, yield optimisation, supply chain demand planning, energy consumption forecasting.
Marketing & AdTech
Click-through rate (CTR) prediction, customer lifetime value estimation, marketing attribution modelling, audience segmentation and targeting, bid optimisation.
Cybersecurity
Network intrusion detection, malware classification, phishing email detection, threat scoring and prioritisation, user and entity behaviour analytics (UEBA).

Benchmarks & Model Comparisons

Performance benchmarks for leading predictive model architectures on tabular tasks.

Tabular ML Benchmarks (Accuracy %)

Model Properties (Radar)

Market & Adoption Data

Market sizing and growth projections for the Predictive Analytics ecosystem.

Predictive Analytics Market ($B)

Market Growth 2024 → 2030 (CAGR 23%)

Risks & Limitations

Critical challenges and failure modes when deploying predictive / discriminative AI systems.

Bias & Fairness

Historical data encodes discrimination; models can produce disparate impact on protected groups without careful auditing.

Data Leakage

Target leakage or train-test contamination inflates evaluation metrics; models fail catastrophically in production.

Concept Drift

Data distribution shifts over time; model accuracy degrades silently without proper monitoring and retraining pipelines.

Overfitting

Model memorises training data noise rather than learning true patterns; poor generalisation to unseen data.

Feature Attribution

Black-box models cannot explain individual predictions; creates regulatory risk in finance, healthcare, and hiring.

Adversarial Examples

Small, carefully crafted input perturbations can flip predictions; represents a security vulnerability in deployed models.

Key Terminology Glossary

Essential predictive / discriminative AI terminology — searchable.

AUC-ROCArea Under the Receiver Operating Characteristic curve; measures classifier discrimination across all thresholds.
BaggingBootstrap Aggregating; parallel ensemble method training models on random data subsets to reduce variance.
BoostingSequential ensemble technique where each learner corrects the errors of its predecessor.
CalibrationAligning predicted probabilities with actual outcome frequencies for reliable risk scoring.
Cross-ValidationSplitting data into K folds for robust model evaluation; each fold serves as validation once.
F1 ScoreHarmonic mean of precision and recall; balances false positives and false negatives.
Feature EngineeringCreating informative input features from raw data to improve model performance.
Gradient DescentIterative optimisation algorithm that adjusts parameters in the direction of steepest loss reduction.
HyperparameterConfiguration setting determined before training begins (e.g., learning rate, tree depth, regularisation).
ImputationFilling in missing values in a dataset using statistical methods (mean, median, KNN, MICE).
L1/L2 RegularisationPenalty terms (Lasso/Ridge) added to the loss function to prevent overfitting and improve generalisation.
OverfittingModel captures noise rather than signal in training data; indicates need for regularisation or more data.
PrecisionTrue positives divided by (True positives + False positives); measures prediction exactness.
RecallTrue positives divided by (True positives + False negatives); measures sensitivity / completeness.
SHAPSHapley Additive exPlanations; game-theoretic approach to explain individual feature contributions.

Visual Infographics

Animation infographics for Predictive / Discriminative AI — overview and full technology stack.

Regulation

Detailed reference content for regulation.

Regulation & Governance

Predictive AI has been regulated longer than generative and agentic AI because it is most commonly deployed for high-stakes decisioning — credit, hiring, insurance, healthcare, and law enforcement. The 2025–2026 regulatory wave formalises documentation, transparency, and risk-management requirements that many regulated industries already partially implement.

EU AI Act — Predictive AI Implications

Many predictive models fall under the high-risk classification when deployed in employment, education, credit, insurance, public benefits, healthcare, biometric identification, or law enforcement contexts.

Requirement Area Practical Implication for Predictive AI Systems
Risk Management System Formal risk process covering model development, deployment, and incident handling
Data Governance Document training data sources, representativeness, bias mitigation measures, and preprocessing steps
Technical Documentation Model design, intended use, performance characteristics, known limitations, and monitoring plan
Logging & Traceability Maintain logs of inputs, outputs, and key decisions for auditability and incident investigation
Transparency Users must be informed when subject to AI-driven decisions in high-risk contexts
Human Oversight Ability for a human to intervene, override, and review automated decisions
Accuracy, Robustness & Cybersecurity Stress-test and monitor for failures and attacks; ensure resilience across edge cases
Post-Market Monitoring Ongoing monitoring and mandatory reporting of serious incidents post-deployment

Sector-Specific Regulation

Domain Regulatory Drivers Typical Controls Required
Credit & Lending ECOA, FCRA, CFPB guidance (US); FCA (UK); EU consumer credit directive Explainability, adverse action notices, disparate impact testing, model validation
Insurance State regulators (US); EIOPA (EU); FCA (UK) Pricing transparency, discrimination constraints, audit trails, actuarial sign-off
Healthcare FDA / EMA / MHRA; HIPAA (US); MDR (EU) Clinical validation, safety monitoring, privacy safeguards, regulatory clearance for SaMD
Employment EEOC guidance (US); NYC Local Law 144; EU AI Act Title III Bias audits, candidate notification, documentation of model use
Data Privacy GDPR (EU); CCPA / CPRA (US); LGPD (Brazil); PDPA (various) Data minimisation, lawful basis for processing, right to explanation, data subject rights
Financial Markets SR 11-7 (US Fed); EBA guidelines (EU); MAS (Singapore) Model risk management, validation, governance, documentation of assumptions
Public Sector EU AI Act prohibited & high-risk lists; national AI strategies Mandatory human oversight; prohibition on certain social scoring applications

Model Risk Management (SR 11-7 Framework)

The US Federal Reserve's SR 11-7 guidance is the most widely adopted framework for model governance in financial services — and its principles are increasingly applied across industries.

SR 11-7 Component What It Requires
Model Development Sound methodology; well-reasoned assumptions; documented design choices
Model Validation Independent validation by a team separate from model developers
Conceptual Soundness Review of the theory and assumptions underlying the model
Ongoing Monitoring Track model performance and stability over time; flag material degradation
Outcomes Analysis Compare model predictions to actual outcomes; assess accuracy and discrimination
Benchmarking Compare model against alternative approaches or challenger models
Model Inventory Maintain a register of all models in use, their status, risk tier, and validation history

Responsible AI Toolkits

Toolkit Provider Capabilities
Fairlearn Microsoft (open-source) Fairness assessment and mitigation for classification and regression models
AI Fairness 360 (AIF360) IBM (open-source) 70+ fairness metrics; 10+ bias mitigation algorithms
What-If Tool Google (open-source) Interactive exploration of model behaviour and fairness across subgroups
SHAP Open-source Model-agnostic feature attribution; supports all major ML frameworks
LIME Open-source Local interpretable model-agnostic explanations; approximates model locally
InterpretML Microsoft (open-source) Explainable Boosting Machines (EBM); glass-box model family
Responsible AI Dashboard Microsoft (Azure ML) Integrated fairness, explainability, error analysis, and causal analysis
Fiddler AI SaaS Enterprise model monitoring, explainability, and fairness auditing
Arthur AI SaaS Bias detection, performance monitoring, explainability for regulated industries
Credo AI SaaS AI governance platform; policy-as-code; compliance evidence generation

Documentation Standards

Standard Description Who Uses It
Model Cards Standardised model documentation covering intended use, performance, limitations, and ethical considerations Google, Hugging Face, industry-wide
Datasheets for Datasets Documentation standard for training datasets covering motivation, composition, and known biases Academic and enterprise ML teams
FactSheets IBM's structured transparency document for AI services IBM customers and enterprise deployments
AI BoM (Bill of Materials) Inventory of all components, data sources, and dependencies in an AI system Enterprise AI governance; supply chain compliance
EU AI Act Technical File Mandatory documentation package for high-risk AI systems under the EU AI Act EU market deployers of high-risk AI

MLOps

Detailed reference content for mlops.

MLOps — Model Deployment & Lifecycle

MLOps (Machine Learning Operations) is the discipline of deploying, monitoring, and maintaining predictive models in production reliably and at scale.

Model Deployment Patterns

Pattern Description Best For
Real-Time Inference (REST API) Model served as an API endpoint; responds to requests in milliseconds Fraud detection, credit scoring, recommendation
Batch Prediction Model runs over a dataset at scheduled intervals; results stored Daily churn scores, overnight demand forecasts
Streaming Inference Model applied to events in real time as they arrive (Kafka, Kinesis) Transaction monitoring, log anomaly detection
Edge Deployment Model deployed on device; no server round-trip required Mobile apps, IoT sensors, autonomous vehicles
Embedded ML Model compiled into a software product or firmware Spam filters in email clients, ABS in vehicles
Shadow Mode New model runs in parallel with production; results compared without serving to users Safe A/B testing before full cutover
A/B Testing Route a portion of traffic to new model; compare business metrics Gradual model rollout with statistical confidence
Canary Deployment Gradually increase traffic to new model while monitoring for degradation Risk-managed production transitions

Model Serving Infrastructure

Tool Type Highlights
TensorFlow Serving Open-source Production-grade serving for TF models; gRPC + REST
TorchServe Open-source PyTorch's official model server
Triton Inference Server Open-source (NVIDIA) Multi-framework; GPU-optimised; high-throughput batching
BentoML Open-source / SaaS Framework-agnostic model serving; easy containerisation
Seldon Core Open-source Kubernetes-native model serving; explainability integration
KServe Open-source Kubernetes-based serverless inference; multi-model serving
Ray Serve Open-source Distributed inference; Python-native; composable pipelines
FastAPI Open-source Lightweight REST API framework; common for custom serving

Model Monitoring

Monitoring Type What It Detects Tools
Data Drift Input feature distributions shift from training data Evidently, NannyML, WhyLogs, Arize
Concept Drift Relationship between inputs and output changes over time Evidently, NannyML, Fiddler
Label / Prediction Drift Output distributions shift unexpectedly Arize Phoenix, Fiddler, Arthur AI
Data Quality Missing values, outliers, schema violations in production data Great Expectations, Monte Carlo, Soda
Model Performance Decay Accuracy, precision, or AUC drops below threshold MLflow, Arize, Fiddler, SageMaker Monitor
Infrastructure Monitoring Latency, throughput, error rates of serving infrastructure Prometheus, Grafana, Datadog

MLOps Platforms

Platform Type Highlights
MLflow Open-source Experiment tracking + model registry + deployment; industry standard
Kubeflow Open-source Kubernetes-native ML pipelines; production-grade orchestration
ZenML Open-source Clean MLOps abstraction; framework-agnostic pipelines
Metaflow (Netflix) Open-source Python-first ML workflow management; data science-friendly
Prefect / Airflow Open-source General workflow orchestration used for ML pipelines
Weights & Biases SaaS Experiment tracking + artifact management + model registry
Arize AI SaaS Model observability; drift detection; explainability
Fiddler AI SaaS Enterprise ML monitoring; explainability; bias detection
Evidently AI Open-source / SaaS Data and model monitoring reports; drift analysis
NannyML Open-source / SaaS Performance monitoring without ground truth labels
Arthur AI SaaS ML monitoring; bias detection; explainability for regulated industries

The Full MLOps Lifecycle

┌─────────────────────────────────────────────────────────────────────┐
│ MLOPS LIFECYCLE │
│ │
│ DATA EXPERIMENTATION DEPLOYMENT │
│ ───────────── ───────────────── ────────────── │
│ Ingest → Train → Evaluate → Package → Serve │
│ Validate → Track → Compare → via API / Batch │
│ Feature Store Register Model / Edge │
│ │
│ MONITORING GOVERNANCE RETRAINING │
│ ───────────── ───────────────── ────────────── │
│ Drift → Audit logs → Trigger → │
│ Performance → Compliance → Retrain → │
│ Alerts Lineage Validate → Redeploy │
└─────────────────────────────────────────────────────────────────────┘

AutoML

Detailed reference content for automl.

AutoML & No-Code ML

AutoML automates the most labour-intensive steps of the ML pipeline — algorithm selection, feature engineering, and hyperparameter tuning — making predictive AI accessible to non-specialist practitioners.

AutoML Capabilities

Capability What It Automates
Algorithm Selection Tests multiple algorithms and selects the best performer
Feature Engineering Automatically creates, selects, and transforms features
Hyperparameter Optimisation Searches for optimal model parameters
Ensemble Construction Combines the best models into an ensemble
Model Evaluation Runs cross-validation and computes performance metrics
Neural Architecture Search (NAS) Discovers optimal DNN architectures automatically
Data Preprocessing Handles missing values, encoding, and normalisation automatically

Leading AutoML Platforms

Platform Type Highlights
H2O AutoML Open-source / SaaS Best open-source AutoML; Driverless AI; widely used in enterprise
Auto-sklearn Open-source Bayesian optimisation over scikit-learn pipelines; academic standard
TPOT Open-source Genetic programming for pipeline optimisation; exports clean Python code
AutoGluon Open-source (Amazon) State-of-the-art AutoML; strong stacking; competitive with manual models
Google Vertex AutoML SaaS No-code tabular, image, text, and video ML on Google Cloud
AWS SageMaker Autopilot SaaS AutoML with full pipeline visibility; explainability built-in
Azure Automated ML SaaS Enterprise AutoML with responsible AI integration
DataRobot SaaS Market-leading enterprise AutoML; MLOps; time-to-value focus
Dataiku SaaS Collaborative data science and ML platform; low-code/pro-code
RapidMiner SaaS Visual ML pipeline builder; no-code + Python integration
Obviously AI SaaS One-click predictive models for business users; natural language interface
Pecan AI SaaS Automated predictive analytics for business metrics

Foundation Models for Time-Series (Zero-Shot Forecasting)

A rapidly emerging category: large pre-trained models that forecast without task-specific training.

Model Organisation Context Length Highlights
Chronos Amazon 512 time steps Tokenises time-series; transformer-based; strong zero-shot
TimesFM Google 512 time steps Decoder-only transformer; strong zero-shot forecasting
Moirai Salesforce 1–5,000 steps Any-variate; any-frequency; universal forecasting
Lag-Llama Community 1,024 steps Open-source; univariate probabilistic forecasting
MOMENT CMU 512 time steps Multi-task time-series foundation model

Deep Dives

Detailed reference content for deep dives.

Feature Engineering & Data Preparation

Feature engineering is the process of transforming raw data into the numeric representations that ML models learn from. It remains one of the most impactful steps in the predictive AI pipeline.

Structured / Tabular Feature Types

Feature Type Description Handling Technique
Numeric Continuous or discrete numbers Normalisation, scaling, log transform
Categorical (Low Cardinality) Variables with few categories (e.g., country) One-hot encoding, label encoding
Categorical (High Cardinality) Variables with many categories (e.g., product ID) Target encoding, embeddings, hashing
Ordinal Ordered categories (e.g., education level) Integer encoding preserving order
Datetime Timestamps and date values Extract day, month, year, hour, day-of-week, cyclical encoding
Text Free-form strings TF-IDF, word2vec, BERT embeddings
Geospatial Latitude, longitude, region Geohashing, distance features, clustering
Interaction Features Products or ratios of two features Manual engineering or automated discovery
Lag Features Prior time period values as features For time-series; critical for forecasting
Aggregation Features Grouped statistics (user's last 30-day average spend) Window functions; feature stores

Feature Engineering Techniques

Technique Description When to Use
Normalisation (Min-Max) Scale features to [0, 1] range Neural networks; distance-based models
Standardisation (Z-score) Transform to zero mean, unit variance Logistic Regression, SVM, KNN
Log Transformation Reduce skewness of heavily right-skewed distributions Revenue, transaction amounts, counts
Binning / Discretisation Convert continuous variable to ordinal buckets Age groups, income bands, risk tiers
Polynomial Features Create squared and interaction terms Capture non-linearity in linear models
Target Encoding Replace category with mean of target variable High-cardinality categoricals in GBT
PCA / Dimensionality Reduction Compress many correlated features into fewer components High-dimensional data; reduce noise
Embeddings Dense vector representations of categorical entities Users, products, documents in deep models
Window Statistics Rolling mean, max, std over time window Time-series; fraud; user behaviour

Data Quality & Preprocessing

Challenge Approach
Missing Values Impute with mean/median/mode; model-based imputation; flag missingness as a feature
Class Imbalance Oversample minority (SMOTE), undersample majority, adjust class weights, use AUC not accuracy
Outliers Detect via IQR/Z-score; clip, transform, or flag as a separate category
Data Leakage Ensure no future information is used as a feature; strict train/test temporal splits
Concept Drift Monitor feature distributions in production; trigger retraining when distributions shift
Label Noise Clean labels via majority vote, label smoothing, or confident learning
Duplicates Deduplicate before train/test split to avoid data contamination

Feature Stores

Feature stores are the infrastructure layer for managing, storing, and serving features consistently across training and inference.

Tool Type Highlights
Feast Open-source Most widely used open-source feature store; online + offline serving
Tecton SaaS Enterprise feature platform; real-time streaming features
Databricks Feature Store SaaS Native integration with MLflow and Delta Lake
Vertex AI Feature Store SaaS Google's managed feature store; scalable low-latency serving
AWS SageMaker Feature Store SaaS Online and offline store; integrated with SageMaker
Hopsworks Open-source / SaaS Full-stack feature store with model registry

Model Training Techniques

Supervised Learning

The foundational paradigm for all predictive AI.

Aspect Detail
Core Mechanism Learn from (input, label) pairs; minimise prediction error on known outcomes
Requirement Labelled training data — each example must have a known correct output
Training Process Iterate over data; compute loss between predictions and labels; update weights via gradient descent or tree-splitting
Used For All classification, regression, risk scoring, and ranking tasks

Semi-Supervised Learning

Aspect Detail
Core Mechanism Use a small amount of labelled data plus a large amount of unlabelled data
Why It Matters Labelling is expensive; unlabelled data is abundant
Techniques Self-training, label propagation, pseudo-labelling, consistency regularisation
Used For Medical imaging (few labelled scans), NLP text classification, fraud detection

Transfer Learning

Aspect Detail
Core Mechanism Fine-tune a large pre-trained model on a small domain-specific labelled dataset
Why It Works Pre-trained models encode rich general representations; domain data adapts the final layers
Key Benefit Achieves strong performance with orders of magnitude less labelled data
Examples BERT fine-tuned for contract classification; ImageNet-pre-trained CNN for medical imaging

Hyperparameter Optimisation

Method How It Works Best For
Grid Search Exhaustively try all combinations in a specified grid Small parameter spaces; sanity checks
Random Search Randomly sample parameter combinations More efficient than grid search; general use
Bayesian Optimisation (Optuna, Hyperopt) Build a surrogate model of performance; intelligently explore space Complex, expensive search spaces
Successive Halving / Hyperband Early-stop underperforming configurations; allocate budget to promising ones Large-scale ML at resource-constrained budgets
Neural Architecture Search (NAS) Automatically discover optimal model architectures Deep learning; automated model design

Ensemble Methods

Method How It Works When to Use
Bagging Train multiple models on random data subsets; aggregate predictions (e.g., Random Forest) Reduce variance; avoid overfitting
Boosting Train models sequentially; each corrects predecessor errors (e.g., XGBoost) Reduce bias; high accuracy on tabular data
Stacking Use predictions of base models as features for a meta-model Kaggle competitions; maximum accuracy
Voting Combine predictions from diverse models by majority vote or averaging Robust, easy to implement ensemble
Blending Similar to stacking but uses a holdout set rather than cross-validation Production ensembles

Class Imbalance Handling

Technique Description When to Use
SMOTE Synthetic Minority Over-sampling Technique — generates synthetic minority examples Moderate imbalance; tabular data
Class Weights Increase the training loss for minority class errors All gradient-based models; simple and effective
Undersampling Reduce majority class to balance dataset Very high imbalance; large majority class
Threshold Adjustment Lower classification threshold to increase minority recall When false negatives are more costly
Ensemble for Imbalance (BalancedBagging) Bagging with balanced sampling per tree Random Forest on imbalanced data

Overview

Detailed reference content for overview.

Definition & Core Concept

Predictive / Discriminative AI is the branch of artificial intelligence focused on systems that learn to map inputs to outputs — classifying what something is, scoring the probability of an outcome, ranking items by relevance, or forecasting future values — based entirely on patterns extracted from historical labelled data.

This is the most widely deployed form of AI in production today. It powers spam filters, credit scores, fraud detection, medical diagnosis, demand forecasting, recommender systems, and the vast majority of "AI in the enterprise" that predated the generative AI era.

Dimension Detail
Core Capability Discriminates — learns the boundary between classes or the mapping from inputs to outputs
How It Learns Supervised or semi-supervised learning on labelled datasets; minimises a loss function
What It Produces Labels, probabilities, scores, rankings, or numeric forecasts — not new content
Key Differentiator Predicts from existing patterns; it does not create, reason autonomously, or generate

Predictive AI vs. Other AI Types

AI Type What It Does Example
Predictive / Discriminative AI Classifies, scores, and forecasts from patterns in historical data Fraud score, churn probability, demand forecast
Agentic AI Pursues goals autonomously using tools, memory, and planning Research agent, coding agent
Analytical AI Extracts insights and explanations from existing data Dashboard, root-cause analysis
Autonomous AI (Non-Agentic) Operates independently within fixed boundaries without human input Autopilot, auto-scaling, algorithmic trading
Bayesian / Probabilistic AI Reasons under uncertainty using probability distributions Clinical trial analysis, A/B testing, risk modelling
Cognitive / Neuro-Symbolic AI Combines neural learning with symbolic reasoning LLM + knowledge graph, physics-informed neural net
Conversational AI Manages multi-turn dialogue between humans and machines Customer service chatbot, voice assistant
Evolutionary / Genetic AI Optimises solutions through population-based search inspired by natural selection Neural architecture search, logistics scheduling
Explainable AI (XAI) Makes AI decisions understandable to humans SHAP explanations, LIME, Grad-CAM
Generative AI Creates new original content from learned distributions Write an essay, generate an image
Multimodal Perception AI Fuses vision, language, audio, and other modalities GPT-4o processing image + text, AV sensor fusion
Optimisation / Operations Research AI Finds optimal solutions to constrained mathematical problems Vehicle routing, supply chain planning, scheduling
Physical / Embodied AI Acts in the physical world through sensors and actuators Autonomous vehicle, robot arm, drone
Privacy-Preserving AI Trains and runs AI without exposing raw data Federated hospital models, differential privacy
Reactive AI Responds to current input with no learning or memory Chess engine, rule-based spam filter
Recommendation / Retrieval AI Surfaces relevant items from large catalogues based on user signals Netflix suggestions, Google Search, Spotify playlists
Reinforcement Learning AI Learns optimal behaviour from reward signals via trial and error AlphaGo, robotic locomotion, RLHF
Scientific / Simulation AI Solves scientific problems and models physical systems AlphaFold, climate simulation, molecular dynamics
Symbolic / Rule-Based AI Reasons over explicit rules and knowledge to derive conclusions Medical expert system, legal reasoning engine

Key Distinction from Generative AI: Generative AI produces new content that did not exist. Predictive AI evaluates what already exists — assigning it to a category, estimating its probability, or forecasting what will happen next.

Key Distinction from Analytical AI: Analytical AI answers "what does this data mean?" and surfaces insights. Predictive AI answers "what will happen?" or "which category does this belong to?" — outputting a specific, actionable prediction.