Core Concepts
Reasoning
Memory & Retrieval
Agent Types
Design Patterns
Training & Alignment
Frameworks
Tools
Safety & Security
Evaluation
Meta
Core Concepts
Reasoning
Memory & Retrieval
Agent Types
Design Patterns
Training & Alignment
Frameworks
Tools
Safety & Security
Evaluation
Meta
Artificial intelligence encompasses a wide spectrum of systems with vastly different capabilities, approaches, and architectures. Classifying AI into distinct types helps clarify what current systems can actually do versus what remains theoretical, guides investment and regulatory decisions, and sets appropriate expectations for AI deployment.
There are four primary frameworks for classifying AI: by capability level, by functional behavior, by machine learning approach, and by architectural design. Each framework provides a different lens for understanding the AI landscape.1)
The most widely recognized classification divides AI into three levels based on the scope of intelligence:
Narrow AI — also called Weak AI — is designed to perform specific tasks within a bounded domain. It cannot generalize beyond its training or transfer knowledge to unrelated tasks. Despite this limitation, narrow AI powers virtually all commercial AI applications in use today.2)
Examples include:
Importantly, modern narrow AI is significantly more capable than earlier generations. Multimodal models like GPT-5 and Claude Opus 4.6 can reason across text, code, images, audio, and tool use — leading some researchers to use terms like “proto-AGI” or “agentic AI” to describe these systems. However, they remain bounded by their training and cannot autonomously learn entirely new domains without retraining.3)
Current status (2026): Dominant paradigm. All production AI systems are forms of narrow AI.
General AI would match human-level cognitive abilities across diverse domains — learning any intellectual task that a human can, transferring knowledge between unrelated fields, and reasoning flexibly about novel situations without specific training.
AGI does not yet exist. However, fragments of general-like capability are emerging within narrow systems:
Leading AI researchers have varying timelines for AGI. Geoffrey Hinton, Nobel Laureate for his work in AI, estimates that most field experts believe AGI will emerge within 20 years. OpenAI's GPT-5.2 achieved 52.9% on the ARC-AGI-2 benchmark (up from 17.6% for GPT-5.1), suggesting measurable progress on general reasoning tasks.
Current status (2026): Theoretical. No true AGI exists, but advanced models approach narrow benchmarks for general reasoning.
Super AI would surpass human intelligence across all dimensions — scientific creativity, social intelligence, general wisdom, and problem-solving. ASI is a hypothetical concept that raises profound questions about control, alignment, and existential risk.
No credible development milestones toward ASI have been achieved. It remains the subject of philosophical and safety research rather than engineering efforts.
Current status (2026): Purely speculative. No development pathway established.
This framework categorizes AI by how it processes information and interacts with its environment:4)
Reactive machines respond to current inputs without any memory of past interactions. They cannot learn from experience or adjust behavior over time. They follow fixed rules for each input.
Examples: IBM's Deep Blue (chess), basic game AI, simple rule-based chatbots
Current status: Mature but largely superseded by more capable memory-enabled systems. Still used in simple, deterministic applications.
Limited memory systems use historical data and past experiences to inform current decisions. They can learn from training data and, in some cases, from recent interactions within a session. This is the category that encompasses virtually all modern AI.
Examples: Autonomous vehicles (learning from driving data), LLMs (trained on text corpora), predictive analytics systems, medical imaging AI
Current status: Dominant paradigm. Powers self-driving cars, all LLMs, recommendation engines, and most commercial AI applications.
Theory of Mind AI would understand human emotions, beliefs, intentions, and social dynamics. It would be able to model the mental states of other agents and adjust its behavior accordingly — similar to how humans intuitively understand that others have thoughts and feelings different from their own.
Examples: No operational implementations exist. Research prototypes explore empathetic response generation and social reasoning in conversational AI.
Current status: Emerging research area. No practical implementations, though advanced NLP models show rudimentary social awareness.
Self-aware AI would possess consciousness and self-perception — understanding its own internal states, having subjective experiences, and being aware of its own existence.
Examples: Entirely hypothetical.
Current status: Theoretical. No evidence of development or even a clear scientific pathway. The nature of machine consciousness remains an open philosophical question.
Machine learning — the primary method by which modern AI systems learn — itself encompasses several distinct approaches:5)
The model trains on labeled data — input-output pairs where the correct answer is provided. The model learns to map inputs to outputs and can then predict outputs for new, unseen inputs.
| Aspect | Details |
|---|---|
| How it works | Learn from labeled examples (input-output pairs) |
| Common algorithms | Linear regression, logistic regression, decision trees, neural networks |
| Applications | Email spam detection, medical diagnosis, credit scoring, image classification, speech recognition |
| Strengths | High accuracy when quality labeled data is available |
| Limitations | Requires large amounts of labeled data, which can be expensive to produce |
The model finds patterns in unlabeled data without predefined correct answers. It discovers hidden structures, groupings, and relationships within the data.
| Aspect | Details |
|---|---|
| How it works | Discover patterns in data without labels |
| Common algorithms | K-means clustering, hierarchical clustering, autoencoders, principal component analysis |
| Applications | Customer segmentation, anomaly detection, topic discovery, data compression |
| Strengths | Can discover unexpected patterns; does not require labeled data |
| Limitations | Results can be harder to interpret and validate |
The model learns through trial and error, receiving rewards for desirable actions and penalties for undesirable ones. It develops strategies to maximize cumulative reward over time.
| Aspect | Details |
|---|---|
| How it works | Agent interacts with environment, receives rewards/penalties |
| Common algorithms | Q-learning, policy gradients, proximal policy optimization (PPO) |
| Applications | Game playing (AlphaGo, AlphaZero), robotics control, autonomous driving, RLHF for LLMs |
| Strengths | Can learn complex strategies in dynamic environments |
| Limitations | Computationally expensive; requires careful reward design |
Hybrid approaches that combine elements of the above:
Self-supervised learning is the foundation of modern LLMs and has been one of the most important developments in AI, enabling models to learn from vast unlabeled datasets.
Expert systems are rule-based programs that encode human expert knowledge as if-then rules. Dominant in the 1980s, they excel in narrow, well-defined domains but are brittle, expensive to maintain, and cannot learn from new data.
Examples: Medical diagnostic systems (MYCIN), financial advisory tools, industrial process control
Current status: Established but declining. Increasingly augmented or replaced by machine learning approaches that can learn from data and generalize better.6)
Neural networks are computing systems inspired by biological neural networks in the brain. They consist of layers of interconnected nodes (neurons) that process information and learn pattern recognition through adjusting connection weights during training.
Examples: Image recognition, speech processing, basic pattern recognition
Current status: Foundational technology underlying deep learning and all modern AI.
Deep learning uses neural networks with many layers (hence “deep”) to learn hierarchical representations of data. Deeper layers capture increasingly abstract features — for example, early layers might detect edges in images while later layers detect objects and scenes.
Examples: Convolutional Neural Networks (CNNs) for image classification, Recurrent Neural Networks (RNNs) for sequential data, generative models
Current status: State-of-the-art across computer vision, NLP, and content generation.
Transformers use self-attention mechanisms to process entire sequences in parallel, weighting the importance of different parts of the input. Introduced in 2017, they revolutionized NLP and have been extended to vision, audio, and multimodal applications.
Examples: GPT series, Claude, Gemini, Llama, BERT, T5
Current status: Leading-edge architecture powering all frontier AI models. Hybrid models incorporating Mixture-of-Experts and other innovations continue to extend the architecture.7)
| Classification | Types | Current Dominant |
|---|---|---|
| By Capability | Narrow, General, Super | Narrow AI |
| By Functionality | Reactive, Limited Memory, Theory of Mind, Self-Aware | Limited Memory |
| By ML Approach | Supervised, Unsupervised, Reinforcement, Self-Supervised | Self-Supervised (for LLMs) |
| By Architecture | Expert Systems, Neural Networks, Deep Learning, Transformers | Transformers |