AI Tooling Landscape — 2026 Overview¶
This is a high-level map of the entire AI tooling ecosystem as documented in this repository. It serves as the main entry point for understanding how everything connects.
Purpose¶
A living overview that maps the AI tooling landscape into layers, showing how tools relate to each other.
The Stack (layered view)¶
┌───────────────────────────────────────────────────────────────────────────┐
│ Layer 7: Applications (ChatGPT, Perplexity, Open WebUI) │
├───────────────────────────────────────────────────────────────────────────┤
│ Layer 6: Agents & Orchestration (CrewAI, AutoGen, LangGraph, n8n) │
├───────────────────────────────────────────────────────────────────────────┤
│ Layer 5: Frameworks (LangChain, LlamaIndex, Haystack, DSPy) │
├───────────────────────────────────────────────────────────────────────────┤
│ Layer 4: Protocols & Standards (MCP, Tool Calling, A2A) │
├───────────────────────────────────────────────────────────────────────────┤
│ Layer 3: Inference & Serving (vLLM, TGI, Ollama, SGLang) │
├───────────────────────────────────────────────────────────────────────────┤
│ Layer 2: Models (GPT-4, Claude, Llama, Mistral, Gemini, Qwen) │
├───────────────────────────────────────────────────────────────────────────┤
│ Layer 1: Providers (OpenAI, Anthropic, Google, Meta, Mistral, OpenRouter) │
├───────────────────────────────────────────────────────────────────────────┤
│ Layer 0: Infrastructure (GPUs, quantization, vector DBs) │
└───────────────────────────────────────────────────────────────────────────┘
Layer 7: Applications¶
User-facing interfaces and platforms where humans interact with AI. These provide the final product experience, abstracting the underlying layers for end-users. This layer includes both general-purpose chat interfaces and specialized AI-enhanced development environments. - Relevant Pages: ChatGPT, Perplexity, Open WebUI, Claude Code, Cursor, Aider, Zed, Obsidian, Logseq, TeamOut, Valyu, ansigpt, Continue Dev, Codeium, GitHub Copilot, VS Code, Tabnine, Mentat, GPT Engineer, Melty, Superconductor, Terminus 2, Junie CLI. - Key Trends: Moving from simple chat to agentic IDEs and multimodal research assistants.
Layer 6: Agents & Orchestration¶
Systems that coordinate multiple steps, tools, and agents to achieve complex goals. This layer handles reasoning, planning, and task execution using underlying models and frameworks. - Relevant Pages: Mistral Agents, CrewAI, AutoGen, LangGraph, n8n, Agency Swarm, Agentic Automation Canvas, Agno, Bee Agent Framework, Composio, Phidata, OpenHands, Droid, Plandex, OpenSwarm, OpenClaw, Jules, Browser Use, Zapier, Make, Skyvern, Atlassian Jira MCP, ServiceNow MCP, CliHub. Systems that coordinate multiple steps, tools, and agents to achieve complex goals. This layer handles reasoning, planning, and task execution using underlying models and frameworks. It is where autonomous decision-making and environment interaction are managed. - Key Trends: Shift from linear chains to complex, stateful multi-agent graphs.
Layer 5: Frameworks¶
Development libraries used to build AI applications, handling prompt management, tool integration, and RAG logic. They provide the abstraction layer between models and applications. These frameworks simplify the process of constructing complex AI workflows and integrating various data sources. - Relevant Pages: LangChain, LlamaIndex, Haystack, DSPy, Semantic Kernel, Smolagents, Mycelium, Dify, Flowise, RAGFlow. - Key Trends: Increased focus on programmatic prompt optimization and modular RAG.
Layer 4: Protocols & Standards¶
The "glue" that allows models to interact with tools and other agents consistently. These standards ensure interoperability across the ecosystem. By establishing common interfaces, they prevent vendor lock-in and enable tool reuse across different frameworks. - Relevant Pages: Model Context Protocol (MCP), Agent Client Protocol (ACP), Tool Calling & MCP Patterns, Mistral AI (Native MCP), MCP Registry. - Key Trends: Rapid adoption of MCP as the standard for model-to-tool communication.
Layer 3: Inference & Serving¶
Engines that run model weights and provide APIs for applications to consume. This layer is responsible for the actual execution of model inference. It optimizes performance, handles concurrency, and provides the necessary scaling for production deployments. - Relevant Pages: vLLM, Text Generation Inference (TGI), Ollama, SGLang, Aphrodite Engine, ExLlamaV2, llama.cpp, MLX, LiteLLM. - Key Trends: Layer 3 is consolidating around vLLM and SGLang for high-performance serving.
Layer 2: Models¶
The core reasoning engines (LLMs, VLMs) that process information and generate text or actions. These are the fundamental units of intelligence in the stack. This layer includes both general-purpose foundation models and specialized models for coding, reasoning, or multimodality. - Relevant Pages: OpenAI Models, Anthropic Claude, Meta Llama, Mistral, Google Gemini, DeepSeek, Model Classes. - Key Trends: Rise of specialized reasoning models using test-time compute.
Layer 1: Providers¶
Companies and platforms that host models and provide them as-a-service via API. They handle the scale and infrastructure required for model access. These providers offer varying levels of cost, speed, and privacy, allowing users to choose the best fit for their needs. - Relevant Pages: OpenRouter, Groq, Fireworks AI, Together AI, Replicate, Mistral AI, Cohere. - Key Trends: Providers are competing on speed (tokens/sec) and lower costs.
Layer 0: Infrastructure¶
The underlying hardware, storage, and low-level optimizations like quantization and vector databases that power the entire stack. This foundation ensures that higher-level services run efficiently and securely. It also includes the critical data supply chain components for ingestion and preparation. - Relevant Pages: Home Lab Architecture, TrueNAS SCALE, Tailscale, OpenPipe (Fine-tuning), Crawl4AI, Firecrawl, OCRmyPDF, PageIndex, CalDAV, ZSE. - Key Trends: Move towards hybrid infrastructure combining local GPU power with cloud scaling.
Key Patterns¶
- Retrieval-Augmented Generation (RAG): Grounding models with external data to improve accuracy.
- Tool Calling & MCP: Standardized interaction between models and external functions.
- LLM Trust Boundaries: Security and privacy considerations in agentic systems.
- Agent Skills Best Practices: Optimizing how agents use tools.
- Claude Tool Search: Specific patterns for maximizing Anthropic's tool use.
- OpenClaw Workflow Prompts: Library of prompts for specialized workflows.
How to use this repo¶
- "I want to run LLMs locally" → Ollama, MLX, llama.cpp, ExLlamaV2
- "I want to build an AI agent" → CrewAI/AutoGen + LangGraph + MCP
- "I want to add AI to my app" → LangChain/LlamaIndex + OpenRouter (provider API)
- "I want to evaluate models" → Benchmarking tools
- "I want to stay current" → Essential AI Reading List
- "I want the highest-signal repos from my GitHub stars" → Starred AI / Agent Repositories Over 10K Stars
- "I want the shortest practical stack for an AI-driven company" → AI Company Starter Stack
- "I want to build a website or small app on free infrastructure" → AI Builder Index and Free AI Website Playbook
Sources / references¶
- Sequoia: Generative AI's Act Two
- A16Z: Emerging Architectures for LLM Applications
- MAD Landscape 2024
Contribution Metadata¶
- Last reviewed: 2026-03-15
- Confidence: high