The largest side of the ecosystem
All tracked agentic AI projects written primarily in Python. LangChain, CrewAI, AutoGen, LlamaIndex, and dozens more.
A high-throughput and memory-efficient inference and serving engine for LLMs
Data infrastructure for AI
Open-source observability for your GenAI or LLM application, based on OpenTelemetry
A programming framework for agentic AI
AI Agent Framework, the Pydantic way
Supercharge Your LLM Application Evaluations 🚀
DSPy: The framework for programming—not prompting—language models
Convert documents to structured data effortlessly. Unstructured is open-source ETL solution for transforming complex documents into clean, structured formats for language models. Visit our website to learn more about our enterprise grade Platform product for production grade workflows, partitioning, enrichments, chunking and embedding.
Langflow is a powerful tool for building and deploying AI-powered agents and workflows.
Production-ready platform for agentic workflow development.
The agent engineering platform
🙌 OpenHands: AI-Driven Development
Universal memory layer for AI Agents
LlamaIndex is the leading document agent and OCR platform
Framework for orchestrating role-playing, autonomous AI agents. By fostering collaborative intelligence, CrewAI empowers agents to work together seamlessly, tackling complex tasks.
aider is AI pair programming in your terminal
Python SDK, Proxy Server (AI Gateway) to call 100+ LLM APIs in OpenAI (or native) format, with cost tracking, guardrails, loadbalancing and logging. [Bedrock, Azure, OpenAI, VertexAI, Cohere, Anthropic, Sagemaker, HuggingFace, VLLM, NVIDIA NIM]
Build resilient language agents as graphs.
The official Python SDK for Model Context Protocol servers and clients
Letta is the platform for building stateful agents: AI with advanced memory that can learn and self-improve over time.
A guidance language for controlling large language models.
SWE-agent takes a GitHub issue and tries to automatically fix it, using your LM of choice. It can also be employed for offensive cybersecurity or competitive coding challenges. [NeurIPS 2024]
Debug, evaluate, and monitor your LLM applications, RAG systems, and agentic workflows with comprehensive tracing, automated evaluations, and production-ready dashboards.
Evals is a framework for evaluating LLMs and LLM systems, and an open-source registry of benchmarks.
The LLM Evaluation Framework
Structured Outputs
Access large language models from the command-line
AI Observability & Evaluation