Observability, Testing & Evaluation

Monitor, debug, and evaluate your AI agents. Discover essential LLMOps and AgentOps tools like LangSmith for reliable deployment.

An evals and observability platform designed for building reliable AI agents.

Galileo AI is a company specializing in generative AI evaluation and monitoring solutions.

AgentOps provides AI agent development tools for monitoring, evaluation, and debugging to enhance performance and reliability.

An experimentation and human annotation platform designed for AI development teams.

Open-source platform for managing and enhancing LLM chatbots.

LangSmith is a platform designed to streamline the development and deployment of LLM applications from prototype to production.

HoneyHive is a unified LLMOps platform offering AI evaluation, testing, observability, and collaborative prompt management for teams building LLM applications.