Evaluation and Tracking for LLM Experiments and AI Agents
-
Updated
Aug 19, 2025 - Python
Evaluation and Tracking for LLM Experiments and AI Agents
Open source platform for AI Engineering: OpenTelemetry-native LLM Observability, GPU Monitoring, Guardrails, Evaluations, Prompt Management, Vault, Playground. 🚀💻 Integrates with 50+ LLM Providers, VectorDBs, Agent Frameworks and GPUs.
Fiddler Auditor is a tool to evaluate language models.
A comprehensive solution for monitoring your AI models in production
A python library to send data to Arize AI!
Monitor, detect, and analyze security threats and Shadow AI in your LLM applications with comprehensive analytics and real-time alerting.
A report generator library for the ML models deployed on the Fiddler AI Observability platform
This repo hosts a chatbot that runs in a docker container to demo Okahu AI Observability Cloud
This repo hosts a chatbot that runs in Github Codespaces to demo Okahu AI Observability Cloud with OpenAI
Official Python Library to monitor your LLM Application with Doku
Open Source Video Understanding API and Large Vision Model Observability Platform.
The Modelmetry Python SDK allows developers to easily integrate Modelmetry’s advanced guardrails and monitoring capabilities into their LLM-powered applications.
FlagWise 🐙 detects Shadow AI and monitors LLM traffic in real time, exposing unauthorized or risky model use for security, compliance, and audit across enterprise systems.
Add a description, image, and links to the ai-observability topic page so that developers can more easily learn about it.
To associate your repository with the ai-observability topic, visit your repo's landing page and select "manage topics."