Handpicked and reviewed AI applications to supercharge your workflow — updated daily.
Page 13 of 48 · 1,004 tools
DeepEval is an open-source LLM evaluation framework with 14+ evaluation metrics including hallucination, answer relevancy, and faithfulness. pytest-style…
Portkey is an AI gateway providing unified access to 200+ LLMs with built-in observability, caching, and fallbacks. Production-grade…
Unify automatically routes LLM requests to the cheapest or fastest provider based on your optimization criteria. Benchmark any…
Eden AI provides a unified API for 100+ AI models across text, image, audio, and video. Test and…
Agenta is an open-source LLMOps platform for prompt management, evaluation, and deployment. Teams collaborate on prompts, run systematic…
HoneyHive is an AI evaluation and observability platform for teams building LLM applications. Dataset management, automated evaluations, and…
Mirascope is a Python toolkit for building LLM applications with clean abstractions for prompts, calls, and extractions. Type-safe…
PromptLayer is a platform for tracking, managing, and evaluating LLM prompts in production. Log every prompt and completion,…
Langfuse is an open-source LLM engineering platform for observability, testing, and prompt management. Debug production AI issues, evaluate…
Helicone provides one-line LLM observability — add a single line to your OpenAI calls and get full logging,…
🔍
We review every submission within 24–48 hours. Free listing, no strings attached.