Handpicked and reviewed AI applications to supercharge your workflow — updated daily.
Page 9 of 21 · 1,004 tools
PromptLayer is a platform for tracking, managing, and evaluating LLM prompts in production. Log every prompt and completion,…
Langfuse is an open-source LLM engineering platform for observability, testing, and prompt management. Debug production AI issues, evaluate…
Helicone provides one-line LLM observability — add a single line to your OpenAI calls and get full logging,…
Opik by Comet is an open-source LLM evaluation framework for testing AI application quality at scale. Automated evaluation…
Braintrust is an enterprise AI evaluation platform for measuring, improving, and shipping AI applications. Logging, evaluation datasets, prompt…
Phoenix by Arize is an open-source AI observability library for ML engineers. Traces LLM and embedding applications, visualizes…
TruLens is an open-source framework for evaluating and tracking LLM applications. Feedback functions assess truthfulness, harmlessness, and helpfulness…
DeepEval is an open-source LLM evaluation framework with 14+ evaluation metrics including hallucination, answer relevancy, and faithfulness. pytest-style…
Portkey is an AI gateway providing unified access to 200+ LLMs with built-in observability, caching, and fallbacks. Production-grade…
Unify automatically routes LLM requests to the cheapest or fastest provider based on your optimization criteria. Benchmark any…
🔍
We review every submission within 24–48 hours. Free listing, no strings attached.