Help developers discover the tools you use. Get visibility for your team's tech choices and contribute to the community's knowledge.
It is a tool for testing and evaluating LLM output quality. With this tool, you can systematically test prompts, models, and RAGs with predefined test cases. It can be utilized as a CLI, a library, or integrated into CI/CD pipelines. | It is the leading observability platform trusted by high-performing teams to help maintain the quality and performance of ML models, LLMs, and data pipelines. |
Evaluate quality and catch regressions;
Speed up evaluations with caching and concurrency;
Score outputs automatically by defining test cases | Detect, prevent, and mitigate risk in your AI applications;
Seamless integration with your existing pipelines and tools;
Model and data health;
LLM security |
Statistics | |
GitHub Stars 9.0K | GitHub Stars - |
GitHub Forks 760 | GitHub Forks - |
Stacks 0 | Stacks 1 |
Followers 0 | Followers 4 |
Votes 0 | Votes 0 |
Integrations | |

Waxell is the AI governance plane for agentic systems in production. It sits above agents, models, and integrations, enforcing constraints and defining what's allowed. Auto-instrumentation for 200+ libraries without code changes. Real-time tracing, token and cost tracking, and 11 categories of agentic governance policy enforcement.

Find what caused your AI bill. Opsmeter gives endpoint, user, model, and prompt-level AI cost attribution in one view.

It is a platform for building production-grade LLM applications. It lets you debug, test, evaluate, and monitor chains and intelligent agents built on any LLM framework and seamlessly integrates with LangChain, the go-to open source framework for building with LLMs.

Transform basic prompts into expert-level AI instructions. Enhance, benchmark & optimize prompts for ChatGPT, Claude, Gemini & more.

The collaborative testing platform for LLM applications and agents. Your whole team defines quality requirements together, Rhesis generates thousands of test scenarios covering edge cases, simulates realistic multi-turn conversations, and delivers actionable reviews. Testing infrastructure built for Gen AI.

Vivgrid is an AI agent infrastructure platform that helps developers and startups build, observe, evaluate, and deploy AI agents with safety guardrails and global low-latency inference. Support for GPT-5, Gemini 2.5 Pro, and DeepSeek-V3. Start free with $200 monthly credits. Ship production-ready AI agents confidently.

The most advanced, consistent, and effective AI humanizer on the market. Instantly transform AI-generated text into undetectable, human-like writing in one click.

Practice your interview skills with AI-powered interviewers. Simulate real interview scenarios and improve your performance. Get instant feedback. Get complete overview and a plan with next steps to improve.

Dominate AI search results with Hikoo. The complete Generative Engine Optimization platform to monitor, analyze, and improve your visibility on ChatGPT, Claude, Perplexity, and all major AI assistants. Start your AI SEO journey today.

Is a debate simulator powered by the top 5 LLM's. Generate endless discussions and debates on any topic. It's like reddit - but powered by AI.