Help developers discover the tools you use. Get visibility for your team's tech choices and contribute to the community's knowledge.
It is a tool for testing and evaluating LLM output quality. With this tool, you can systematically test prompts, models, and RAGs with predefined test cases. It can be utilized as a CLI, a library, or integrated into CI/CD pipelines. | It is a no-code compute platform for language models. It is aimed at AI developers and product builders. You can also vibe-check and compare quality, performance, and cost at once across a wide selection of open-source and proprietary LLMs. |
Evaluate quality and catch regressions;
Speed up evaluations with caching and concurrency;
Score outputs automatically by defining test cases | Query and compare a large selection of open-source and proprietary models at once;
Replace costly APIs with cheap custom AI models;
Airtrain’s LLM-assisted scoring simplifies model grading using your task descriptions;
Cut your AI costs by up to 90% |
Statistics | |
GitHub Stars 9.0K | GitHub Stars - |
GitHub Forks 760 | GitHub Forks - |
Stacks 0 | Stacks 0 |
Followers 0 | Followers 2 |
Votes 0 | Votes 0 |
Integrations | |

It is a platform for building production-grade LLM applications. It lets you debug, test, evaluate, and monitor chains and intelligent agents built on any LLM framework and seamlessly integrates with LangChain, the go-to open source framework for building with LLMs.

The collaborative testing platform for LLM applications and agents. Your whole team defines quality requirements together, Rhesis generates thousands of test scenarios covering edge cases, simulates realistic multi-turn conversations, and delivers actionable reviews. Testing infrastructure built for Gen AI.

The most advanced, consistent, and effective AI humanizer on the market. Instantly transform AI-generated text into undetectable, human-like writing in one click.

Vivgrid is an AI agent infrastructure platform that helps developers and startups build, observe, evaluate, and deploy AI agents with safety guardrails and global low-latency inference. Support for GPT-5, Gemini 2.5 Pro, and DeepSeek-V3. Start free with $200 monthly credits. Ship production-ready AI agents confidently.

Practice your interview skills with AI-powered interviewers. Simulate real interview scenarios and improve your performance. Get instant feedback. Get complete overview and a plan with next steps to improve.
Easily host and share test reports. Gaffer saves developers time and improves test visibility.

Is a training API for researchers and developers.

Is a debate simulator powered by the top 5 LLM's. Generate endless discussions and debates on any topic. It's like reddit - but powered by AI.

Provides comprehensive AI validation and certification services. Get instant AI trust scores, secure badges, and compliance reports. Validate your AI systems for transparency, data protection, governance, and user control. Trusted by startups and enterprises worldwide.

CI failures are painful to debug. SentinelQA gives you run summaries, flaky test detection, regression analysis, visual diffs and AI-generated action items.