Help developers discover the tools you use. Get visibility for your team's tech choices and contribute to the community's knowledge.
CI failures are painful to debug. SentinelQA gives you run summaries, flaky test detection, regression analysis, visual diffs and AI-generated action items. | Track AI API spend, tokens, latency, and errors across major providers with one observability dashboard. |
QA, DevOps, Test Intelligence, AI, Analytics, Test Debugging | Features, Providers, Changelog, Terms of Privacy, Privacy Policy |
Statistics | |
Stacks 0 | Stacks 0 |
Followers 1 | Followers 1 |
Votes 1 | Votes 1 |

BrowserStack is the leading test platform built for developers & QAs to expand test coverage, scale & optimize testing with cross-browser, real device cloud, accessibility, visual testing, test management, and test observability.

Small, fast and scaleable bearbones state-management solution. Has a comfy api based on hooks, that isn't boilerplatey or opinionated, but still just enough to be explicit and flux-like.

TestRail helps you manage and track your software testing efforts and organize your QA department. Its intuitive web-based user interface makes it easy to create test cases, manage test runs and coordinate your entire testing process.

Statsbot is helping you take control of your raw data, providing an all-in-one analysis tool for engineers and non-tech folks alike.

Is the game-changing European modern data quality platform that effortlessly uncovers anomalies and errors in your data with Artificial Intelligence.

Manage all aspects of software quality; integrate with JIRA and various test tools, foster collaboration and gain real-time visibility.

Transform basic prompts into expert-level AI instructions. Enhance, benchmark & optimize prompts for ChatGPT, Claude, Gemini & more.

Find what caused your AI bill. Opsmeter gives endpoint, user, model, and prompt-level AI cost attribution in one view.

It is a platform for building production-grade LLM applications. It lets you debug, test, evaluate, and monitor chains and intelligent agents built on any LLM framework and seamlessly integrates with LangChain, the go-to open source framework for building with LLMs.

The collaborative testing platform for LLM applications and agents. Your whole team defines quality requirements together, Rhesis generates thousands of test scenarios covering edge cases, simulates realistic multi-turn conversations, and delivers actionable reviews. Testing infrastructure built for Gen AI.