Help developers discover the tools you use. Get visibility for your team's tech choices and contribute to the community's knowledge.
Infratailors empowers you to optimize and observe the cost, energy, performance of your GPUs | JarvisLabs is a GPU cloud platform used by AI teams, research labs, and universities to train, fine-tune, and deploy deep learning models. Access NVIDIA H100, H200, A100, and L4 GPUs on-demand with per-minute billing, persistent storage, and pre-configured ML environments. Trusted by 1000+ teams across startups, enterprises, and academic institutions. Multi-region availability with dedicated datacenter infrastructure and 99.5% uptime SLA. |
Infrastructure, GPU Optimization, LLM Infrastructure, Monitoring, Cost Optimization, AI inference optimization | GPU Cloud, AI Infrastructure, Machine Learning, Deep Learning, Model Training, LLM Fine-Tuning, Cloud GPU, MLOps, AI/ML Platform, NVIDIA H100 |
Statistics | |
Stacks 0 | Stacks 0 |
Followers 1 | Followers 1 |
Votes 1 | Votes 1 |

The Docker Platform is the industry-leading container platform for continuous, high-velocity innovation, enabling organizations to seamlessly build and share any application — from legacy to what comes next — and securely run them anywhere

LXD isn't a rewrite of LXC, in fact it's building on top of LXC to provide a new, better user experience. Under the hood, LXD uses LXC through liblxc and its Go binding to create and manage the containers. It's basically an alternative to LXC's tools and distribution template system with the added features that come from being controllable over the network.

LXC is a userspace interface for the Linux kernel containment features. Through a powerful API and simple tools, it lets Linux users easily create and manage system or application containers.

Rocket is a cli for running App Containers. The goal of rocket is to be composable, secure, and fast.

Vagrant Cloud pairs with Vagrant to enable access, insight and collaboration across teams, as well as to bring exposure to community contributions and development environments.

The most advanced, consistent, and effective AI humanizer on the market. Instantly transform AI-generated text into undetectable, human-like writing in one click.

Waxell is the AI governance plane for agentic systems in production. It sits above agents, models, and integrations, enforcing constraints and defining what's allowed. Auto-instrumentation for 200+ libraries without code changes. Real-time tracing, token and cost tracking, and 11 categories of agentic governance policy enforcement.

It is a platform for building production-grade LLM applications. It lets you debug, test, evaluate, and monitor chains and intelligent agents built on any LLM framework and seamlessly integrates with LangChain, the go-to open source framework for building with LLMs.

Find what caused your AI bill. Opsmeter gives endpoint, user, model, and prompt-level AI cost attribution in one view.

Transform basic prompts into expert-level AI instructions. Enhance, benchmark & optimize prompts for ChatGPT, Claude, Gemini & more.