The weights and architecture of Mixture-of-Experts model, Grok-1 (By xAI)
It is a full-stack application and tool suite that enables you to turn any document, resource, or piece of content into a piece of data that any LLM can use as reference during chatting. This application runs with very minimal overhead as by default the LLM and vectorDB are hosted remotely, but can be swapped for local instances. | Boost your customer service with Coldi's AI call center. Try our AI voice agents to handle calls, bookings, and support – fast, efficient, and always on. |
Atomically manage documents to be used in long-term-memory from a simple UI;
Each chat response contains a citation that is linked to the original content;
Simple technology stack for fast iteration;
Fully capable of being hosted remotely | Products, Pricing, Products, Pricing, info@coldi.ai |
Statistics | |
GitHub Stars 50.7K | GitHub Stars - |
GitHub Forks 5.3K | GitHub Forks - |
Stacks 9 | Stacks 0 |
Followers 14 | Followers 1 |
Votes 0 | Votes 1 |
Integrations | |
| No integrations available | |

It allows you to run open-source large language models, such as Llama 2, locally.

It is a platform for building production-grade LLM applications. It lets you debug, test, evaluate, and monitor chains and intelligent agents built on any LLM framework and seamlessly integrates with LangChain, the go-to open source framework for building with LLMs.

It is a new scripting language to automate your interaction with a Large Language Model (LLM), namely OpenAI. The ultimate goal is to create a natural language programming experience. The syntax of GPTScript is largely natural language, making it very easy to learn and use.

It is a framework built around LLMs. It can be used for chatbots, generative question-answering, summarization, and much more. The core idea of the library is that we can “chain” together different components to create more advanced use cases around LLMs.

It is an open-source library designed to help developers build conversational streaming user interfaces in JavaScript and TypeScript. The SDK supports React/Next.js, Svelte/SvelteKit, and Vue/Nuxt as well as Node.js, Serverless, and the Edge Runtime.

Build, train, and deploy state of the art models powered by the reference open source in machine learning.

It is a project that provides a central interface to connect your LLMs with external data. It offers you a comprehensive toolset trading off cost and performance.

It is a Rust ecosystem of libraries for running inference on large language models, inspired by llama.cpp. On top of llm, there is a CLI application, llm-cli, which provides a convenient interface for running inference on supported models.

It is an open-source embedding database. Chroma makes it easy to build LLM apps by making knowledge, facts, and skills pluggable for LLMs.

It is an open-source, drag & drop UI to build your customized LLM flow. It is built on top of LangChainJS, with the aim to make it easy for people to visualize and build LLM apps.