The weights and architecture of Mixture-of-Experts model, Grok-1 (By xAI)
It is an open-source embedding database. Chroma makes it easy to build LLM apps by making knowledge, facts, and skills pluggable for LLMs. | It is an easy to use desktop app for experimenting with local and open-source Large Language Models (LLMs). It supports any ggml Llama, MPT, and StarCoder model on Hugging Face (Llama 2, Orca, Vicuna, Nous Hermes, WizardCoder, MPT, etc.) |
Fully-typed, fully-tested, fully-documented;
Dev, Test, Prod: the same API that runs in your Python notebook, scales to your cluster;
Feature-rich;
Free & open source | Run LLMs on your laptop, entirely offline;
Use models through the in-app Chat UI or an OpenAI compatible local server;
Download any compatible model files from HuggingFace repositories;
Discover new & noteworthy LLMs in the app's home page |
Statistics | |
GitHub Stars 24.2K | GitHub Stars - |
GitHub Forks 1.9K | GitHub Forks - |
Stacks 41 | Stacks 9 |
Followers 15 | Followers 8 |
Votes 0 | Votes 0 |
Integrations | |

Milvus is an open source vector database. Built with heterogeneous computing architecture for the best cost efficiency. Searches over billion-scale vectors take only milliseconds with minimum computing resources.

It allows you to run open-source large language models, such as Llama 2, locally.

It is a platform for building production-grade LLM applications. It lets you debug, test, evaluate, and monitor chains and intelligent agents built on any LLM framework and seamlessly integrates with LangChain, the go-to open source framework for building with LLMs.

It is a new scripting language to automate your interaction with a Large Language Model (LLM), namely OpenAI. The ultimate goal is to create a natural language programming experience. The syntax of GPTScript is largely natural language, making it very easy to learn and use.

It is a framework built around LLMs. It can be used for chatbots, generative question-answering, summarization, and much more. The core idea of the library is that we can “chain” together different components to create more advanced use cases around LLMs.

It is an open-source library designed to help developers build conversational streaming user interfaces in JavaScript and TypeScript. The SDK supports React/Next.js, Svelte/SvelteKit, and Vue/Nuxt as well as Node.js, Serverless, and the Edge Runtime.

Build, train, and deploy state of the art models powered by the reference open source in machine learning.

It is a project that provides a central interface to connect your LLMs with external data. It offers you a comprehensive toolset trading off cost and performance.

It is an open-source Vector Search Engine and Vector Database written in Rust. It deploys as an API service providing search for the nearest high-dimensional vectors. With Qdrant, embeddings or neural network encoders can be turned into full-fledged applications for matching, searching, recommending, and much more.

It is a Rust ecosystem of libraries for running inference on large language models, inspired by llama.cpp. On top of llm, there is a CLI application, llm-cli, which provides a convenient interface for running inference on supported models.