Help developers discover the tools you use. Get visibility for your team's tech choices and contribute to the community's knowledge.
It is a general-purpose speech recognition model. It is trained on a large dataset of diverse audio and is also a multi-task model that can perform multilingual speech recognition as well as speech translation and language identification. | It aims to make large models accessible to everyone by co-development of open models, datasets, systems, and evaluation tools. |
Automatic speech recognition;
Trained on a large dataset of diverse audio;
Multi-task model;
Can perform multilingual speech recognition;
Can perform speech translation and language identification | Open models, datasets, systems, and evaluation tools;
Distributed systems to accelerate LLM training and inference |
Statistics | |
GitHub Stars 90.3K | GitHub Stars - |
GitHub Forks 11.3K | GitHub Forks - |
Stacks 24 | Stacks 0 |
Followers 28 | Followers 0 |
Votes 1 | Votes 0 |
Integrations | |

It is the base model weights and network architecture of Grok-1, the large language model. Grok-1 is a 314 billion parameter Mixture-of-Experts model trained from scratch by xAI.

It is Google’s largest and most capable AI model. It is built to be multimodal, it can generalize, understand, operate across, and combine different types of info — like text, images, audio, video, and code.

It is a state-of-the-art foundational large language model designed to help researchers advance their work in this subfield of AI.

Try Grok 4 on GPT Proto. Access xAI’s most advanced 1.7T LLM with 130K context, multimodal support, and real-time data integration for dynamic analysis.

Use AI on aiasmr.lol to create soothing ASMR videos. Provide other tools for ASMR creators:script generator, idea generator, video to prompt, thumbnail generator, channel name generator.

Use e4tools’ free online paraphrasing tool to quickly rewrite sentences, paragraphs, and articles with AI for clarity, readability, and originality.

Build n8n workflows with AI and deploy in 30 seconds. Free hosting, workflow analyzer, and 100+ LLM models included.

Build smarter coding agents with the GLM-4.7 API, featuring multilingual coding, terminal tasks, and think-before-act reasoning.

Discover DeepSeek V4, a future-ready reasoning model for long context, clear outputs, and reliable automation. Join the preview and shape how teams work.

Kimi K2.5 is an open-weight native multimodal model from Moonshot AI, continued-trained on ~15T multimodal tokens for 256K context, visual coding, and agent swarms.