It aims to pre-train a 1.1 billion parameter language model on 3 trillion tokens of text data. It is compact with only 1.1B parameters. This compactness allows it to cater to a multitude of applications demanding a restricted computation and memory footprint. | Try Grok 4 on GPT Proto. Access xAI’s most advanced 1.7T LLM with 130K context, multimodal support, and real-time data integration for dynamic analysis. |
1.1 billion parameter language model;
Trained on 3 trillion tokens of text data;
Uses the same architecture and tokenizer as Llama 2;
Compact and fast | grok-4, grok-4 api, grok ai |
Statistics | |
GitHub Stars 8.8K | GitHub Stars - |
GitHub Forks 572 | GitHub Forks - |
Stacks 0 | Stacks 0 |
Followers 2 | Followers 1 |
Votes 0 | Votes 1 |

It is the base model weights and network architecture of Grok-1, the large language model. Grok-1 is a 314 billion parameter Mixture-of-Experts model trained from scratch by xAI.

Professional 4K AI-powered image generation and editing. Create stunning 4K visuals with self-correction technology and multi-image understanding.

Is an all-in-one platform featuring GPT-5, Flux, Claude, Qwen Image, Kling, Hailuo, and more. Always the latest AI models, updated regularly

Banana-Pro.com offers fast, high-quality AI image & video generation powered by Nano Banana Pro, Sora2 and more. Built-in prompt optimizer, no watermarks, no invite code.

Creating safe artificial general intelligence that benefits all of humanity. Our work to create safe and beneficial AI requires a deep understanding of the potential risks and benefits, as well as careful consideration of the impact.

It is a next-generation AI assistant. It is accessible through chat interface and API. It is capable of a wide variety of conversational and text-processing tasks while maintaining a high degree of reliability and predictability.

It is Google’s largest and most capable AI model. It is built to be multimodal, it can generalize, understand, operate across, and combine different types of info — like text, images, audio, video, and code.

It is a state-of-the-art foundational large language model designed to help researchers advance their work in this subfield of AI.

It is a large multimodal model (accepting text inputs and emitting text outputs today, with image inputs coming in the future) that can solve difficult problems with greater accuracy than any of our previous models, thanks to its broader general knowledge and advanced reasoning capabilities.

It offers an API to add cutting-edge language processing to any system. Through training, users can create massive models customized to their use case and trained on their data.