Grok models have evolved rapidly: Grok-1 was a 314B parameter MoE model (open-sourced), Grok-2 and Grok-3 are proprietary models that benchmark competitively with frontier models from other labs. The company's differentiator is its integration with X — Grok can reference recent posts, trends, and conversations in a way that models trained on static datasets can't.
xAI's strategy is heavily compute-centric. The Colossus cluster in Memphis, Tennessee, with 100,000+ NVIDIA H100 GPUs, is among the largest AI training installations in the world. The bet is that raw compute scale, combined with the X data advantage, can close or eliminate the gap with labs that have been operating longer. Whether compute alone is sufficient — versus needing the research depth and alignment expertise of established labs — remains to be seen.