Meta AI's story begins in December 2013, when Mark Zuckerberg recruited Yann LeCun — one of the three "godfathers of deep learning" alongside Geoffrey Hinton and Yoshua Bengio — to lead Facebook AI Research (FAIR). LeCun, a professor at NYU and a pioneer of convolutional neural networks, brought instant credibility and a clear research philosophy: fundamental research, published openly, with no short-term product pressure. FAIR quickly became one of the most prolific and respected AI labs in the world, attracting top talent and producing influential work across computer vision, natural language processing, and self-supervised learning. The lab operated with unusual academic freedom for a corporate research division, and LeCun's presence gave it a gravity that few industry labs could match.
If Meta AI had done nothing else, PyTorch would be enough to cement its legacy. Released in 2016 as an evolution of the Torch framework, PyTorch became the dominant deep learning framework in research and, increasingly, in production. Its "define-by-run" dynamic computation graph was more intuitive than TensorFlow's original static graph approach, and the developer experience was simply better. By the early 2020s, the vast majority of AI research papers used PyTorch, and most of the models you interact with today — including those from OpenAI, Anthropic, and Mistral — were trained on it. Meta open-sourced PyTorch and eventually donated it to the Linux Foundation in 2022, making it genuinely community-governed. This wasn't pure altruism; by making PyTorch the standard, Meta ensured that the entire AI ecosystem built on infrastructure it understood deeply.
In February 2023, Meta released LLaMA (Large Language Model Meta AI), a family of models from 7B to 65B parameters, initially restricted to researchers. Within a week, the weights leaked online. Rather than fighting it, Meta leaned in: Llama 2 (July 2023) was released under a permissive license for both research and commercial use, and Llama 3 (April 2024) and Llama 4 (2025) continued the open-weights strategy with increasingly competitive models. This decision reshaped the industry. Before Llama, the prevailing assumption was that frontier models would remain proprietary. After Llama, a massive ecosystem of fine-tunes, quantizations, and derivative models emerged, and the open-weights movement became a serious competitive force. Meta's motivation was partly strategic — if AI models become commoditized, the value shifts to the platforms that deploy them, and Meta has 3+ billion users across Facebook, Instagram, WhatsApp, and Threads — but the impact on democratizing AI access was real regardless of the motive.
Mark Zuckerberg's personal investment in AI has been massive and highly visible. After the metaverse pivot failed to capture public imagination (and cost tens of billions), Zuckerberg repositioned Meta as an "AI-first" company in 2023-2024. The capital expenditure has been staggering: Meta built one of the largest GPU clusters in the world (initially 600,000+ H100 GPUs, with plans for more), spending over $30 billion annually on AI infrastructure by 2025. Meta AI, the company's consumer-facing assistant powered by Llama, was integrated across all Meta apps, making it one of the most widely deployed AI assistants in the world by sheer distribution. Zuckerberg has also been vocal about his philosophical stance: AI models should be open, concentration of AI power in closed labs is dangerous, and Meta's open approach is better for the ecosystem. Whether this is genuine conviction or competitive strategy against OpenAI and Google (whose closed models Meta can undercut with free alternatives) is a question that generates plenty of debate.
Beyond LLMs, Meta AI's research portfolio is remarkably broad. Their work in computer vision (DINOv2, Segment Anything), speech and translation (SeamlessM4T, covering 100+ languages), video generation, and embodied AI is among the best in the industry. The Segment Anything Model (SAM), released in 2023, did for image segmentation what AlphaFold did for protein folding — it made a previously difficult task trivially easy and freely available. FAIR, now led by Joelle Pineau (LeCun remains as Chief AI Scientist in an advisory capacity), continues to publish at an extraordinary pace. The strategic challenge for Meta is integration: turning all this research into products that make Meta's social platforms more engaging and its advertising more effective, without triggering the kind of regulatory backlash that a company with Meta's privacy history is particularly vulnerable to. The open-weights strategy also has limits — as models approach AGI-level capabilities, the question of whether it's responsible to release them openly becomes much harder to answer.