Meta AI: LLaMA and Open Source Strategy
While OpenAI and Google closed their doors, Mark Zuckerberg kicked them open. Why Meta is giving away billions of dollars of IP for free.
Meta AI: LLaMA and Open Source Strategy
In a tech landscape defined by walled gardens, Meta (Facebook) made a shocking pivot. Instead of hoarding their AI research, they released their state-of-the-art models, LLaMA (Large Language Model Meta AI), for free.
Mark Zuckerberg’s strategy is clear: Commoditize the complement. If the “intelligence” layer becomes free and open source, then proprietary competitors (like OpenAI) lose their moat. The value shifts to the platforms where people use the AI—Instagram, WhatsApp, and Facebook.
The LLaMA Timeline
LLaMA 1 (Feb 2023)
Released to researchers only. Ideally, it wouldn’t leak. It leaked immediately via 4chan torrents. Instead of suing, Meta watched as the open-source community optimized it, quantized it, and made it run on MacBooks. The innovation explosion was undeniable.
LLaMA 2 (July 2023)
Meta embraced the community. LLaMA 2 was released with a commercially permissive license. It became the default standard for every startup, enterprise, and hobbyist.
LLaMA 3 (April 2024)
The beast. Trained on 15 Trillion tokens.
- 8B: The best small model in the world.
- 70B: Rivals GPT-4 class models.
- 405B: The first open-weights model to truly challenge the closed frontier models (GPT-4o, Claude 3.5 Sonnet).
Why Open Weights Matter
When Meta releases “Open Source AI,” they are releasing the Weights (the trained neural network), not the training data or the full training code. However, this is enough.
- Privacy: Banks and hospitals can’t send data to OpenAI’s API. They can download LLaMA, run it on their own servers (air-gapped), and ensure zero data leaks.
- Fine-Tuning: You can teach LLaMA to speak “Medical,” “Legal,” or “Harry Potter.” You can’t easily do that with closed models.
- Cost: Once you download the model, you only pay for the electricity to run it. No per-token API fees.
The Ecosystem Effect
Because LLaMA is the standard:
- Hardware is optimized for LLaMA architectures.
- Software (like llama.cpp, vLLM, Ollama) is built for LLaMA first.
- Research papers use LLaMA as the baseline.
Meta has effectively become the “Linux of AI.” Just as Linux powers the internet’s servers, LLaMA powers the world’s custom AI applications.
FAIR vs GenAI
Meta has two wings:
- FAIR (Fundamental AI Research): Led by Yann LeCun (Godfather of AI). Focuses on “World Models” and getting AI to understand physics and reality, not just predict text. LeCun is skeptical of LLMs (“they are just glorified autocomplete”).
- GenAI Product: Focuses on shipping LLaMA and integrating it into Meta Ray-Ban glasses and Instagram.
This internal tension drives rapid progress on both theoretical and practical fronts. By giving away the brain, Meta ensures they own the ecosystem the brain lives in.