Meta has introduced the latest addition to its Llama family of generative AI models: Llama 3.3 70B. This text-only model boasts impressive performance, rivaling that of Meta's largest Llama model, Llama 3.1 405B, but at a significantly lower cost. According to Ahmad Al-Dahle, VP of generative AI at Meta, the model leverages the latest advancements in post-training techniques to achieve this feat.
In a post on X, Al-Dahle shared a chart showcasing Llama 3.3 70B's superior performance compared to other industry-leading models, including Google's Gemini 1.5 Pro, OpenAI's GPT-4o, and Amazon's Nova Pro. The model is expected to deliver improvements in areas such as math, general knowledge, instruction following, and app use.
Llama 3.3 70B is available for download from the AI dev platform Hugging Face and other sources, including the official Llama website. This move is part of Meta's strategy to dominate the AI field with "open" models that can be used and commercialized for a range of purposes. While Meta's terms do impose some restrictions on certain developers, the model's availability has already led to over 650 million downloads.
Meta has also been utilizing Llama models for its own purposes, including powering its AI assistant, Meta AI. With nearly 600 million monthly active users, Meta AI is on track to become the most-used AI assistant in the world, according to CEO Mark Zuckerberg.
However, the open nature of Llama has presented both opportunities and challenges for Meta. In November, a report emerged that Chinese military researchers had used a Llama model to develop a defense chatbot, prompting Meta to make its Llama models available to U.S. defense partners. The company has also expressed concerns about complying with the EU's AI Act, citing uncertainty around the law's implementation and related provisions in the GDPR.
EU regulators have requested that Meta halt training on European user data while they assess the company's GDPR compliance. Meta has relented, while also endorsing an open letter calling for a "modern interpretation" of GDPR that doesn't "reject progress." The company is also investing heavily in its compute infrastructure to train and serve future generations of Llama models, with plans to build a $10 billion AI data center in Louisiana – the largest AI data center it's ever built.
The training of large language models like Llama 3.3 70B can be a costly endeavor. Meta's capital expenditures rose nearly 33% to $8.5 billion in Q2 2024, driven by investments in servers, data centers, and network infrastructure. According to Zuckerberg, the company will need 10x more compute than what was needed to train Llama 3 to train the next major set of Llama models, Llama 4.
As Meta continues to push the boundaries of AI development, the implications of its advancements will be felt across various industries. With Llama 3.3 70B, the company is poised to further solidify its position in the AI landscape, while also driving innovation and progress in areas such as natural language processing and machine learning.