This Week in AI: Meta's Llama 3 and the 'Open Source' Debate

This Week in AI: The Nuances of 'Open Source' and Meta's Llama 3
Keeping pace with the rapidly evolving field of Artificial Intelligence (AI) is a significant challenge. This week's AI landscape is dominated by discussions around the definition and application of "open source" in the context of AI models, particularly following Meta's release of its Llama 3 series.
Meta's Llama 3 Release and the "Open Source" Debate
Meta has launched Llama 3 8B and Llama 3 70B, its latest generative AI models, capable of text analysis and generation. Meta has described these models as "open sourced," positioning them as foundational elements for developers to build upon. The company stated, "We believe these are the best open source models of their class, period," emphasizing their commitment to an "open source ethos of releasing early and often."
However, this claim has ignited a debate, as the Llama 3 models, like their predecessor Llama 2, come with certain licensing restrictions that deviate from the strictest definition of open source. Key restrictions include:
- Prohibition on training other models: Developers cannot use Llama models to train competing AI models.
- Special license for large-scale applications: Companies with over 700 million monthly users must obtain a special license from Meta.
This situation highlights a broader industry trend where the term "open source" is being applied loosely to AI models, leading to philosophical debates about its true meaning in this new technological paradigm.
The Evolving Definition of Open Source in AI
The debate over what constitutes "open source" in AI is not new. A 2023 study by researchers from Carnegie Mellon, the AI Now Institute, and the Signal Foundation pointed out that many AI models marketed as "open source" have significant caveats. These often include:
- Secrecy of training data: The data used to train these models is frequently kept private.
- High computational requirements: Running these models often demands substantial computing power, inaccessible to many developers.
- Prohibitive fine-tuning costs: The expense and complexity of fine-tuning these models can be a barrier.
This ambiguity raises questions about whether core AI components, like model embeddings, can even be subject to traditional intellectual property (IP) mechanisms like copyright, which underpin open source licensing.
Implications of "Open Source" AI
The co-opting of the "open source" phrase by major tech companies like Meta has significant implications. According to the Carnegie Mellon study, these releases often generate substantial free marketing and provide technical and strategic advantages to the companies releasing them. Meanwhile, the broader open source community often sees only marginal benefits.
Instead of democratizing AI, these "open source" initiatives can inadvertently entrench and expand centralized power structures. This is a critical point to consider when evaluating new "open source" model releases.
Other AI News This Week:
- Meta AI Chatbot Upgrade: Meta has enhanced its AI chatbot across Facebook, Messenger, Instagram, and WhatsApp with a Llama 3-powered backend, offering faster image generation and web search integration.
- AI-Generated Pornography: Meta's Oversight Board is examining how the company's platforms handle explicit AI-generated images.
- Snap Watermarks: Snap plans to add watermarks to AI-generated images exported from its app to distinguish them.
- Boston Dynamics' Electric Atlas: Hyundai-owned Boston Dynamics unveiled its next-generation Atlas humanoid robot, now fully electric and with a friendlier appearance.
- MenteeBot: The founder of Mobileye has launched a startup, MenteeBot, focused on developing bipedal robotics systems.
- Reddit Translations: Reddit is working on an AI-powered language translation feature to expand its global reach and an assistive moderation tool.
- LinkedIn Premium AI: LinkedIn is testing a premium subscription for company pages that includes AI content generation tools.
- Project Bellwether: Google's X is using AI to improve the prediction and identification of natural disasters.
- AI for Child Protection: UK regulators are exploring how AI can help detect and remove illegal content online to protect children.
- OpenAI in Japan: OpenAI is opening a Tokyo office and developing a GPT-4 model optimized for the Japanese language.
Research and Insights:
- AI Persuasion: Swiss researchers found that AI chatbots, particularly GPT-4, can be more persuasive than humans in debates, especially when armed with personal information. This raises concerns about AI's influence on public opinion and elections.
- AI Safety and Alignment: Researchers Stuart Russell and Michael Cohen are exploring how to ensure AI remains "human-compatible," addressing the challenge of testing advanced AI agents that might circumvent evaluation methods. They propose restricting hardware access for such agents.
- Neuromorphic Computing: The development of advanced computing systems like Venado (supercomputer for AI research) and Hala Point (neuromorphic system with 1.15 billion artificial neurons) signifies a move towards more brain-like computation, complementing traditional AI approaches.
Event Spotlight:
- TechCrunch All Stage: Taking place on July 15, 2025, in Boston, this event focuses on connecting founders and VCs across all stages of growth. Early bird registration offers significant savings.
This week's AI developments underscore the critical need for clear definitions and ethical considerations as the technology continues its rapid advancement.
Original article available at: https://techcrunch.com/2024/04/20/this-week-in-ai-when-open-source-isnt-so-open/