Trends
Meta releases early versions of Llama 3 multimodal AI model
Meta Platforms released early versions of its latest large language model, Llama 3, with new computer coding capabilities and the ability to process image commands. The equipped image generator will update pictures in real-time while users type prompts, as it races to catch up to generative AI marke…

Headline
Meta Platforms released early versions of its latest large language model, Llama 3, with new computer coding capabilities and the ability to process image commands. The equipped image generator will update pictures in real-time while users type prompts, as it races to catch up…
Context
Meta Platforms released early versions of its latest large language model, Llama 3, with new computer coding capabilities and the ability to process image commands. The equipped image generator will update pictures in real-time while users type prompts, as it races to catch up to generative AI market leader OpenAI. See CEO Mark Zuckerburg’s video explainer. Versions of Llama 3 planned for release in the coming months will also be capable of “multimodality,” meaning they can generate both text and images though for now the model will output only text, Meta chief product officer Chris Cox said in an interview.
Evidence
Pending intelligence enrichment.
Analysis
The models will be integrated into the virtual assistant Meta AI , which the company is pitching as the most sophisticated of its free-to-use peers. More advanced reasoning, like the ability to craft longer multi-step plans, will follow in subsequent versions. Also read: Meta debuts an ‘all-rounder’ MTAI chip 3 times faster than previous The inclusion of images in the training of Llama 3 would enhance an update rolling out this year to the Ray-Ban Meta smart glasses, a partnership with glasses maker Essilor Luxoticca , enabling Meta AI to identify objects seen by the wearer and answer questions about them, said Chris Cox. The Llama 2 model is unable to understand basic context, Meta reduces these problems in Llama 3 by using “high-quality data” to allow the model to recognise nuances. Rival Google has run into similar issues and recently suspended the use of its Gemini AI image-generating tool after it was criticised for inaccurate depictions of historical figures.
Key Points
- Meta Platforms released early versions of its latest large language model, Llama 3, with new computer coding capabilities and the ability to process image commands. The models will be integrated into the virtual assistant Meta AI, which the company is pitching as the most…
- Versions of Llama 3 planned for release in the coming months will also be capable of “multimodality,” meaning they can generate both text and images, as it races to catch up to generative AI market leader OpenAI.
- The Llama 2 model is unable to understand basic context, Meta reduces these problems in Llama 3 by using “high-quality data” to allow the model to recognise nuances. The demand for data for generative AI models has become a major source of tension in the development of the…
Actions
Pending intelligence enrichment.





