Meta Platforms has introduced the largest version of its Llama 3 artificial intelligence models, highlighting significant improvements in language and math capabilities. This new model aims to rival paid models from competitors like OpenAI. The Llama 3 model can now converse in eight languages, produce higher-quality computer code, and solve complex math problems more effectively than its predecessors. According to Meta’s announcements in blog posts and a research paper, the new model features 405 billion parameters, significantly surpassing the previous version, though it remains smaller than some leading models from competitors. In comparison, OpenAI's GPT-4 model reportedly boasts one trillion parameters, and Amazon is working on a model with two trillion parameters. Tech companies are in a fierce race to prove the value of their increasingly resource-intensive large language models, aiming to justify the substantial investments by showcasing significant advancements in areas like advanced reasoning. Meta is also releasing updated versions of its lighter-weight Llama 3 models with 8 billion and 70 billion parameters. All three models are multilingual and can handle larger user requests through an expanded "context window." This feature, according to Ahmad Al-Dahle, Meta’s head of generative AI, will enhance the experience of generating computer code, addressing a major request from the community. “That was the number one feedback we got from the community,” Al-Dahle told Reuters, explaining that the larger context windows provide the models with a longer memory, aiding in the processing of multi-step requests. Meta's strategy of offering its Llama models mostly free-of-charge aims to foster innovation and increase engagement on its core social networks, despite some investor concerns about the costs involved. By providing free models, Meta could attract developers away from paid alternatives, potentially undermining the business models of its rivals. Meta highlighted gains in key math and knowledge tests, suggesting that its largest Llama 3 model is nearly on par with, and in some cases surpasses, leading models like Anthropic's Claude 3.5 Sonnet and OpenAI's GPT-4o. For instance, on the MATH benchmark for competition-level math word problems, Meta's model scored 73.8, compared to GPT-4o's 76.6 and Claude 3.5 Sonnet's 71.1. On the MMLU benchmark, covering various subjects across math, science, and the humanities, Llama 3 scored 88.6, close to GPT-4o's 88.7 and higher than Claude 3.5 Sonnet's 88.3. Meta researchers also hinted at upcoming "multimodal" versions of the models, which will integrate image, video, and speech capabilities with the core Llama 3 text model. Early experiments suggest these models can compete with other multimodal models such as Google's Gemini 1.5 and Anthropic's Claude 3.5 Sonnet. https://github.com/molikikyuhs https://github.com/hedsazaslr https://github.com/molopososlr https://github.com/houtedmsnsr https://github.com/hsnmahlahn https://github.com/dundunduo https://github.com/slurxsumma https://github.com/gdzlkngslur https://pastelink.net/0ubb2juq https://muckrack.com/secafe-faba https://www.bitsdujour.com/profiles/X6HzP7
Technology