Sign up to save tools and stay up to date with the latest in AI
bg
bg
1

Meta says Llama 3 beats most other models, including Gemini

Apr 18, 2024 - theverge.com
Meta has released the next generation of its large language model, Llama 3, which it claims outperforms most current AI models. The new model, available to cloud providers like AWS and model libraries like Hugging Face, features two model weights with 8B and 70B parameters and offers text-based responses. Llama 3 demonstrated more diversity in answering prompts, fewer false refusals, and improved reasoning abilities. It also reportedly understands more instructions and writes better code than its predecessor. Benchmark tests showed Llama 3 outperforming similarly sized models like Google’s Gemma and Gemini, Mistral 7B, and Anthropic’s Claude 3.

Human evaluators also rated Llama 3 higher than other models, including OpenAI’s GPT-3.5. Meta created a new dataset for human evaluators to emulate real-world scenarios where Llama 3 might be used, including advice, summarization, and creative writing. The company is currently training larger versions of Llama 3, with over 400B parameters, which are expected to understand longer strings of instructions and data and be capable of more multimodal responses. However, Meta has not yet released a preview of these larger models.

Key takeaways:

  • Meta's next generation large language model, Llama 3, has been released to cloud providers and model libraries, offering improved performance over most current AI models.
  • Llama 3 features two model weights, with 8B and 70B parameters, and has shown more diversity in answering prompts, fewer false refusals, and better reasoning capabilities.
  • In benchmark testing, both sizes of Llama 3 outperformed similarly sized models like Google’s Gemma and Gemini, Mistral 7B, and Anthropic’s Claude 3.
  • Meta is currently training larger versions of Llama 3, with over 400B parameters, which are expected to understand longer strings of instructions and data and be capable of more multimodal responses.
View Full Article

Comments (0)

Be the first to comment!