Meta is a large-scale language modelLlama 3.3” was released on Saturday, December 6, 2024. At the time of article creation, a model with 70B parameters has been released, and Amazon’s basic model “Amazon Nova ProIt is being promoted as having lower cost and higher performance than the previous generation model, and that it has performance comparable to the previous generation’s higher-end models.
llama-models/models/llama3_3/MODEL_CARD.md at main · meta-llama/llama-models · GitHub
https://github.com/meta-llama/llama-models/blob/main/models/llama3_3/MODEL_CARD.md
As we continue to explore new post-training techniques, today we’re releasing Llama 3.3 — a new open source model that delivers leading performance and quality across text-based use cases such as synthetic data generation at a fraction of the inference cost. pic.twitter.com/BNoV2czGKL
— AI at Meta (@AIatMeta) December 6, 2024
Llama 3.3 70B is trained on over 15 trillion tokens collected until December 2023. The context length is 128,000 and the supported languages are English, German, French, Italian, Portuguese, Hindi, Spanish, and Thai.
Below is a table listing the benchmark results of Llama 3.3 70B and various AI models. Llama 3.3 70B outperforms Llama 3.1 70B in most tests, and also outperforms Amazon Nova Pro in MMLU, which evaluates overall performance, and MATH, which evaluates math-related performance. . In addition, in some benchmarks such as coding performance and math performance, it has exceeded the previous generation top model Llama 3.1 405B.
Meta claims that Llama 3.3 70B uses cost-effective inference to achieve performance similar to Llama 3.1 405B while being able to run locally on a typical developer workstation.
Improvements in Llama 3.3 were driven by a new alignment process and progress in online RL techniques. This model delivers similar performance to Llama 3.1 405B with cost effective inference that’s feasible to run locally on common developer workstations.
— AI at Meta (@AIatMeta) December 6, 2024
Llama 3.3 70BMeta’s official websiteIt is available at , and also distributed by Hugging Face. Please note that the license isLlama 3.3 Community License” and the number of monthly active users is less than 700 million, commercial use is possible.
meta-llama/Llama-3.3-70B-Instruct · Hugging Face
https://huggingface.co/meta-llama/Llama-3.3-70B-Instruct
Copy the title and URL of this article
・Related articles
Meta uses over 100,000 NVIDIA H100s to train Llama-4 – GIGAZINE
Amazon announces “Amazon Nova”, a multimodal generative AI model available on AWS – GIGAZINE
Google releases “PaliGemma 2” visual language model that is easy to fine tune – GIGAZINE
Mistral AI releases 124B size multimodal AI model “Pixtral Large”, outperforming GPT-4o, Gemini-1.5 Pro, etc. – GIGAZINE
A list summarizing the AI inference processing capabilities of NVIDIA graphics cards and Apple chips, which will help you decide which graphics card or Mac to buy – GIGAZINE
in Software, Posted by log1o_hf
You can read the machine translated English article here.