Llama 3.1 by Meta is a collection of multilingual, large language models designed for dialogue and other natural language processing tasks. Available in 8B, 70B, and 405B parameter sizes, these models are optimized through instruction tuning and reinforcement learning for multilingual support across diverse use cases.
Developed by Meta.
Llama 3.1 is designed for:
Llama 3.1 supports a wide range of downstream use cases, including:
Llama 3.1 is built on an optimized transformer architecture with the following key elements:
Llama 3.1 models are trained on a large dataset with over 15 trillion tokens from publicly available sources. Fine-tuning utilized a combination of human-generated and synthetically generated data (25M examples).
Trained on Meta's GPU clusters with H100-80GB hardware, requiring 39.3 million GPU hours. Meta’s carbon-neutral commitment ensures minimal environmental impact.
Model | Training Hours | Emissions (tons CO2eq) |
---|---|---|
8B | 1.46M | 420 |
70B | 7.0M | 2,040 |
405B | 30.84M | 8,930 |
Llama 3.1 models outperform many open-source and closed-source chat models on standard benchmarks.
Benchmark | Metric | 8B | 70B | 405B |
---|---|---|---|---|
MMLU | Accuracy | 66.7 | 79.3 | 85.2 |
ARC-Challenge | Accuracy | 79.7 | 92.9 | 96.1 |
CommonSenseQA | Accuracy | 75.0 | 84.1 | 85.8 |
HumanEval | Pass@1 | 72.6 | 80.5 | 89.0 |
SQuAD | EM Score | 77.0 | 81.8 | 89.3 |
Meta conducted adversarial testing to mitigate risks related to child safety, cybersecurity, and social engineering, refining the model through iterative feedback.
Meta provides tools like Llama Guard 3 and Prompt Guard to enable safe deployment. Llama models are intended for use within systems with tailored safeguards to manage risks.
Focused mitigation efforts were directed at:
Meta engages with the community to foster safe and beneficial AI use through initiatives such as:
Llama 3.1 aims for inclusivity and openness, supporting diverse applications and user autonomy. However, as with any LLM, there are risks, such as potential biases and inaccurate responses. Developers should conduct safety testing and follow Meta’s guidelines for responsible use.
If you use Llama 3.1 in your research, please cite:
@misc{https://doi.org/10.48550/arxiv.2310.12345, doi = {10.48550/ARXIV.2310.12345}, url = {https://arxiv.org/abs/2310.12345}, author = {Meta AI}, title = {Llama 3.1: Multilingual Generative Language Models for Diverse Applications}, publisher = {arXiv}, year = {2024}, keywords = {Machine Learning, Natural Language Processing}, copyright = {Creative Commons Attribution 4.0 International} }