Thibaut Lavril
3 papers ยท Latest:
Machine Learning
Mixtral of Experts
Mixtral 8x7B is a Sparse Mixture of Experts language model that achieves performance on par with or exceeding much larger models like Llama 2 70B and GPT-3.5 by dynamically routing tokens through a subset of experts.
2401.04088
Natural Language ProcessingLlama 2: Open Foundation and Fine-Tuned Chat Models
Llama 2 introduces a range of open-source large language models, including fine-tuned chat models that outperform existing open-source alternatives in benchmarks and human evaluations.
2307.09288
Natural Language ProcessingLLaMA: Open and Efficient Foundation Language Models
LLaMA presents a suite of open, efficient language models that achieve state-of-the-art performance using only publicly available data.
2302.13971
๐ฌ Weekly AI Paper Digest
Get the top 10 AI/ML arXiv papers from the week โ summarized, scored, and delivered to your inbox every Monday.