Guillaume Lample
2 papers ยท Latest:
Machine Learning
Mixtral of Experts
Mixtral 8x7B is a Sparse Mixture of Experts language model that achieves performance on par with or exceeding much larger models like Llama 2 70B and GPT-3.5 by dynamically routing tokens through a subset of experts.
2401.04088
Natural Language ProcessingLLaMA: Open and Efficient Foundation Language Models
LLaMA presents a suite of open, efficient language models that achieve state-of-the-art performance using only publicly available data.
2302.13971
๐ฌ Weekly AI Paper Digest
Get the top 10 AI/ML arXiv papers from the week โ summarized, scored, and delivered to your inbox every Monday.