Lukasz Kaiser
2 papers ยท Latest:
Machine Learning
Evaluating Large Language Models Trained on Code
Codex, a GPT model fine-tuned on GitHub code, significantly outperforms prior models in generating correct Python programs from docstrings, demonstrating strong code synthesis capabilities.
2107.03374
Natural Language ProcessingAttention Is All You Need
The paper introduces the Transformer, a novel neural network architecture based solely on attention mechanisms that outperforms traditional recurrent and convolutional models in sequence transduction tasks like machine translation.
1706.03762
๐ฌ Weekly AI Paper Digest
Get the top 10 AI/ML arXiv papers from the week โ summarized, scored, and delivered to your inbox every Monday.