Christopher Ré
2 papers · Latest:
Machine Learning
FlashAttention: Fast and Memory-Efficient Exact Attention with IO-Awareness
FlashAttention is an IO-aware exact attention algorithm that significantly speeds up Transformer training and enables longer context lengths by optimizing GPU memory access patterns.
2205.14135
Machine LearningCan Foundation Models Wrangle Your Data?
This paper demonstrates that large foundation models can effectively perform classical data cleaning and integration tasks without task-specific training, achieving state-of-the-art results.
2205.09911
📬 Weekly AI Paper Digest
Get the top 10 AI/ML arXiv papers from the week — summarized, scored, and delivered to your inbox every Monday.