Sewon Min
2 papers ยท Latest:
Natural Language Processing
EMO: Pretraining Mixture of Experts for Emergent Modularity
EMO is a new Mixture-of-Experts model that achieves emergent modularity, allowing efficient selective expert use for memory-constrained LLM deployment.
2605.06663
Information RetrievalRAG over Thinking Traces Can Improve Reasoning Tasks
This paper shows that using "thinking traces" as a retrieval corpus significantly enhances RAG performance on complex reasoning tasks like math and code.
2605.03344
๐ฌ Weekly AI Paper Digest
Get the top 10 AI/ML arXiv papers from the week โ summarized, scored, and delivered to your inbox every Monday.