Xiangru Tang
4 papers ยท Latest:
The Last Human-Written Paper: Agent-Native Research Artifacts
Ara is a new protocol for machine-executable research packages, enhancing AI's ability to understand, reproduce, and extend scientific work by preserving full research context.
StarCoder 2 and The Stack v2: The Next Generation
StarCoder2 is a next-generation open-source Code LLM trained on a vastly expanded and diverse dataset, achieving state-of-the-art performance on multiple code benchmarks while being more parameter-efficient than larger models.
OctoPack: Instruction Tuning Code Large Language Models
OctoPack introduces instruction tuning for code LLMs using a massive dataset of Git commits, achieving state-of-the-art results on multi-language coding benchmarks without relying on OpenAI data.
Crosslingual Generalization through Multitask Finetuning
This paper demonstrates that multitask finetuning of large multilingual language models on English and machine-translated prompts enables strong zero-shot crosslingual generalization to many languages, including those unseen during training.
๐ฌ Weekly AI Paper Digest
Get the top 10 AI/ML arXiv papers from the week โ summarized, scored, and delivered to your inbox every Monday.