Rewon Child
3 papers ยท Latest:
Natural Language Processing
PaLM: Scaling Language Modeling with Pathways
PaLM is a 540-billion parameter Transformer language model that achieves state-of-the-art few-shot learning performance across diverse benchmarks, demonstrating significant benefits from scaling.
2204.02311
Natural Language ProcessingLanguage Models are Few-Shot Learners
GPT-3, a 175 billion parameter language model, demonstrates strong few-shot learning abilities across diverse NLP tasks without task-specific fine-tuning.
2005.14165
Machine LearningScaling Laws for Neural Language Models
This paper identifies power-law scaling relationships between language model performance and factors like model size, dataset size, and compute, enabling optimal training strategies under fixed compute budgets.
2001.08361
๐ฌ Weekly AI Paper Digest
Get the top 10 AI/ML arXiv papers from the week โ summarized, scored, and delivered to your inbox every Monday.