ArXiv TLDR

Jason Wei

6 papers ยท Latest:

Natural Language Processing

GPT-4 Technical Report

GPT-4 is a large-scale multimodal Transformer model achieving human-level performance on professional and academic benchmarks through advanced training and alignment techniques.

2303.08774
Natural Language Processing

Transcending Scaling Laws with 0.1% Extra Compute

UL2R fine-tuning significantly improves large language model performance and scaling efficiency with only 0.1% extra compute, enabling substantial computational savings and emergent abilities.

2210.11399
Natural Language Processing

PaLM: Scaling Language Modeling with Pathways

PaLM is a 540-billion parameter Transformer language model that achieves state-of-the-art few-shot learning performance across diverse benchmarks, demonstrating significant benefits from scaling.

2204.02311
Natural Language Processing

Self-Consistency Improves Chain of Thought Reasoning in Language Models

Self-consistency is a new decoding strategy that improves chain-of-thought reasoning in language models by sampling diverse reasoning paths and selecting the most consistent answer.

2203.11171
Natural Language Processing

Chain-of-Thought Prompting Elicits Reasoning in Large Language Models

Chain of thought prompting, which involves providing intermediate reasoning steps in prompts, significantly enhances large language models' performance on complex reasoning tasks.

2201.11903
Natural Language Processing

Finetuned Language Models Are Zero-Shot Learners

Instruction tuning large language models on diverse NLP tasks significantly enhances their zero-shot learning capabilities, outperforming much larger models like GPT-3.

2109.01652

๐Ÿ“ฌ Weekly AI Paper Digest

Get the top 10 AI/ML arXiv papers from the week โ€” summarized, scored, and delivered to your inbox every Monday.