Yuanzhi Li
3 papers ยท Latest:
Textbooks Are All You Need II: phi-1.5 technical report
phi-1.5 is a 1.3B parameter Transformer model trained on high-quality textbook-style data that achieves reasoning and coding performance comparable to much larger models.
TinyStories: How Small Can Language Models Be and Still Speak Coherent English?
TinyStories demonstrates that very small language models under 10 million parameters can generate coherent, fluent English stories using a simple synthetic dataset and a novel GPT-4-based evaluation framework.
LoRA: Low-Rank Adaptation of Large Language Models
LoRA introduces a low-rank adaptation method that enables efficient fine-tuning of large language models by injecting trainable low-rank matrices while freezing original weights, drastically reducing trainable parameters and resource usage without sacrificing performance.
๐ฌ Weekly AI Paper Digest
Get the top 10 AI/ML arXiv papers from the week โ summarized, scored, and delivered to your inbox every Monday.