Skip to content

puneet-chandna/awesome-LLM-papers

Folders and files

NameName
Last commit message
Last commit date

Latest commit

ย 

History

14 Commits
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 
ย 

Repository files navigation

๐Ÿง  Awesome LLM Papers

Daily LLM Papers Awesome Stars Updates

PRs Welcome

Read what matters. Skip the noise. ๐ŸŽฏ

Today's Pick โ€ข This Week โ€ข Hall of Fame โ€ข Categories โ€ข Contribute


Papers New Papers Contributors Ratings

๐Ÿ”ฅ Today's Pick

Authors: Parshin Shojaee, Iman Mirzadeh, Keivan Alizadeh, Maxwell Horton, Samy Bengio, Mehrdad Farajtabar โ€ข Apple Machine Learning Research

Why this matters: This groundbreaking paper systematically exposes fundamental limitations of current "reasoning" models through controllable puzzle environments. It reveals that LRMs face complete accuracy collapse beyond certain complexities and paradoxically reduce reasoning effort as problems get harder..

Key Innovations:

  • ๐Ÿ”ธ Controllable puzzle environments for systematic complexity manipulation
  • ๐Ÿ”ธ Three performance regimes identified: low (standard models win), medium (LRMs excel), high (both fail)
  • ๐Ÿ”ธ Counter-intuitive scaling discovery: reasoning effort decreases with problem complexity despite token budget
  • ๐Ÿ”ธ Rigorous methodology avoiding data contamination issues of standard benchmarks

Resources:

Impact Score:

+ Performance: โ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–‘ 92%
+ Innovation:  โ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆ 98%
+ Practicality:โ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–ˆโ–‘โ–‘ 85% 

๐Ÿ”ฅ Trending Topics

"Hot research areas this month - Where the field is moving"

Topic Papers Why It's Trending
๐Ÿš€ Mixture of Experts (MoE) 8 papers Efficient scaling beyond dense models - Mixtral, DeepSeek success
๐Ÿ’ก Million Token Context 6 papers Breaking the context barrier - full books & codebases in one prompt
๐Ÿค– Autonomous Agents 12 papers From chatbots to actual workers - AutoGPT evolution
๐Ÿ”„ Self-Improving Models 5 papers Models that enhance themselves without human intervention

๐Ÿ“† This Week's Essential Reads

Click to expand this week's papers (January 13-19, 2025)
Day Paper Impact TL;DR
Mon ๐Ÿงฎ ๐Ÿง  Self-Questioning Language Models reasoning self-improvement LLMs improve without external data by generating and solving their own questions through asymmetric self-play
Tue โšก R-Zero: Self-Evolving Reasoning LLM from Zero Data autonomous-learning reasoning Challenger-Solver framework boosts Qwen3-4B by +6.49 points on math benchmarks without any human data
Wed โšก Mercury: Ultra-Fast Language Models Based on Diffusion speed inference Maintains character consistency across multiple generations
Thu ๐Ÿค Multi-Agent Collaboration Mechanisms: A Survey of LLMs collaboration agents Comprehensive framework for LLM-based multi-agent systems - covers cooperation, competition, and coordination protocols
Fri Tracing the thoughts of a large language model research testing We investigate the internal mechanisms used by Claude 3.5 Haiku โ€” Anthropic's lightweight production model

๐Ÿ“š Must-Read Papers (Hall of Fame)

๐Ÿ›๏ธ Papers that fundamentally changed the field

Paper Impact Why Essential Resources

Attention Is All You Need
Vaswani et al., 2017

๐Ÿ† Foundational
architecture

Created the Transformer architecture that powers all modern LLMs. Replaced RNNs with self-attention, enabling parallelization and scaling.

๐Ÿ“„ ๐Ÿ’ป ๐Ÿ“Š

GPT-3: Language Models are Few-Shot Learners
Brown et al., 2020

๐Ÿš€ Scale
emergence

Proved that scale leads to emergent abilities. In-context learning without fine-tuning revolutionized how we use LLMs.

๐Ÿ“„ ๐Ÿ” ๐Ÿ“Š

Constitutional AI: Harmlessness from AI Feedback
Bai et al., 2022

๐Ÿ›ก๏ธ Safety
alignment

Introduced RLAIF - training AI systems to be helpful and harmless using AI feedback instead of human feedback.

๐Ÿ“„ ๐Ÿ’ป ๐ŸŽฅ

Chain-of-Thought Prompting
Wei et al., 2022

๐Ÿงฎ Reasoning
prompting

Simple prompting technique that dramatically improves reasoning by asking models to think step-by-step.

๐Ÿ“„ ๐Ÿ’ก ๐Ÿ“Š

RLHF: Training with Human Feedback
Christiano et al., 2017

๐ŸŽฏ Alignment
training

The technique behind ChatGPT's success. Aligns model outputs with human preferences through reinforcement learning.

๐Ÿ“„ ๐Ÿ’ป ๐Ÿ“š

View all foundational papers โ†’


๐Ÿ“š Browse by Category

Transformers, SSMs, MoE, Novel designs
๐Ÿ“„ 10 papers ย |ย  ๐Ÿ”ฅ

CoT, Planning, Tool use, Autonomous systems
๐Ÿ“„ 6 papers ย |ย  ๐Ÿ”ฅ๐Ÿ”ฅ

Quantization, Pruning, Fast inference
๐Ÿ“„ 7 papers ย |ย  ๐Ÿ”ฅ

RLHF, DPO, Fine-tuning, PEFT methods
๐Ÿ“„ 5 papers ย |ย  โ†’

Vision-Language, Audio, Video, Any-to-any
๐Ÿ“„ 8 papers ย |ย  ๐Ÿ”ฅ

๐Ÿ“š RAG & Knowledge

Retrieval systems, Long context, Memory
๐Ÿ“„ 8 papers ย |ย  ๐Ÿ”ฅ๐Ÿ”ฅ๐Ÿ”ฅ

๐Ÿ›ก๏ธ Safety & Security

Jailbreaks, Alig nment, Robustness, Ethics
๐Ÿ“„ 4 papers ย |ย  โ†’

Interpretability, Mechanistic, Evaluations
๐Ÿ“„ 3 papers ย |ย  โ†’


๐Ÿ—๏ธ Model Architectures

View papers (10 total) โ€ข ๐Ÿ”ฅ Hot area

Latest Additions:

Foundational Papers:

Tags: transformer state-space mixture-of-experts attention-mechanisms

View all architecture papers โ†’


๐Ÿงฎ Reasoning & Agents

View papers (6 total) โ€ข ๐Ÿ”ฅ๐Ÿ”ฅ Very hot area

Agent Systems:

Reasoning Methods:

Tags: agents chain-of-thought planning tool-use reasoning

View all reasoning papers โ†’


โšก Efficiency & Scaling

View papers (7 total) โ€ข ๐Ÿ”ฅ Hot area

Inference Optimization:

Model Compression:

Tags: quantization pruning distillation inference deployment

View all efficiency papers โ†’


๐ŸŽฏ Training & Alignment

View papers (5 total)

Alignment Methods:

Fine-tuning Techniques:

Tags: rlhf fine-tuning peft instruction-tuning alignment

View all training papers โ†’


๐ŸŽจ Multimodal Models

View papers (8 total) โ€ข ๐Ÿ”ฅ Hot area

Vision-Language:

Generation:

Tags: vision-language image-generation video audio multimodal

View all multimodal papers โ†’


๐Ÿ“š RAG & Knowledge

View papers (4 total) โ€ข ๐Ÿ”ฅ๐Ÿ”ฅ๐Ÿ”ฅ Hottest area

RAG Systems:

Long Context:

Tags: retrieval rag long-context memory knowledge-bases

View all RAG papers โ†’


๐Ÿ›ก๏ธ Safety & Security

View papers (4 total)

Safety Research:

Alignment:

Tags: jailbreaks alignment safety robustness red-teaming

View all safety papers โ†’


๐Ÿ”ฌ Analysis & Theory

View papers (3 total)

Tags: interpretability mechanistic theory analysis evaluation

View all analysis papers โ†’


๐Ÿท๏ธ Explore by Tags

moe long-context rag agents

๐Ÿ“Š Most Used Tags

transformer (67) โ€ข efficient (45) โ€ข reasoning (38) โ€ข open-source (35) โ€ข multimodal (28)

production-ready (25) โ€ข breakthrough (22) โ€ข sota (20) โ€ข chain-of-thought (18) โ€ข rlhf (15)

๐ŸŽฏ Quick Filters

By Impact: ๐Ÿ† breakthrough โ€ข โญ sota โ€ข ๐Ÿš€ production-ready โ€ข ๐Ÿงช experimental

By Org: ๐ŸŸ  openai โ€ข ๐Ÿ”ท anthropic โ€ข ๐Ÿ”ด google โ€ข ๐Ÿ”ต meta โ€ข ๐ŸŽ“ academic

By Size: <1B โ€ข 1B-7B โ€ข 7B-30B โ€ข 30B+


๐Ÿ“ˆ Research Trends Dashboard

%%{init: {'theme':'dark'}}%%
graph TD
    A[2024 Q4] -->|RAG Revolution| B[2025 Q1]
    B -->|Agents & Tools| C[2025 Q2]
    B -->|1M+ Context| D[2025 Q2]
    B -->|MoE Everything| E[2025 Q2]
    
    style A fill:#1f2937
    style B fill:#374151
    style C fill:#4b5563
    style D fill:#6b7280
    style E fill:#9ca3af
Loading

This Month's Momentum:

  • ๐Ÿ“ˆ Rising: RAG Systems (+450%), Autonomous Agents (+320%), MoE Models (+280%)
  • ๐Ÿ“‰ Cooling: Basic Prompting (-60%), Small Models (-40%)
  • ๐Ÿ”ฎ Next Wave: Self-improving models, Multimodal reasoning, Edge deployment
  • Current Hot Topics: ๐Ÿ”ฅ Long Context (>1M tokens) | ๐Ÿ”ฅ Reasoning without CoT | ๐Ÿ”ฅ Efficient Fine-tuning

๐Ÿค Contributing

Add a paper in 30 seconds!

Submit Paper

Just need: Paper link + One sentence on why it matters

โœ… Reviewed within 24 hours | ๐Ÿ† Contributors get credit | ๐Ÿ’ฌ Join discussions


โฌ† Back to Top

Made with โค๏ธ for the AI Research Community

Last updated: August 10, 2025, 9:00 AM IST

Follow on Twitter GitHub followers

About

Your one-stop destination for the latest LLM research papers and breakthroughs

Topics

Resources

License

Contributing

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages