oguzhanercan 's Collections Large Language Models
updated
Treasure Hunt: Real-time Targeting of the Long Tail using Training-Time
Markers
Paper
• 2506.14702
• Published
• 3
MiniMax-M1: Scaling Test-Time Compute Efficiently with Lightning
Attention
Paper
• 2506.13585
• Published
• 273
Scaling Test-time Compute for LLM Agents
Paper
• 2506.12928
• Published
• 63
A Survey on Latent Reasoning
Paper
• 2507.06203
• Published
• 93
Pre-Trained Policy Discriminators are General Reward Models
Paper
• 2507.05197
• Published
• 39
First Return, Entropy-Eliciting Explore
Paper
• 2507.07017
• Published
• 24
Jan-nano Technical Report
Paper
• 2506.22760
• Published
• 9
Test-Time Scaling with Reflective Generative Model
Paper
• 2507.01951
• Published
• 108
Geometric-Mean Policy Optimization
Paper
• 2507.20673
• Published
• 32
Paper
• 2507.22879
• Published
• 38
CLEAR: Error Analysis via LLM-as-a-Judge Made Easy
Paper
• 2507.18392
• Published
• 20
SAND-Math: Using LLMs to Generate Novel, Difficult and Useful
Mathematics Questions and Answers
Paper
• 2507.20527
• Published
• 7
A Survey of Self-Evolving Agents: On Path to Artificial Super
Intelligence
Paper
• 2507.21046
• Published
• 84
UltraMemV2: Memory Networks Scaling to 120B Parameters with Superior
Long-Context Learning
Paper
• 2508.18756
• Published
• 36
LLaMAX2: Your Translation-Enhanced Model also Performs Well in Reasoning
Paper
• 2510.09189
• Published
• 5
Cache-to-Cache: Direct Semantic Communication Between Large Language
Models
Paper
• 2510.03215
• Published
• 98
Scaling Latent Reasoning via Looped Language Models
Paper
• 2510.25741
• Published
• 229
The End of Manual Decoding: Towards Truly End-to-End Language Models
Paper
• 2510.26697
• Published
• 117
Kimi Linear: An Expressive, Efficient Attention Architecture
Paper
• 2510.26692
• Published
• 127
Latent Collaboration in Multi-Agent Systems
Paper
• 2511.20639
• Published
• 121
Recursive Language Models
Paper
• 2512.24601
• Published
• 89
Dynamic Large Concept Models: Latent Reasoning in an Adaptive Semantic Space
Paper
• 2512.24617
• Published
• 65