Collections
Discover the best community collections!
Collections including paper arxiv:2411.12580
-
Unleashing the Power of Pre-trained Language Models for Offline Reinforcement Learning
Paper • 2310.20587 • Published • 18 -
Chain-of-Thought Reasoning Without Prompting
Paper • 2402.10200 • Published • 110 -
LLM2LLM: Boosting LLMs with Novel Iterative Data Enhancement
Paper • 2403.15042 • Published • 28 -
LIMA: Less Is More for Alignment
Paper • 2305.11206 • Published • 26
-
TuCo: Measuring the Contribution of Fine-Tuning to Individual Responses of LLMs
Paper • 2506.23423 • Published • 1 -
Stochastic Parameter Decomposition
Paper • 2506.20790 • Published • 1 -
Decomposing MLP Activations into Interpretable Features via Semi-Nonnegative Matrix Factorization
Paper • 2506.10920 • Published • 6 -
From Flat to Hierarchical: Extracting Sparse Representations with Matching Pursuit
Paper • 2506.03093 • Published • 2
-
SELF: Language-Driven Self-Evolution for Large Language Model
Paper • 2310.00533 • Published • 2 -
GrowLength: Accelerating LLMs Pretraining by Progressively Growing Training Length
Paper • 2310.00576 • Published • 2 -
A Pretrainer's Guide to Training Data: Measuring the Effects of Data Age, Domain Coverage, Quality, & Toxicity
Paper • 2305.13169 • Published • 3 -
Transformers Can Achieve Length Generalization But Not Robustly
Paper • 2402.09371 • Published • 15
-
Same Task, More Tokens: the Impact of Input Length on the Reasoning Performance of Large Language Models
Paper • 2402.14848 • Published • 20 -
Teaching Large Language Models to Reason with Reinforcement Learning
Paper • 2403.04642 • Published • 51 -
How Far Are We from Intelligent Visual Deductive Reasoning?
Paper • 2403.04732 • Published • 24 -
Learning to Reason and Memorize with Self-Notes
Paper • 2305.00833 • Published • 5
-
SELF: Language-Driven Self-Evolution for Large Language Model
Paper • 2310.00533 • Published • 2 -
GrowLength: Accelerating LLMs Pretraining by Progressively Growing Training Length
Paper • 2310.00576 • Published • 2 -
A Pretrainer's Guide to Training Data: Measuring the Effects of Data Age, Domain Coverage, Quality, & Toxicity
Paper • 2305.13169 • Published • 3 -
Transformers Can Achieve Length Generalization But Not Robustly
Paper • 2402.09371 • Published • 15
-
Unleashing the Power of Pre-trained Language Models for Offline Reinforcement Learning
Paper • 2310.20587 • Published • 18 -
Chain-of-Thought Reasoning Without Prompting
Paper • 2402.10200 • Published • 110 -
LLM2LLM: Boosting LLMs with Novel Iterative Data Enhancement
Paper • 2403.15042 • Published • 28 -
LIMA: Less Is More for Alignment
Paper • 2305.11206 • Published • 26
-
Same Task, More Tokens: the Impact of Input Length on the Reasoning Performance of Large Language Models
Paper • 2402.14848 • Published • 20 -
Teaching Large Language Models to Reason with Reinforcement Learning
Paper • 2403.04642 • Published • 51 -
How Far Are We from Intelligent Visual Deductive Reasoning?
Paper • 2403.04732 • Published • 24 -
Learning to Reason and Memorize with Self-Notes
Paper • 2305.00833 • Published • 5
-
TuCo: Measuring the Contribution of Fine-Tuning to Individual Responses of LLMs
Paper • 2506.23423 • Published • 1 -
Stochastic Parameter Decomposition
Paper • 2506.20790 • Published • 1 -
Decomposing MLP Activations into Interpretable Features via Semi-Nonnegative Matrix Factorization
Paper • 2506.10920 • Published • 6 -
From Flat to Hierarchical: Extracting Sparse Representations with Matching Pursuit
Paper • 2506.03093 • Published • 2