Asankhaya Sharma's picture

Asankhaya Sharma PRO

codelion

AI & ML interests

AI/ML, Dev Tools and Application Security

Recent Activity

Organizations

meraGPT's profile picture Lambda Security's profile picture National University of Singapore's profile picture Patched's profile picture ZeroGPU Explorers's profile picture MLX Community's profile picture Social Post Explorers's profile picture Hugging Face Discord Community's profile picture Adaptive Classifier's profile picture Reasoning datasets competition 's profile picture Cerebras Hugging Face Hackathon's profile picture Agents-MCP-Hackathon's profile picture

codelion's activity

reacted to their post with โž•โค๏ธ๐Ÿš€๐Ÿ”ฅ 4 days ago
view post
Post
3311
๐Ÿง  We just implemented Andrej Karpathy's "third paradigm" for LLM learning!

System Prompt Learning (SPL) enables LLMs to automatically learn problem-solving strategies from experience, rather than relying on static prompts.

๐Ÿš€ How it works:
Your LLM builds a database of effective strategies, selects the best ones for each problem, and refines them over time based on success rates.

๐Ÿ“Š Results across math benchmarks:
Arena Hard: 29% โ†’ 37.6% (+8.6%)
AIME24: 23.33% โ†’ 30% (+6.67%)
OptILLMBench: 61% โ†’ 65% (+4%)

The best part? All strategies are human-readable and the system gets progressively better at problem types you use frequently.

โœจ Key benefits:
๐Ÿ”„ Cumulative learning over time
๐Ÿ“– Transparent, inspectable strategies
๐Ÿ”Œ Works with any OpenAI-compatible API
โšก Simple integration: just add "spl-" prefix to your model

Built as an open-source plugin in optillm. After 500 queries, our system developed 129 strategies and refined 97 of them!

This feels like a genuine step toward AI that learns from experience while staying completely interpretable.

๐Ÿ”— GitHub: https://github.com/codelion/optillm/tree/main/optillm/plugins/spl
๐Ÿ“– Full article: https://huggingface.co/blog/codelion/system-prompt-learning
๐Ÿฆ Original Karpathy tweet: https://x.com/karpathy/status/1921368644069765486

Have you experimented with advanced system prompting? What strategies would you want your LLM to learn?
posted an update 4 days ago
view post
Post
3311
๐Ÿง  We just implemented Andrej Karpathy's "third paradigm" for LLM learning!

System Prompt Learning (SPL) enables LLMs to automatically learn problem-solving strategies from experience, rather than relying on static prompts.

๐Ÿš€ How it works:
Your LLM builds a database of effective strategies, selects the best ones for each problem, and refines them over time based on success rates.

๐Ÿ“Š Results across math benchmarks:
Arena Hard: 29% โ†’ 37.6% (+8.6%)
AIME24: 23.33% โ†’ 30% (+6.67%)
OptILLMBench: 61% โ†’ 65% (+4%)

The best part? All strategies are human-readable and the system gets progressively better at problem types you use frequently.

โœจ Key benefits:
๐Ÿ”„ Cumulative learning over time
๐Ÿ“– Transparent, inspectable strategies
๐Ÿ”Œ Works with any OpenAI-compatible API
โšก Simple integration: just add "spl-" prefix to your model

Built as an open-source plugin in optillm. After 500 queries, our system developed 129 strategies and refined 97 of them!

This feels like a genuine step toward AI that learns from experience while staying completely interpretable.

๐Ÿ”— GitHub: https://github.com/codelion/optillm/tree/main/optillm/plugins/spl
๐Ÿ“– Full article: https://huggingface.co/blog/codelion/system-prompt-learning
๐Ÿฆ Original Karpathy tweet: https://x.com/karpathy/status/1921368644069765486

Have you experimented with advanced system prompting? What strategies would you want your LLM to learn?
upvoted an article 4 days ago
view article
Article

System Prompt Learning: Teaching LLMs to Learn Problem-Solving Strategies from Experience

By codelion โ€ข
โ€ข 10
published an article 4 days ago
view article
Article

System Prompt Learning: Teaching LLMs to Learn Problem-Solving Strategies from Experience

By codelion โ€ข
โ€ข 10
replied to their post 9 days ago
reacted to their post with โค๏ธ๐Ÿ‘€๐Ÿš€๐Ÿ”ฅ 10 days ago
view post
Post
2318
Introducing AutoThink: Adaptive reasoning for LLMs that improves performance by 43% on reasoning benchmarks!

Instead of using fixed thinking budgets, AutoThink:
- Classifies query complexity (HIGH/LOW) using adaptive classification
- Dynamically allocates thinking tokens based on complexity
- Uses steering vectors derived from Pivotal Token Search to guide reasoning patterns

Results on DeepSeek-R1-Distill-Qwen-1.5B:
- GPQA-Diamond: 31.06% vs 21.72% baseline (+9.34 points)
- MMLU-Pro: 26.38% vs 25.58% baseline (+0.8 points)
- Uses fewer tokens than baseline approaches

Works with any local reasoning model - DeepSeek, Qwen, Llama, custom models. The technique combines our research on Pivotal Token Search (PTS) implementation and adaptive classification frameworks.

Paper: AutoThink: efficient inference for reasoning LLMs
https://papers.ssrn.com/sol3/papers.cfm?abstract_id=5253327

Code and examples:
https://github.com/codelion/optillm/tree/main/optillm/autothink

PTS implementation and technical details:
https://github.com/codelion/pts
https://huggingface.co/blog/codelion/pts

Adaptive classifier framework:
https://github.com/codelion/adaptive-classifier

Would love to hear your thoughts on adaptive resource allocation for LLM reasoning! Have you experimented with similar approaches?
  • 5 replies
ยท