Hacker News
24.2K subscribers
118K links
Top stories from https://news.ycombinator.com (with 100+ score)
Contribute to the development here: https://github.com/phil-r/hackernewsbot
Also check https://t.me/designer_news

Contacts: @philr
Download Telegram
Show HN: Lazy Tetris (Score: 150+ in 8 hours)

Link: https://readhacker.news/s/6v9Jz
Comments: https://readhacker.news/c/6v9Jz

I made a tetris variant
Aims to remove all stress, and focus the game on what I like the best - stacking.
No timer, no score, no gravity. Move to the next piece when you are ready, and clear lines when you are ready.
Separate mobile + desktop controls
LumoSQL (Score: 151+ in 4 hours)

Link: https://readhacker.news/s/6vaim
Comments: https://readhacker.news/c/6vaim
Show HN: AutoThink – Boosts local LLM performance with adaptive reasoning (🔥 Score: 153+ in 3 hours)

Link: https://readhacker.news/c/6vcs8

I built AutoThink, a technique that makes local LLMs reason more efficiently by adaptively allocating computational resources based on query complexity.
The core idea: instead of giving every query the same "thinking time," classify queries as HIGH or LOW complexity and allocate thinking tokens accordingly. Complex reasoning gets 70-90% of tokens, simple queries get 20-40%.
I also implemented steering vectors derived from Pivotal Token Search (originally from Microsoft's Phi-4 paper) that guide the model's reasoning patterns during generation. These vectors encourage behaviors like numerical accuracy, self-correction, and thorough exploration.
Results on DeepSeek-R1-Distill-Qwen-1.5B:
- GPQA-Diamond: 31.06% vs 21.72% baseline (+43% relative improvement)
- MMLU-Pro: 26.38% vs 25.58% baseline
- Uses fewer tokens than baseline approaches
Works with any local reasoning model - DeepSeek, Qwen, custom fine-tuned models. No API dependencies.
The technique builds on two things I developed: an adaptive classification framework that can learn new complexity categories without retraining, and an open source implementation of Pivotal Token Search.
Technical paper: https://papers.ssrn.com/sol3/papers.cfm?abstract_id=5253327
Code and examples: https://github.com/codelion/optillm/tree/main/optillm/autoth...
PTS implementation: https://github.com/codelion/pts
I'm curious about your thoughts on adaptive resource allocation for AI reasoning. Have you tried similar approaches with your local models?