Are polynomial features the root of all evil? (2024) (Score: 151+ in 11 hours)
Link: https://readhacker.news/s/6tdpP
Comments: https://readhacker.news/c/6tdpP
Link: https://readhacker.news/s/6tdpP
Comments: https://readhacker.news/c/6tdpP
Alex Shtoff
“Are polynomial features the root of all evil?”
There is a well-known myth in the machine learning community - high degree polynomials are bad for modeling. In this post we debunk this myth.
How long does it take to create a new habit? (2015) (Score: 151+ in 10 hours)
Link: https://readhacker.news/s/6tdHn
Comments: https://readhacker.news/c/6tdHn
Link: https://readhacker.news/s/6tdHn
Comments: https://readhacker.news/c/6tdHn
The Logical Optimist - Ottawa Life Coach
The Logical Optimist - how long it takes to create a new habit
How long it really takes to create a new habit, not 21 days.
Verus: Verified Rust for low-level systems code (❄️ Score: 150+ in 2 days)
Link: https://readhacker.news/s/6t7Cm
Comments: https://readhacker.news/c/6t7Cm
Link: https://readhacker.news/s/6t7Cm
Comments: https://readhacker.news/c/6t7Cm
GitHub
GitHub - verus-lang/verus: Verified Rust for low-level systems code
Verified Rust for low-level systems code. Contribute to verus-lang/verus development by creating an account on GitHub.
Show HN: Morphik – Open-source RAG that understands PDF images, runs locally (Score: 151+ in 14 hours)
Link: https://readhacker.news/s/6tdiG
Comments: https://readhacker.news/c/6tdiG
Hey HN, we’re Adi and Arnav. A few months ago, we hit a wall trying to get LLMs to answer questions over research papers and instruction manuals. Everything worked fine, until the answer lived inside an image or diagram embedded in the PDF. Even GPT‑4o flubbed it (we recently tried O3 with the same, and surprisingly it flubbed it too). Naive RAG pipelines just pulled in some text chunks and ignored the rest.
We took an invention disclosure PDF (https://drive.google.com/file/d/1ySzQgbNZkC5dPLtE3pnnVL2rW_9...) containing an IRR‑vs‑frequency graph and asked GPT “From the graph, at what frequency is the IRR maximized?”. We originally tried this on gpt-4o, but while writing this used the new natively multimodal model o4‑mini‑high. After a 30‑second thinking pause, it asked for clarifications, then churned out buggy code, pulled data from the wrong page, and still couldn’t answer the question. We wrote up the full story with screenshots here: https://docs.morphik.ai/blogs/gpt-vs-morphik-multimodal.
We got frustrated enough to try fixing it ourselves.
We built Morphik to do multimodal retrieval over documents like PDFs, where images and diagrams matter as much as the text.
To do this, we use Colpali-style embeddings, which treat each document page as an image and generate multi-vector representations. These embeddings capture layout, typography, and visual context, allowing retrieval to get a whole table or schematic, not just nearby tokens. Along with vector search, this could now retrieve exact pages with relevant diagrams and pass them as images to the LLM to get relevant answers. It’s able to answer the question with an 8B llama 3.1 vision running locally!
Early pharma testers hit our system with queries like "Which EGFR inhibitors at 50 mg showed ≥ 30% tumor reduction?" We correctly returned the right tables and plots, but still hit a bottleneck, we weren’t able to join the dots across multiple reports. So we built a knowledge graph: we tag entities in both text and images, normalize synonyms (Erlotinib → EGFR inhibitor), infer relations (e.g. administered_at, yields_reduction), and stitch everything into a graph. Now a single query could traverse that graph across documents and surface a coherent, cross‑document answer along with the correct pages as images.
To illustrate that, and just for fun, we built a graph of 100 Paul Graham’s essays here: https://pggraph.streamlit.app/ You can search for various nodes, (eg. startup, sam altman, paul graham and see corresponding connections). In our system, we create graphs and store the relevant text chunks along with the entities, so on querying, we can extract the relevant entity, do a search on the graph and pull in the text chunks of all connected nodes, improving cross document queries.
For longer or multi-turn queries, we added persistent KV caching, which stores intermediate key-value states from transformer attention layers. Instead of recomputing attention from scratch every time, we reuse prior layers, speeding up repeated queries and letting us handle much longer context windows.
We’re open‑source under the MIT Expat license: https://github.com/morphik-org/morphik-core
Would love to hear your RAG horror stories, what worked, what didn’t and any feedback on Morphik. We’re here for it.
Link: https://readhacker.news/s/6tdiG
Comments: https://readhacker.news/c/6tdiG
Hey HN, we’re Adi and Arnav. A few months ago, we hit a wall trying to get LLMs to answer questions over research papers and instruction manuals. Everything worked fine, until the answer lived inside an image or diagram embedded in the PDF. Even GPT‑4o flubbed it (we recently tried O3 with the same, and surprisingly it flubbed it too). Naive RAG pipelines just pulled in some text chunks and ignored the rest.
We took an invention disclosure PDF (https://drive.google.com/file/d/1ySzQgbNZkC5dPLtE3pnnVL2rW_9...) containing an IRR‑vs‑frequency graph and asked GPT “From the graph, at what frequency is the IRR maximized?”. We originally tried this on gpt-4o, but while writing this used the new natively multimodal model o4‑mini‑high. After a 30‑second thinking pause, it asked for clarifications, then churned out buggy code, pulled data from the wrong page, and still couldn’t answer the question. We wrote up the full story with screenshots here: https://docs.morphik.ai/blogs/gpt-vs-morphik-multimodal.
We got frustrated enough to try fixing it ourselves.
We built Morphik to do multimodal retrieval over documents like PDFs, where images and diagrams matter as much as the text.
To do this, we use Colpali-style embeddings, which treat each document page as an image and generate multi-vector representations. These embeddings capture layout, typography, and visual context, allowing retrieval to get a whole table or schematic, not just nearby tokens. Along with vector search, this could now retrieve exact pages with relevant diagrams and pass them as images to the LLM to get relevant answers. It’s able to answer the question with an 8B llama 3.1 vision running locally!
Early pharma testers hit our system with queries like "Which EGFR inhibitors at 50 mg showed ≥ 30% tumor reduction?" We correctly returned the right tables and plots, but still hit a bottleneck, we weren’t able to join the dots across multiple reports. So we built a knowledge graph: we tag entities in both text and images, normalize synonyms (Erlotinib → EGFR inhibitor), infer relations (e.g. administered_at, yields_reduction), and stitch everything into a graph. Now a single query could traverse that graph across documents and surface a coherent, cross‑document answer along with the correct pages as images.
To illustrate that, and just for fun, we built a graph of 100 Paul Graham’s essays here: https://pggraph.streamlit.app/ You can search for various nodes, (eg. startup, sam altman, paul graham and see corresponding connections). In our system, we create graphs and store the relevant text chunks along with the entities, so on querying, we can extract the relevant entity, do a search on the graph and pull in the text chunks of all connected nodes, improving cross document queries.
For longer or multi-turn queries, we added persistent KV caching, which stores intermediate key-value states from transformer attention layers. Instead of recomputing attention from scratch every time, we reuse prior layers, speeding up repeated queries and letting us handle much longer context windows.
We’re open‑source under the MIT Expat license: https://github.com/morphik-org/morphik-core
Would love to hear your RAG horror stories, what worked, what didn’t and any feedback on Morphik. We’re here for it.
GitHub
GitHub - morphik-org/morphik-core: Open source multi-modal RAG for building AI apps over private knowledge.
Open source multi-modal RAG for building AI apps over private knowledge. - GitHub - morphik-org/morphik-core: Open source multi-modal RAG for building AI apps over private knowledge.
Solidjs: Simple and performant reactivity for building user interfaces (❄️ Score: 150+ in 3 days)
Link: https://readhacker.news/s/6t46z
Comments: https://readhacker.news/c/6t46z
Link: https://readhacker.news/s/6t46z
Comments: https://readhacker.news/c/6t46z
Solidjs
Solid is a purely reactive library. It was designed from the ground up with a reactive core. It's influenced by reactive principles developed by previous libraries.
Native visionOS platform support (🔥 Score: 153+ in 3 hours)
Link: https://readhacker.news/s/6teLX
Comments: https://readhacker.news/c/6teLX
Link: https://readhacker.news/s/6teLX
Comments: https://readhacker.news/c/6teLX
GitHub
Native visionOS platform support by rsanchezsaez · Pull Request #105628 · godotengine/godot
Dear Godot community,
I'm on Apple's visionOS engineering team, and we would like to contribute Vision Pro support to the Godot engine. This is the first PR that lays the foundation...
I'm on Apple's visionOS engineering team, and we would like to contribute Vision Pro support to the Godot engine. This is the first PR that lays the foundation...
π0.5: A VLA with open-world generalization (Score: 153+ in 14 hours)
Link: https://readhacker.news/s/6tdvR
Comments: https://readhacker.news/c/6tdvR
Link: https://readhacker.news/s/6tdvR
Comments: https://readhacker.news/c/6tdvR
www.pi.website
A VLA with Open-World Generalization
Our latest generalist policy, π0.5, extends π0 and enables open-world generalization. Our new model can control a mobile manipulator to clean up an entirely new kitchen or bedroom.
Open Source Projects Receive Funding to Reclaim the Public Internet (🔥 Score: 156+ in 2 hours)
Link: https://readhacker.news/s/6tf7U
Comments: https://readhacker.news/c/6tf7U
Link: https://readhacker.news/s/6tf7U
Comments: https://readhacker.news/c/6tf7U
Advanced Python Features (Score: 150+ in 4 hours)
Link: https://readhacker.news/s/6tf7Y
Comments: https://readhacker.news/c/6tf7Y
Link: https://readhacker.news/s/6tf7Y
Comments: https://readhacker.news/c/6tf7Y
Edward Li's Blog
14 Advanced Python Features
Python is one of the most widely adopted programming languages in the world. Yet, because of it’s ease and simplicity to just “get something working”, it’s also one of the most underappreciated.
If you search for Top 10 Advanced Python Tricks on Google or…
If you search for Top 10 Advanced Python Tricks on Google or…
CSS Hell (Score: 150+ in 13 hours)
Link: https://readhacker.news/s/6teev
Comments: https://readhacker.news/c/6teev
Link: https://readhacker.news/s/6teev
Comments: https://readhacker.news/c/6teev
Csshell
CSS Hell
Web site created using create-react-app
The raccoons who made computer magazine ads great (Score: 151+ in 22 hours)
Link: https://readhacker.news/s/6tcBK
Comments: https://readhacker.news/c/6tcBK
Link: https://readhacker.news/s/6tcBK
Comments: https://readhacker.news/c/6tcBK
Technologizer by Harry McCracken
The raccoons who made computer magazine ads great
In the 1980s and 1990s, PC Connection built its brand on a campaign starring folksy small-town critters. They’ll still charm your socks off.
Ping, You've Got Whale: AI detection system alerts ships of whales in their path (Score: 150+ in 18 hours)
Link: https://readhacker.news/s/6tdEm
Comments: https://readhacker.news/c/6tdEm
Link: https://readhacker.news/s/6tdEm
Comments: https://readhacker.news/c/6tdEm
bioGraphic
Ping, You've Got Whale - bioGraphic
A new artificial intelligence-powered detection system is giving ship captains real-time alerts when a whale is in their path.
America's cyber defenses are being dismantled from the inside (🔥 Score: 154+ in 2 hours)
Link: https://readhacker.news/s/6tfpY
Comments: https://readhacker.news/c/6tfpY
Link: https://readhacker.news/s/6tfpY
Comments: https://readhacker.news/c/6tfpY
The Register
America's cyber defenses are being dismantled from the inside
Opinion: The CVE system nearly dying shows that someone has lost the plot
Show HN: I open-sourced my AI toy company that runs on ESP32 and OpenAI realtime (Score: 150+ in 23 hours)
Link: https://readhacker.news/s/6tcRB
Comments: https://readhacker.news/c/6tcRB
Hi HN! Last year the project I launched here got a lot of good feedback on creating speech to speech AI on the ESP32. Recently I revamped the whole stack, iterated on that feedback and made our project fully open-source—all of the client, hardware, firmware code.
This Github repo turns an ESP32-S3 into a realtime AI speech companion using the OpenAI Realtime API, Arduino WebSockets, Deno Edge Functions, and a full-stack web interface. You can talk to your own custom AI character, and it responds instantly.
I couldn't find a resource that helped set up a reliable, secure websocket (WSS) AI speech to speech service. While there are several useful Text-To-Speech (TTS) and Speech-To-Text (STT) repos out there, I believe none gets Speech-To-Speech right. OpenAI launched an embedded-repo late last year which sets up WebRTC with ESP-IDF. However, it's not beginner friendly and doesn't have a server side component for business logic.
This repo is an attempt at solving the above pains and creating a great speech to speech experience on Arduino with Secure Websockets using Edge Servers (with Deno/Supabase Edge Functions) for fast global connectivity and low latency.
Link: https://readhacker.news/s/6tcRB
Comments: https://readhacker.news/c/6tcRB
Hi HN! Last year the project I launched here got a lot of good feedback on creating speech to speech AI on the ESP32. Recently I revamped the whole stack, iterated on that feedback and made our project fully open-source—all of the client, hardware, firmware code.
This Github repo turns an ESP32-S3 into a realtime AI speech companion using the OpenAI Realtime API, Arduino WebSockets, Deno Edge Functions, and a full-stack web interface. You can talk to your own custom AI character, and it responds instantly.
I couldn't find a resource that helped set up a reliable, secure websocket (WSS) AI speech to speech service. While there are several useful Text-To-Speech (TTS) and Speech-To-Text (STT) repos out there, I believe none gets Speech-To-Speech right. OpenAI launched an embedded-repo late last year which sets up WebRTC with ESP-IDF. However, it's not beginner friendly and doesn't have a server side component for business logic.
This repo is an attempt at solving the above pains and creating a great speech to speech experience on Arduino with Secure Websockets using Edge Servers (with Deno/Supabase Edge Functions) for fast global connectivity and low latency.
GitHub
GitHub - akdeb/ElatoAI: Realtime AI speech with OpenAI Realtime API on Arduino ESP32 with Secure Websockets and Deno edge functions…
Realtime AI speech with OpenAI Realtime API on Arduino ESP32 with Secure Websockets and Deno edge functions with >10min uninterrupted conversations globally for AI toys, AI companions, AI de...
The Gruen Transfer is consuming the internet (Score: 152+ in 5 hours)
Link: https://readhacker.news/s/6tfg2
Comments: https://readhacker.news/c/6tfg2
Link: https://readhacker.news/s/6tfg2
Comments: https://readhacker.news/c/6tfg2
sebs.website
The Gruen Transfer is consuming the internet
<p>The Gruen Transfer is taking over social media. What is it and how can we avoid it?</p>
Apple and Meta fined millions for breaching EU law (Score: 151+ in 4 hours)
Link: https://readhacker.news/s/6tfpb
Comments: https://readhacker.news/c/6tfpb
Link: https://readhacker.news/s/6tfpb
Comments: https://readhacker.news/c/6tfpb
Yahoo Finance
Apple, Meta fined as EU presses ahead with tech probes
BRUSSELS (Reuters) -Apple was fined 500 million euros ($570 million) on Wednesday and Meta 200 million euros, as European Union antitrust regulators handed out the first sanctions under landmark legislation aimed at curbing the power of Big Tech. The EU…
MinC Is Not Cygwin (Score: 150+ in 4 hours)
Link: https://readhacker.news/s/6tfr7
Comments: https://readhacker.news/c/6tfr7
Link: https://readhacker.news/s/6tfr7
Comments: https://readhacker.news/c/6tfr7
minc.commandlinerevolution.nl
MinC Is Not Cygwin
How a 20 year old bug in GTA San Andreas surfaced in Windows 11 24H2 (🔥 Score: 174+ in 1 hour)
Link: https://readhacker.news/s/6tg2r
Comments: https://readhacker.news/c/6tg2r
Link: https://readhacker.news/s/6tg2r
Comments: https://readhacker.news/c/6tg2r
Silent’s Blog
How a 20 year old bug in GTA San Andreas surfaced in Windows 11 24H2
After over two decades, players are now forbidden from flying a seaplane, all thanks to undefined code behavior.
How I Blog with Obsidian, Hugo, GitHub, and Cloudflare – Zero Cost, Fully Owned (Score: 150+ in 4 hours)
Link: https://readhacker.news/s/6tfNx
Comments: https://readhacker.news/c/6tfNx
Link: https://readhacker.news/s/6tfNx
Comments: https://readhacker.news/c/6tfNx
ingau
How I write my blogs in Obsidian and publish instantly
I’ve been using Obsidian for all my writing lately, and it’s been a game changer. The local-first model means everything lives as plain text on my machine, and with the Minimal theme, the interface stays clean and distraction-free.
My vault lives in iCloud…
My vault lives in iCloud…
AI Horseless Carriages (🔥 Score: 161+ in 2 hours)
Link: https://readhacker.news/s/6tgvf
Comments: https://readhacker.news/c/6tgvf
Link: https://readhacker.news/s/6tgvf
Comments: https://readhacker.news/c/6tgvf
koomen.dev
AI Horseless Carriages | koomen.dev
An essay about bad AI app design