Hacker News
23.8K subscribers
117K links
Top stories from https://news.ycombinator.com (with 100+ score)
Contribute to the development here: https://github.com/phil-r/hackernewsbot
Also check https://t.me/designer_news

Contacts: @philr
Download Telegram
Show HN: I open-sourced my AI toy company that runs on ESP32 and OpenAI realtime (Score: 150+ in 23 hours)

Link: https://readhacker.news/s/6tcRB
Comments: https://readhacker.news/c/6tcRB

Hi HN! Last year the project I launched here got a lot of good feedback on creating speech to speech AI on the ESP32. Recently I revamped the whole stack, iterated on that feedback and made our project fully open-source—all of the client, hardware, firmware code.
This Github repo turns an ESP32-S3 into a realtime AI speech companion using the OpenAI Realtime API, Arduino WebSockets, Deno Edge Functions, and a full-stack web interface. You can talk to your own custom AI character, and it responds instantly.
I couldn't find a resource that helped set up a reliable, secure websocket (WSS) AI speech to speech service. While there are several useful Text-To-Speech (TTS) and Speech-To-Text (STT) repos out there, I believe none gets Speech-To-Speech right. OpenAI launched an embedded-repo late last year which sets up WebRTC with ESP-IDF. However, it's not beginner friendly and doesn't have a server side component for business logic.
This repo is an attempt at solving the above pains and creating a great speech to speech experience on Arduino with Secure Websockets using Edge Servers (with Deno/Supabase Edge Functions) for fast global connectivity and low latency.
Show HN: Node.js video tutorials where you can edit and run the code (Score: 150+ in 7 hours)

Link: https://readhacker.news/c/6tfHx

Hey HN,
I'm Sindre, CTO of Scrimba (YC S20). We originally launched Scrimba to make video learning more interactive for aspiring frontend developers. So instead of passively watching videos, you can jump in an experiment with the code directly inside the video player. Since launch, almost two million people have used Scrimba to grow their skills.
However, one limitation is that we've only supported frontend code, as our interactive videos run in the browser, whereas most of our learners want to go fullstack—building APIs, handling auth, working with databases, and so forth.
To fix this, we spent the last 6 months integrating StackBlitz WebContainers into Scrimba. This enables a full Node.js environment—including a terminal, shell, npm access, and a virtual file system—directly inside our video player. Everything runs in the browser.
Here is a 2-minute recorded demo: https://scrimba.com/s08dpq3nom
If you want to see more, feel free to enroll into any of the seven fullstack courses we've launched so far, on subject like Node, Next, Express, SQL, Vite, and more. We've opened them up for Hacker News today so that you don't even need to create an account to watch the content:
https://scrimba.com/fullstack
Other notable highlights about our "IDE videos":
- Based on events (code edits, cursor moves, etc) instead of pixels
- Roughly 100x smaller than traditional videos
- Recording is simple: just talk while you code
- Can be embedded in blogs, docs, or courses, like MDN does here: https://developer.mozilla.org/en-US/curriculum/core/css-fund...
- Entirely built in Imba, a language I created myself: https://news.ycombinator.com/item?id=28207662
We think this format could be useful for open-source maintainers and API-focused teams looking to create interactive docs or walkthroughs. Our videos are already embedded by MDN, LangChain, and Coursera.
If you maintain a library or SDK and want an interactive video about it, let us know—happy to record one for free that you can use however you like.
Would love to answer any questions or hear people's feedback!
They made computers behave like annoying salesmen (Score: 153+ in 4 hours)

Link: https://readhacker.news/s/6tgtq
Comments: https://readhacker.news/c/6tgtq
How to quickly charge your smartphone: fast charging technologies in detail (Score: 150+ in 1 day)

Link: https://readhacker.news/s/6teeJ
Comments: https://readhacker.news/c/6teeJ
Shortest-possible walking tour to 81,998 bars in South Korea (Score: 150+ in 4 hours)

Link: https://readhacker.news/s/6thRT
Comments: https://readhacker.news/c/6thRT
Show HN: Rowboat – Open-source IDE for multi-agent systems (Score: 150+ in 1 day)

Link: https://readhacker.news/s/6tdmr
Comments: https://readhacker.news/c/6tdmr

Hi HN! We’re Arjun, Ramnique, and Akhilesh, and we are building Rowboat (https://www.rowboatlabs.com/), an AI-assisted IDE for building and managing multi-agent systems. You start with a single agent, then scale up to teams of agents that work together, use MCP tools, and improve over time - all through a chat-based copilot.
Our repo is https://github.com/rowboatlabs/rowboat, docs are at https://docs.rowboatlabs.com/, and there’s a demo video here: https://youtu.be/YRTCw9UHRbU
It’s becoming clear that real-world agentic systems work best when multiple agents collaborate, rather than having one agent attempt to do everything. This isn’t too surprising - it’s a bit like how good code consists of multiple functions that each do one thing, rather than cramming everything into one function.
For example, a travel assistant works best when different agents handle specialized tasks: one agent finds the best flights, another optimizes hotel selections, and a third organizes the itinerary. This modular approach makes the system easier to manage, debug, and improve over time.
OpenAI’s Agents SDK provides a neat Python library to support this, but building reliable agentic systems requires constant iterations and tweaking - e.g. updating agent instructions (which can quickly get as complex as actual code), connecting tools, and testing the system and incorporating feedback. Rowboat is an AI IDE to do all this. Rowboat is to AI agents what Cursor is to code.
We’ve taken a code-like approach to agent instructions (prompts). There are special keywords to directly reference other agents, tools or prompts - which are highlighted in the UI. The copilot is the best way to create and edit these instructions - each change comes with a code-style diff.
You can give agents access to tools by integrating any MCP server or connecting your own functions through a webhook. You can instruct the agents on when to use specific tools via ‘@mentions’ in the agent instruction. To enable quick testing, we added a way to mock tool responses using LLM calls.
Rowboat playground lets you test and debug the assistants as you build them. You can see agent transfers, tool invocations and tool responses in real-time. The copilot has the context of the chat, and can improve the agent instructions based on feedback. For example, you could say ‘The agent shouldn’t have done x here. Fix this’ and the copilot can go and make this fix.
You can integrate agentic systems built in Rowboat into your application via the HTTP API or the Python SDK (‘pip install rowboat’). For example, you can build user-facing chatbots, enterprise workflows and employee assistants using Rowboat.
We’ve been working with LLMs since GPT-1 launched in 2018. Most recently, we built Coinbase’s support chatbot after our last AI startup was acquired by them.
Rowboat is Apache 2.0 licensed, giving you full freedom to self-host, modify, or extend it however you like.
We’re excited to share Rowboat with everyone here. We’d love to hear your thoughts!
YAGRI: You are gonna read it (Score: 151+ in 8 hours)

Link: https://readhacker.news/s/6thvz
Comments: https://readhacker.news/c/6thvz
Show HN: My from-scratch OS kernel that runs DOOM (Score: 150+ in 6 hours)

Link: https://readhacker.news/s/6thRt
Comments: https://readhacker.news/c/6thRt

Hi there! I've been on-and-off working on TacOS for a few months, which follows some UNIX-derived concepts (exec/fork, unix-style VFS, etc) and is now able to run a port of Doom, with a fairly small amount of modifications, using my from-scratch libc. The performance is actually decent compared to what I expected. Very interested to hear your thoughts. Thank you!