living in uncertainty
each of us is wrong sometimes. when a prediction you made turns out to be false, it's a good exercise to back-propagate it to your world model (what did i miss?) and then do a forward pass to understand what other implications this discovery has (what else must be true that i thought is false). usually it's not a big deal. often it is some wrong assumption/bias which impacts few things. you update your world model and move on
the ai of today was in the realm of fantasy a few years ago. i tried to find what we missed, but all plausible theories explaining llms suggest that we missed something very fundamental, and which has a lot of profound implications. among all world model updates i consider, the delta is large (the rabbit hole is deep), but also i'm not certain which one is the most accurate. just as an example, i am not aware of a theory that would explain why llms are getting good at coding but wont explain why they will eventually become good at everything else.
moreover it is dynamic. last year the talk was about prompt engineering. this year it's agents. i expect it to be different next year, and in general i expect things to continue to evolve. more capabilities will be developed, more work automated and more population waking up
it can be challenging to live like that: we like certainty. you might be used to have some life plan that you follow (say you are a prospective student who chooses your major). any plan has to assume that things that were true during planning will stay true. ok, maybe instead of a static plan you have a more dynamic strategy, essentially a function that accepts the current state and returns suggested actions. well, that function is still compressed knowledge of what works and what doesn't, but that knowledge might also get stale as the world changes
certainty is a luxury. i expect it to continue decreasing. i'm not even sure it will be coming back in our lifetimes, so I'd suggest to start getting used to a world where you can't be very confident in the increasing number of things (this is particularly hard for very smart people who are right most of the time and got used to that)
the only advice i have is gain some humility, be open minded and get ready to constantly adapt to the world changing under our feet
each of us is wrong sometimes. when a prediction you made turns out to be false, it's a good exercise to back-propagate it to your world model (what did i miss?) and then do a forward pass to understand what other implications this discovery has (what else must be true that i thought is false). usually it's not a big deal. often it is some wrong assumption/bias which impacts few things. you update your world model and move on
the ai of today was in the realm of fantasy a few years ago. i tried to find what we missed, but all plausible theories explaining llms suggest that we missed something very fundamental, and which has a lot of profound implications. among all world model updates i consider, the delta is large (the rabbit hole is deep), but also i'm not certain which one is the most accurate. just as an example, i am not aware of a theory that would explain why llms are getting good at coding but wont explain why they will eventually become good at everything else.
moreover it is dynamic. last year the talk was about prompt engineering. this year it's agents. i expect it to be different next year, and in general i expect things to continue to evolve. more capabilities will be developed, more work automated and more population waking up
it can be challenging to live like that: we like certainty. you might be used to have some life plan that you follow (say you are a prospective student who chooses your major). any plan has to assume that things that were true during planning will stay true. ok, maybe instead of a static plan you have a more dynamic strategy, essentially a function that accepts the current state and returns suggested actions. well, that function is still compressed knowledge of what works and what doesn't, but that knowledge might also get stale as the world changes
certainty is a luxury. i expect it to continue decreasing. i'm not even sure it will be coming back in our lifetimes, so I'd suggest to start getting used to a world where you can't be very confident in the increasing number of things (this is particularly hard for very smart people who are right most of the time and got used to that)
the only advice i have is gain some humility, be open minded and get ready to constantly adapt to the world changing under our feet
π26β€βπ₯11π₯6β€2π2π1π³1
I appreciate being mentioned in top-30 AI people in Uzbekistan. Thanks.
https://yuksalish.org/uz/news_detail/835
https://yuksalish.org/uz/news_detail/835
π₯55π14β€8π5π€ͺ1
don't let AIs help you spiral into your craziness. if you are vulnerable to that, don't use sycophantic AIs
https://thezvi.wordpress.com/2025/09/16/ai-craziness-notes/
https://thezvi.wordpress.com/2025/09/16/ai-craziness-notes/
π11π»3
Anthropic's first postmortem about the model being dumber, with a level of detail that we usually don't share
https://www.anthropic.com/engineering/a-postmortem-of-three-recent-issues
https://www.anthropic.com/engineering/a-postmortem-of-three-recent-issues
Anthropic
A postmortem of three recent issues
This is a technical report on three bugs that intermittently degraded responses from Claude. Below we explain what happened, why it took time to fix, and what we're changing.
π₯16π’2
what an idiot. i mean, the fact that he is an idiot is not new, but this is the new level
https://x.com/whitehouse/status/1969147079478989220?s=46
https://x.com/whitehouse/status/1969147079478989220?s=46
X (formerly Twitter)
The White House (@WhiteHouse) on X
π° NEW @Bloomberg: Trump to Add New $100,000 Fee for H-1B Visas in Latest Crackdown.
π±11π―8π€―7π2π’2β€1π1
OpenAI released a new eval that measures performance on economically valuable, real-world tasks across 44 occupations.
https://openai.com/index/gdpval/
https://openai.com/index/gdpval/
π₯18β€3π3
great example of why i didn't even consider applying for openai. think of implications in a country with a racist president
https://x.com/gabrielpeterss4/status/1973120058907041902?s=46
https://x.com/gabrielpeterss4/status/1973120058907041902?s=46
X (formerly Twitter)
gabriel (@gabriel1) on X
i have the most liked video on sora 2 right now, i will be enjoying this short moment while it lasts
cctv footage of sam stealing gpus at target for sora inference
cctv footage of sam stealing gpus at target for sora inference
π21β€6π1π1π1
You can now connect Slack to Claude. It can search your workspace channels, DMs, and files/gdocs to provide context for deep work.
You can also connect Claude app to slack, e.g. ask something in the app and claude can read your slack, search info there, etc.
Video below
https://x.com/claudeai/status/1973445694305468597?s=46
You can also connect Claude app to slack, e.g. ask something in the app and claude can read your slack, search info there, etc.
Video below
https://x.com/claudeai/status/1973445694305468597?s=46
X (formerly Twitter)
Claude (@claudeai) on X
Claude is now available in Slack.
Chat with Claude through DMs, tag @.Claude in threads, or use the AI assistant panelβwith access to web search, document analysis, and your connected tools.
Chat with Claude through DMs, tag @.Claude in threads, or use the AI assistant panelβwith access to web search, document analysis, and your connected tools.
π9π₯8β€5
π¦ i recommend spending a year with Rust
i don't think i can explain all the reasons why do that in a way that's both short and clear. most likely i'll lose the reader in the middle of the post before i'd get to the point. it is only after some first-hand prolonged experience of learning the Rust way you start getting it.
just trust me on this π go ahead and do yourself a favor
fair warning: first 6mo can be painful, but we have LLMs now that help a lot
i don't think i can explain all the reasons why do that in a way that's both short and clear. most likely i'll lose the reader in the middle of the post before i'd get to the point. it is only after some first-hand prolonged experience of learning the Rust way you start getting it.
just trust me on this π go ahead and do yourself a favor
fair warning: first 6mo can be painful, but we have LLMs now that help a lot
π«‘41β€11π4π₯4π4π2π2
haiku 4.5 (just released) is as smart as sonnet 4.0, but it's 2x faster and 3x cheaper. i've been using it in claude code for a while (primarily because of speed) and i can recommend it. i use it more often than sonnet 4.5 and definitely more than opus
https://www.anthropic.com/news/claude-haiku-4-5
https://www.anthropic.com/news/claude-haiku-4-5
π23β€9π₯8
Addressing seemingly common misunderstanding.
- Sonnet 4.5 is smarter than Opus 4.1.
- Haiku 4.5 nearly as smart than Sonnet 4.0
how come? Scaling laws suggest that the intelligence of models grows with scale (aka the bitter lesson). We increase training scale all the time, so it is not surprising that a newer model is more intelligent than an older model.
Besides, smaller models are:
- much faster, so you are getting more done
- cheaper, so your quota lasts longer
- Sonnet 4.5 is smarter than Opus 4.1.
- Haiku 4.5 nearly as smart than Sonnet 4.0
how come? Scaling laws suggest that the intelligence of models grows with scale (aka the bitter lesson). We increase training scale all the time, so it is not surprising that a newer model is more intelligent than an older model.
Besides, smaller models are:
- much faster, so you are getting more done
- cheaper, so your quota lasts longer
π₯18π7β€2πΎ2
i started feeling the agi with this model
X (formerly Twitter)
Claude (@claudeai) on X
Introducing Claude Opus 4.5: the best model in the world for coding, agents, and computer use.
Opus 4.5 is a step forward in what AI systems can do, and a preview of larger changes to how work gets done.
Opus 4.5 is a step forward in what AI systems can do, and a preview of larger changes to how work gets done.
π22β€5
if you are building frontend, enable the frontend plugin
https://x.com/trq212/status/1993786552233939042?s=46
https://x.com/trq212/status/1993786552233939042?s=46
X (formerly Twitter)
Thariq (@trq212) on X
To try this yourself add our marketplace in Claude Code:
/plugin marketplace add anthropics/claude-code
and then install the plugin:
/plugin install frontend-design@claude-code-plugins
/plugin marketplace add anthropics/claude-code
and then install the plugin:
/plugin install frontend-design@claude-code-plugins
π11
This media is not supported in your browser
VIEW IN TELEGRAM
π€£18β€6π4
RIP coding
I started coding approx 26-28 years ago. There were many months that i wrote code every day. It was my main hobby. I no longer write code and I don't think I will.
I still produce a lot of code, but i don't type it myself. I mostly direct agent(s) and review their code
It was fun π«π«‘
I started coding approx 26-28 years ago. There were many months that i wrote code every day. It was my main hobby. I no longer write code and I don't think I will.
I still produce a lot of code, but i don't type it myself. I mostly direct agent(s) and review their code
It was fun π«π«‘
π71π13πΎ13β€7π€‘5π₯4