xAI announces Grok 3. Here is everything you need to know
Elon mentioned that Grok 3 is an order of magnitude more capable than Grok 2.
Total GPUs: 200K
The capacity was doubled in 92 days!
All of this compute was used to improve Grok -- which has lead to Grok 3.
Grok 3 involved 10x more training than Grok 2!
Grok finished pretraining in early January!
The model is still training.
Here are the benchmark numbers:
Grok 3 significantly outperforms other models in its category such as Gemini 2 Pro and GPT-4o. Even Grok-3 mini shows to be competitive.
Results of early Grok 3 in the Chatbot Arena (LMSYS)
It reached an Elo score of 1400 which no other model has achieved.
The model score keeps improving.
Grok 3 also has reasoning capabilities too!
The Grok team has been testing these capabilities which they have unlocked using RL.
The model is good, especially in coding.
Grok 3 coding example:
Thinking traces as generated as the model tries to solve the problem.
Elon confirmed that the thinking steps have been obscured to avoid getting copied.
Grok 3 also excels at creative coding like generating creative and novel games.
Elon emphasized Grok 3's creative emergent capabilities.
You can also use the Big Brain mode to use more compute and reasoning with Grok 3.
Grok 3 Reasoning performance:
The results correspond to the beta version of Grok-3 Reasoning.
It outperforms o1 and DeepSeek-R1 when given more test-time compute (allowing it to think longer).
The Grok 3 mini reasoning model is also very capable.
Grok 3 Reasoning Beta performance on AIME 2025.
Grok 3 shows generalization capabilities.
It not only does coding and math problem-solving, but it can also do other creative and useful real-world tasks.
One of the results generated with Grok 3 mini.
Bejeweled Tetris generated by Grok 3.
Grok 3 cannot only unlock test-time compute, it also enables capable agents.
These capabilities have led to a new product called DeepSearch.
"Next generation of search agents to understand the universe".
More on DeepSearch:
- the model can think deeply about user intent
- what facts to consider
- how many websites to browse
- it can cross-validate different sources.
DeepSearch also exposes the steps that it takes to conduct the search itself.
Improvements will happen rapidly and almost daily according to the team.
There is also a Grok-powered voice app coming too -- about a week away!
Open-source approach:
The last version will be open-sourced when the most recent version is fully out.
After Grok 3 stable version is out, it is highly likely Grok 2 will be open-sourced. (within a few months).
SuperGrok dedicated app is also available with a polished experience.
Try on the web as well: grok.com
The web will include the latest Grok features.
Elon mentioned that Grok 3 is an order of magnitude more capable than Grok 2.
Total GPUs: 200K
The capacity was doubled in 92 days!
All of this compute was used to improve Grok -- which has lead to Grok 3.
Grok 3 involved 10x more training than Grok 2!
Grok finished pretraining in early January!
The model is still training.
Here are the benchmark numbers:
Grok 3 significantly outperforms other models in its category such as Gemini 2 Pro and GPT-4o. Even Grok-3 mini shows to be competitive.
Results of early Grok 3 in the Chatbot Arena (LMSYS)
It reached an Elo score of 1400 which no other model has achieved.
The model score keeps improving.
Grok 3 also has reasoning capabilities too!
The Grok team has been testing these capabilities which they have unlocked using RL.
The model is good, especially in coding.
Grok 3 coding example:
Thinking traces as generated as the model tries to solve the problem.
Elon confirmed that the thinking steps have been obscured to avoid getting copied.
Grok 3 also excels at creative coding like generating creative and novel games.
Elon emphasized Grok 3's creative emergent capabilities.
You can also use the Big Brain mode to use more compute and reasoning with Grok 3.
Grok 3 Reasoning performance:
The results correspond to the beta version of Grok-3 Reasoning.
It outperforms o1 and DeepSeek-R1 when given more test-time compute (allowing it to think longer).
The Grok 3 mini reasoning model is also very capable.
Grok 3 Reasoning Beta performance on AIME 2025.
Grok 3 shows generalization capabilities.
It not only does coding and math problem-solving, but it can also do other creative and useful real-world tasks.
One of the results generated with Grok 3 mini.
Bejeweled Tetris generated by Grok 3.
Grok 3 cannot only unlock test-time compute, it also enables capable agents.
These capabilities have led to a new product called DeepSearch.
"Next generation of search agents to understand the universe".
More on DeepSearch:
- the model can think deeply about user intent
- what facts to consider
- how many websites to browse
- it can cross-validate different sources.
DeepSearch also exposes the steps that it takes to conduct the search itself.
Improvements will happen rapidly and almost daily according to the team.
There is also a Grok-powered voice app coming too -- about a week away!
Open-source approach:
The last version will be open-sourced when the most recent version is fully out.
After Grok 3 stable version is out, it is highly likely Grok 2 will be open-sourced. (within a few months).
SuperGrok dedicated app is also available with a polished experience.
Try on the web as well: grok.com
The web will include the latest Grok features.
β€166π158π142π₯137
Forwarded from BlockChainWORLD.ai - daily crypto and AI news and promos!
β
ARE YOU LOOKING FOR QUALITY CRYPTO & AI TRAFFIC? WE ARE HERE TO HELP!
Promote your Crypto, AI Tools, Apps, Exchanges, Play-to-Earn Games, NFTs, DeFi, Wallets, Tokens, Meme Coins, Telegram Bots, P2E, and more!
Get a high-converting video review on our 5 MASSIVE channels:
BLOCKCHAINWORLD.ai β Massive Crypto Community (5,000,000+)
πͺ BLOCKCHAIN WORLD β 1,150,000 Subscribers https://www.youtube.com/@Block-Chain-World/videos
Promotes P2E, Crypto, NFTs, Apps, Wallets, Tokens, Meme Coins, DeFi, GameFi, Web3, Metaverses, AI Projects, Blockchain projects, ICOs, Presales, Coin Launches, etc.
πͺ YouTube INFINITE DIGITAL CHANNEL β 2,750,000 Subscribers https://www.youtube.com/@Infinite-Digital-YT
Covers Crypto, Blockchain, AI Tools, Meme Coins, P2E, NFTs, Apps, Metaverses, Tokens, and AI App reviews.
πͺ WEB3WORLD β 4,346,736 Subscribers https://www.youtube.com/@Web-3-World
Features Web3-related projects, including AI tools, Crypto, NFTs, P2E, Apps, Tokens, Wallets, and Games.
π₯ NEW! CRYPTO BROS & VORTEX π₯
On our massive channels, we review and promote next-gen tech in the crypto space, from brand-new AI tools to must-have apps, Meme Coins, Play-to-Earn games, Tap2Earn, Telegram bots & games, Clickers, NFTs, wallets, tokens, and everything blockchain and Web3-related!
πͺ Crypto Bros β 1.2 Million Subscribers https://www.youtube.com/@CryptoBrosVortex
Focus on Meme Coins, AI Tools, Play2Earn, Tap2Earn, and Crypto presales.
πͺ Vortex: Next Gen β 1.6 Million Subscribers https://www.youtube.com/@VortexNextGen
Covers Tokens, Meme Coins, Exchanges, GameFi, DeFi, Crypto, AI tools, and trends.
π’ MASSIVE PROMOTIONS ON TELEGRAM & TWITTER
πͺ Telegram Crypto & AI Promos β 200,000 Subs https://t.me/web3worldchannel
πͺ Telegram AI & Web3 Promos β 150,000 Subs https://t.me/VortexNextGen
πͺ Twitter X Shoutouts & Promotions β 130,000 Followers https://x.com/VORTEX_Promos
π₯ NEW! 3 Massive 1M+ TikTok Accounts π₯
3 Massive 1M+ TikTok Accounts
Unleashing the Future: Cutting-Edge AI, Crypto Reviews, Blockchain Evolution & Web3 Magic:
https://www.tiktok.com/@vortexnextgen
https://www.tiktok.com/@infinitedigitalyt
https://www.tiktok.com/@web3.world.yt
πͺ BlockchainWorld.ai Website
Order premium listings, top banners, list your projects, and get high-quality Crypto/AI/Apps/Tokens/Wallets/NFT/DeFi/Play-and-Earn traffic. https://blockchainworld.ai/go
π WHY CHOOSE US? GET PROVEN RESULTS!
β 1. Get Massive Traffic & Rank #1 on Google!
Our massive channel size ensures all our video reviews rank #1 on Google, delivering top-quality organic traffic and maximum reach.
Our expert hosts have industry-specific knowledge, guaranteeing high-quality reviews.
With 1,300+ successful video reviews, we know how to drive results!
β 2. Unmatched Social Proof
Being featured on a 1M+ subscriber channel skyrockets credibility and boosts community engagement.
Share the video across your community to create hype and attract smaller KOLs to amplify exposure.
β 3. Trusted by Industry Leaders
Binance, MEXC, CryptoCom, Gate, Nexo, Bitget, KuCoin, Ledger, and more have trusted our services.
We promote cutting-edge AI tools, essential apps, Play-to-Earn games, NFTs, wallets, tokens, and everything Web3.
π° Need more proof? See Case Studies & Success Stories: https://blockchainworld.ai/casestudies
Promote your Crypto, AI Tools, Apps, Exchanges, Play-to-Earn Games, NFTs, DeFi, Wallets, Tokens, Meme Coins, Telegram Bots, P2E, and more!
Get a high-converting video review on our 5 MASSIVE channels:
BLOCKCHAINWORLD.ai β Massive Crypto Community (5,000,000+)
πͺ BLOCKCHAIN WORLD β 1,150,000 Subscribers https://www.youtube.com/@Block-Chain-World/videos
Promotes P2E, Crypto, NFTs, Apps, Wallets, Tokens, Meme Coins, DeFi, GameFi, Web3, Metaverses, AI Projects, Blockchain projects, ICOs, Presales, Coin Launches, etc.
πͺ YouTube INFINITE DIGITAL CHANNEL β 2,750,000 Subscribers https://www.youtube.com/@Infinite-Digital-YT
Covers Crypto, Blockchain, AI Tools, Meme Coins, P2E, NFTs, Apps, Metaverses, Tokens, and AI App reviews.
πͺ WEB3WORLD β 4,346,736 Subscribers https://www.youtube.com/@Web-3-World
Features Web3-related projects, including AI tools, Crypto, NFTs, P2E, Apps, Tokens, Wallets, and Games.
π₯ NEW! CRYPTO BROS & VORTEX π₯
On our massive channels, we review and promote next-gen tech in the crypto space, from brand-new AI tools to must-have apps, Meme Coins, Play-to-Earn games, Tap2Earn, Telegram bots & games, Clickers, NFTs, wallets, tokens, and everything blockchain and Web3-related!
πͺ Crypto Bros β 1.2 Million Subscribers https://www.youtube.com/@CryptoBrosVortex
Focus on Meme Coins, AI Tools, Play2Earn, Tap2Earn, and Crypto presales.
πͺ Vortex: Next Gen β 1.6 Million Subscribers https://www.youtube.com/@VortexNextGen
Covers Tokens, Meme Coins, Exchanges, GameFi, DeFi, Crypto, AI tools, and trends.
π’ MASSIVE PROMOTIONS ON TELEGRAM & TWITTER
πͺ Telegram Crypto & AI Promos β 200,000 Subs https://t.me/web3worldchannel
πͺ Telegram AI & Web3 Promos β 150,000 Subs https://t.me/VortexNextGen
πͺ Twitter X Shoutouts & Promotions β 130,000 Followers https://x.com/VORTEX_Promos
π₯ NEW! 3 Massive 1M+ TikTok Accounts π₯
3 Massive 1M+ TikTok Accounts
Unleashing the Future: Cutting-Edge AI, Crypto Reviews, Blockchain Evolution & Web3 Magic:
https://www.tiktok.com/@vortexnextgen
https://www.tiktok.com/@infinitedigitalyt
https://www.tiktok.com/@web3.world.yt
πͺ BlockchainWorld.ai Website
Order premium listings, top banners, list your projects, and get high-quality Crypto/AI/Apps/Tokens/Wallets/NFT/DeFi/Play-and-Earn traffic. https://blockchainworld.ai/go
π WHY CHOOSE US? GET PROVEN RESULTS!
β 1. Get Massive Traffic & Rank #1 on Google!
Our massive channel size ensures all our video reviews rank #1 on Google, delivering top-quality organic traffic and maximum reach.
Our expert hosts have industry-specific knowledge, guaranteeing high-quality reviews.
With 1,300+ successful video reviews, we know how to drive results!
β 2. Unmatched Social Proof
Being featured on a 1M+ subscriber channel skyrockets credibility and boosts community engagement.
Share the video across your community to create hype and attract smaller KOLs to amplify exposure.
β 3. Trusted by Industry Leaders
Binance, MEXC, CryptoCom, Gate, Nexo, Bitget, KuCoin, Ledger, and more have trusted our services.
We promote cutting-edge AI tools, essential apps, Play-to-Earn games, NFTs, wallets, tokens, and everything Web3.
π° Need more proof? See Case Studies & Success Stories: https://blockchainworld.ai/casestudies
π80β€75π₯75π68
Vortex Next Gen Trends pinned Β«β
ARE YOU LOOKING FOR QUALITY CRYPTO & AI TRAFFIC? WE ARE HERE TO HELP! Promote your Crypto, AI Tools, Apps, Exchanges, Play-to-Earn Games, NFTs, DeFi, Wallets, Tokens, Meme Coins, Telegram Bots, P2E, and more! Get a high-converting video review on our 5β¦Β»
Breakthrough in Robot Design: Universal Controllers Transform How We Build Robots?
Northwestern University researchers have made a significant breakthrough in robotics design, introducing a method that could revolutionize how we create and evolve robots.
Their paper "Accelerated co-design of robots through morphological pretraining" presents a novel approach that solves a decades-old challenge in robotics.
Code is coming soon.
And here are more robots
Key Innovations:
1. Universal Controller
- Developed a single controller that can work with multiple robot body types
- Pre-trained on millions of different robot morphologies
- Uses gradient-based optimization through differentiable simulation
- Can immediately adapt to new robot designs without extensive retraining
2. Zero-Shot Evolution
- Allows rapid testing of new robot body designs
- Enables immediate evaluation of design changes
- Supports successful recombination of robot parts
- Dramatically speeds up the design process
3. Diversity Maintenance
- Identified and solved "diversity collapse" - a previously unknown problem in robot co-design
- Developed methods to maintain morphological diversity while improving performance
- Enabled successful crossover between different robot designs
Technical Details:
- Controllers are trained on over 10 million distinct robot morphologies
- Uses differentiable simulation for gradient-based optimization
- Supports complex 3D environments with varying terrains
- Enables robots to perform adaptive behaviors like phototaxis (movement toward light)
Future Implications:
- Could dramatically accelerate robot design and development
- Opens new possibilities for self-reconfigurable robots
- Provides a framework for more complex multi-material robots
- May help bridge the simulation-to-reality gap in robotics
Northwestern University researchers have made a significant breakthrough in robotics design, introducing a method that could revolutionize how we create and evolve robots.
Their paper "Accelerated co-design of robots through morphological pretraining" presents a novel approach that solves a decades-old challenge in robotics.
Code is coming soon.
And here are more robots
Key Innovations:
1. Universal Controller
- Developed a single controller that can work with multiple robot body types
- Pre-trained on millions of different robot morphologies
- Uses gradient-based optimization through differentiable simulation
- Can immediately adapt to new robot designs without extensive retraining
2. Zero-Shot Evolution
- Allows rapid testing of new robot body designs
- Enables immediate evaluation of design changes
- Supports successful recombination of robot parts
- Dramatically speeds up the design process
3. Diversity Maintenance
- Identified and solved "diversity collapse" - a previously unknown problem in robot co-design
- Developed methods to maintain morphological diversity while improving performance
- Enabled successful crossover between different robot designs
Technical Details:
- Controllers are trained on over 10 million distinct robot morphologies
- Uses differentiable simulation for gradient-based optimization
- Supports complex 3D environments with varying terrains
- Enables robots to perform adaptive behaviors like phototaxis (movement toward light)
Future Implications:
- Could dramatically accelerate robot design and development
- Opens new possibilities for self-reconfigurable robots
- Provides a framework for more complex multi-material robots
- May help bridge the simulation-to-reality gap in robotics
Google
co-design mpt
Abstract
The co-design of robot morphology and neural control typically requires using reinforcement learning to approximate a unique control policy gradient for each body plan, demanding massive amounts of training data to measure the performance of eachβ¦
The co-design of robot morphology and neural control typically requires using reinforcement learning to approximate a unique control policy gradient for each body plan, demanding massive amounts of training data to measure the performance of eachβ¦
π124π111π₯108β€107
HuggingFace released the "Ultra-Scale Playbook"
A free, open-source, book to learn everything about 5D parallelism, ZeRO, fast CUDA kernels, how and why overlap compute & communication β all scaling bottlenecks and tools introduced with motivation, theory, interactive plots from our 4000+ scaling experiments and even NotebookLM podcasters to tag along with you.
- How was DeepSeek trained for $5M only?
- Why did Mistral trained an MoE?
- Why is PyTorch native Data Parallelism implementation so complex under the hood?
- What are all the parallelism techniques and why were they invented?
- Should I use ZeRO-3 or Pipeline Parallelism when scaling and what's the story behind both techniques?
- What is this Context Parallelism that Meta used to train Llama 3? Is it different from Sequence Parallelism?
- What is FP8? how does it compares to BF16?
The largest factor for democratizing AI will always be teaching everyone how to build AI and in particular how to create, train and fine-tune high performance models. In other word making accessible to everybody the techniques that power all recent large language models and efficient training is possibly one of the most essential of them.
A free, open-source, book to learn everything about 5D parallelism, ZeRO, fast CUDA kernels, how and why overlap compute & communication β all scaling bottlenecks and tools introduced with motivation, theory, interactive plots from our 4000+ scaling experiments and even NotebookLM podcasters to tag along with you.
- How was DeepSeek trained for $5M only?
- Why did Mistral trained an MoE?
- Why is PyTorch native Data Parallelism implementation so complex under the hood?
- What are all the parallelism techniques and why were they invented?
- Should I use ZeRO-3 or Pipeline Parallelism when scaling and what's the story behind both techniques?
- What is this Context Parallelism that Meta used to train Llama 3? Is it different from Sequence Parallelism?
- What is FP8? how does it compares to BF16?
The largest factor for democratizing AI will always be teaching everyone how to build AI and in particular how to create, train and fine-tune high performance models. In other word making accessible to everybody the techniques that power all recent large language models and efficient training is possibly one of the most essential of them.
huggingface.co
The Ultra-Scale Playbook - a Hugging Face Space by nanotron
The ultimate guide to training LLM on large GPU Clusters
β€57π₯47π46π46
Wow, DeepSeek announced Day 0: Warming up for OpenSourceWeek
Starting next week, they'll be open-sourcing 5 repos, sharing sincere progress with full transparency.
These humble building blocks in their online service have been documented, deployed and battle-tested in production.
Daily unlocks are coming soon. No ivory towers - just pure garage-energy and community-driven innovation.
Starting next week, they'll be open-sourcing 5 repos, sharing sincere progress with full transparency.
These humble building blocks in their online service have been documented, deployed and battle-tested in production.
Daily unlocks are coming soon. No ivory towers - just pure garage-energy and community-driven innovation.
π83β€81π₯74π72
This media is not supported in your browser
VIEW IN TELEGRAM
The Pika neural network has introduced the Pikaswaps feature, which changes objects in a video to any kind of thing. You can replace a pancake with a human face, a dog with an iguana, or your hand with a cyber prosthesis.
Special effects are no longer needed
Special effects are no longer needed
π115π₯115β€107π105
This media is not supported in your browser
VIEW IN TELEGRAM
A new android has been created in Norway β NEO Gamma. The mechanical servant from 1X Technologies will do housework and bring coffee to its owners.
Robots do the hard work, not humans
Robots do the hard work, not humans
π₯104π96π95β€82
This media is not supported in your browser
VIEW IN TELEGRAM
Anthropic to release Claude Sonnet 3.7 on Feb 26
Itβs expected to have step-by-step thinking, never before seen coding capabilities and web search.
The best coding model which powers Cursor and Windsurf is about to get a whole lot better.
Claude 3.7 Sonnet is Anthropic's most intelligent model to date and the first Claude model to offer extended thinking - the ability to solve complex problems with careful, step-by-step reasoning.
Anthropic is the first AI lab to introduce a single model where users can balance speed and quality by choosing between standard thinking for near-instant responses or extended thinking or advanced reasoning.
Claude 3.7 Sonnet is state-of-the-art for coding, and delivers advancements in computer use, agentic capabilities, complex reasoning, and content generation. With frontier performance and more control over speed, Claude 3.7 Sonnet is the ideal choice for powering AI agents, especially customer-facing agents, and complex AI workflows.
Supported use cases: RAG or search & retrieval over vast amounts of knowledge, product recommendations, forecasting, targeted marketing, code generation, quality control, parse text from images, agentic computer use, content generation
Model attributes: Reasoning, Text generation, Code generation, Rich text formatting, Agentic computer use
Itβs expected to have step-by-step thinking, never before seen coding capabilities and web search.
The best coding model which powers Cursor and Windsurf is about to get a whole lot better.
Claude 3.7 Sonnet is Anthropic's most intelligent model to date and the first Claude model to offer extended thinking - the ability to solve complex problems with careful, step-by-step reasoning.
Anthropic is the first AI lab to introduce a single model where users can balance speed and quality by choosing between standard thinking for near-instant responses or extended thinking or advanced reasoning.
Claude 3.7 Sonnet is state-of-the-art for coding, and delivers advancements in computer use, agentic capabilities, complex reasoning, and content generation. With frontier performance and more control over speed, Claude 3.7 Sonnet is the ideal choice for powering AI agents, especially customer-facing agents, and complex AI workflows.
Supported use cases: RAG or search & retrieval over vast amounts of knowledge, product recommendations, forecasting, targeted marketing, code generation, quality control, parse text from images, agentic computer use, content generation
Model attributes: Reasoning, Text generation, Code generation, Rich text formatting, Agentic computer use
π174π168π₯159β€149
AI models now handle voice/speech yet building with them in Python is very frustrating
FastRTC is here to solve
- Automatic Voice Detection
- Handling WebRTC & the backend for real-time apps
- Calling Phones
Github
FastRTC is here to solve
- Automatic Voice Detection
- Handling WebRTC & the backend for real-time apps
- Calling Phones
Github
huggingface.co
FastRTC: The Real-Time Communication Library for Python
Weβre on a journey to advance and democratize artificial intelligence through open source and open science.
β€72π₯72π68π55
DeepSeek makes 2 major announcements
1. Starting today, DeepSeek is offering significant discounts on their API Platform during off-peak hours (16:30-00:30 UTC daily):
β’ DeepSeek-V3: 50% OFF
β’ DeepSeek-R1: Massive 75% OFF
This means you can access powerful AI models at a fraction of the cost during these hours. For example, DeepSeek-R1 output cost drops from $2.19 to just $0.550 per 1M tokens!
2. DeepSeek has also released DeepGEMM - an impressive FP8 GEMM library that supports both dense and MoE GEMMs, powering their V3/R1 models.
Key features:
- Up to 1350+ FP8 TFLOPS on Hopper GPUs
- Lightweight with no heavy dependencies
- Fully Just-In-Time compiled
- Core logic at just ~300 lines of code
- Outperforms expert-tuned kernels on most matrix sizes
- Supports dense layout and two MoE layouts
1. Starting today, DeepSeek is offering significant discounts on their API Platform during off-peak hours (16:30-00:30 UTC daily):
β’ DeepSeek-V3: 50% OFF
β’ DeepSeek-R1: Massive 75% OFF
This means you can access powerful AI models at a fraction of the cost during these hours. For example, DeepSeek-R1 output cost drops from $2.19 to just $0.550 per 1M tokens!
2. DeepSeek has also released DeepGEMM - an impressive FP8 GEMM library that supports both dense and MoE GEMMs, powering their V3/R1 models.
Key features:
- Up to 1350+ FP8 TFLOPS on Hopper GPUs
- Lightweight with no heavy dependencies
- Fully Just-In-Time compiled
- Core logic at just ~300 lines of code
- Outperforms expert-tuned kernels on most matrix sizes
- Supports dense layout and two MoE layouts
π₯213π180π177β€168
New announcements from DeepSeek Optimized Parallelism Strategies
1. DualPipe - a bidirectional pipeline parallelism algorithm for computation-communication overlap in V3/R1 training.
2. EPLB - an expert-parallel load balancer for V3/R1.
3. Analyze computation-communication overlap in V3/R1.
1. DualPipe - a bidirectional pipeline parallelism algorithm for computation-communication overlap in V3/R1 training.
2. EPLB - an expert-parallel load balancer for V3/R1.
3. Analyze computation-communication overlap in V3/R1.
GitHub
GitHub - deepseek-ai/DualPipe: A bidirectional pipeline parallelism algorithm for computation-communication overlap in DeepSeekβ¦
A bidirectional pipeline parallelism algorithm for computation-communication overlap in DeepSeek V3/R1 training. - deepseek-ai/DualPipe
π81π75π₯69β€64
GPT-4.5 is out! Knowledge Still Stuck in October 2023, itβs not going to blow your mind, but it might befriend you.
It's more like a personality, communication, and creativity upgrade than a huge intelligence leap. It's like OpenAI is pivoting its base model from "bland assistant" to "AI bestie."
What it does do well:
- OpenAI says it scores 64% on SimpleQA (double GPT-4's score)
- Much better writing with cleaner, better structured, more human-like prose
- Genuinely warmer and more emotionally intelligent (gave me some good advice!)
- Less robotic, more opinionated responses
4.5 is more extroverted, agreeable, and less neurotic than 4o.
It's sometimes worse at following instructions and because it's less sycophantic and more creative.
The model received approximately 10x more computational resources during pre-training compared to GPT-4. Training occurred simultaneously across multiple data centers.
Pricing $75 per million input tokens and $150 per million output tokens β 15-30x more expensive than GPT-4o! This pricing reflects the model's scale and resource requirements.
Performance and Context Generation is noticeably slower than its predecessors, context length remains at 128K tokens. Knowledge cutoff stays at October 2023, which is disappointing for many users.
Functionality Supports Canvas, search, and file uploads. Currently lacks multimodal features like voice mode or video.
Availability:
Already available to Pro users and developers of all API tiers
Coming to Plus subscribers ($20) next week
OpenAI plans to add "tens of thousands of GPUs" next week to expand access
Independent Benchmark Results:
Aider Polyglot Coding Benchmark: Recent tests show that GPT-4.5 Preview significantly outperforms its predecessor but lags behind specialized models:
Claude 3.7 Sonnet with thinking mode (32k tokens) β 65%
Claude 3.7 Sonnet without thinking mode β 60%
DeepSeek V3 β 48%
GPT-4.5 Preview β 45%
ChatGPT-4o β 27%
GPT-4o β 23%
It's more like a personality, communication, and creativity upgrade than a huge intelligence leap. It's like OpenAI is pivoting its base model from "bland assistant" to "AI bestie."
What it does do well:
- OpenAI says it scores 64% on SimpleQA (double GPT-4's score)
- Much better writing with cleaner, better structured, more human-like prose
- Genuinely warmer and more emotionally intelligent (gave me some good advice!)
- Less robotic, more opinionated responses
4.5 is more extroverted, agreeable, and less neurotic than 4o.
It's sometimes worse at following instructions and because it's less sycophantic and more creative.
The model received approximately 10x more computational resources during pre-training compared to GPT-4. Training occurred simultaneously across multiple data centers.
Pricing $75 per million input tokens and $150 per million output tokens β 15-30x more expensive than GPT-4o! This pricing reflects the model's scale and resource requirements.
Performance and Context Generation is noticeably slower than its predecessors, context length remains at 128K tokens. Knowledge cutoff stays at October 2023, which is disappointing for many users.
Functionality Supports Canvas, search, and file uploads. Currently lacks multimodal features like voice mode or video.
Availability:
Already available to Pro users and developers of all API tiers
Coming to Plus subscribers ($20) next week
OpenAI plans to add "tens of thousands of GPUs" next week to expand access
Independent Benchmark Results:
Aider Polyglot Coding Benchmark: Recent tests show that GPT-4.5 Preview significantly outperforms its predecessor but lags behind specialized models:
Claude 3.7 Sonnet with thinking mode (32k tokens) β 65%
Claude 3.7 Sonnet without thinking mode β 60%
DeepSeek V3 β 48%
GPT-4.5 Preview β 45%
ChatGPT-4o β 27%
GPT-4o β 23%
Openai
Introducing GPT-4.5
Weβre releasing a research preview of GPTβ4.5βour largest and best model for chat yet. GPTβ4.5 is a step forward in scaling up pre-training and post-training.
π₯73β€70π70π63
Media is too big
VIEW IN TELEGRAM
The Magnific neural network can change the style of pictures in seconds, while accurately preserving their essence. AI can turn even a simple sketch on paper into a high-quality render or photorealistic picture.
π122π₯111π109β€97
Want next-level access control that is fast, secure, and scalable? π Mercury Access by JM Digital is a cutting-edge solution designed to provide seamless authentication, advanced security, and unparalleled access management for businesses and institutions.
β Smart & secure access control for any environment
β Scalable & flexible to meet your security needs
β Seamless integration with modern digital infrastructure
β Enhanced authentication & encrypted security measures
π Discover Mercury Access today: https://www.jmdigital.tech/mercury-access
https://vt.tiktok.com/ZSM5LXeq1/
β Smart & secure access control for any environment
β Scalable & flexible to meet your security needs
β Seamless integration with modern digital infrastructure
β Enhanced authentication & encrypted security measures
π Discover Mercury Access today: https://www.jmdigital.tech/mercury-access
https://vt.tiktok.com/ZSM5LXeq1/
JMD
Mercury Access NFT Membership | Exclusive Blockchain Rewards
Get exclusive access to NFT membership benefits with Mercury Access. Enjoy private communities, discounts, and early product launches. Join today!
π₯93π77β€72π68
This media is not supported in your browser
VIEW IN TELEGRAM
AI has learned to play Pokemon. The newest model Claude 3.7 Sonnet was ordered to play Pokemon Red, and it has already beaten the leader of the first stadium.
β€176π164π₯156π154
DeepSeek introduced DeepSeek-V3/R1 Inference System Overview
Optimized throughput and latency via:
1. Cross-node EP-powered batch scaling
2. Computation-communication overlap
3. Load balancing
Statistics of DeepSeek's Online Service:
- 73.7k/14.8k input/output tokens per second per H800 node
- Cost profit margin 545%
Optimized throughput and latency via:
1. Cross-node EP-powered batch scaling
2. Computation-communication overlap
3. Load balancing
Statistics of DeepSeek's Online Service:
- 73.7k/14.8k input/output tokens per second per H800 node
- Cost profit margin 545%
GitHub
open-infra-index/202502OpenSourceWeek/day_6_one_more_thing_deepseekV3R1_inference_system_overview.md at main Β· deepseek-ai/openβ¦
Production-tested AI infrastructure tools for efficient AGI development and community-driven innovation - deepseek-ai/open-infra-index
π130β€113π₯100π96
Huge VLM release from Cohere for AI is just in
Aya-Vision is a new VLM family based on SigLIP and Aya, and it outperforms many larger models.
> 8B and 32B models covering 23 languages and two new benchmark dataset
> supported by HF transformers from get-go
Aya-Vision is a new VLM family based on SigLIP and Aya, and it outperforms many larger models.
> 8B and 32B models covering 23 languages and two new benchmark dataset
> supported by HF transformers from get-go
huggingface.co
Cohere Labs Aya Vision - a CohereLabs Collection
Aya Vision is a state-of-the-art family of vision models that brings multimodal capabilities to 23 languages.
β€225π216π₯202π197
Today Anthropic submitted their recommendations to the OSTP for the U.S. AI Action Plan
Anthropic predicts powerful AI systems will appear by late 2026 or early 2027, with intellectual abilities matching Nobel Prize winners, able to autonomously handle digital tasks (text, audio, video, internet browsing), reason independently over hours or weeks, and control physical equipment digitally
They recommend stronger national security actions, including government testing of AI models for security risks, stricter export controls on key chips like the H20, and secure communication channels between AI labs and intelligence agencies
They suggest the government build 50 gigawatts of additional power capacity dedicated to AI by 2027, speed up AI adoption across federal agencies, and improve economic data collection to prepare for AIβs impact on jobs and society
Anthropic predicts powerful AI systems will appear by late 2026 or early 2027, with intellectual abilities matching Nobel Prize winners, able to autonomously handle digital tasks (text, audio, video, internet browsing), reason independently over hours or weeks, and control physical equipment digitally
They recommend stronger national security actions, including government testing of AI models for security risks, stricter export controls on key chips like the H20, and secure communication channels between AI labs and intelligence agencies
They suggest the government build 50 gigawatts of additional power capacity dedicated to AI by 2027, speed up AI adoption across federal agencies, and improve economic data collection to prepare for AIβs impact on jobs and society
Anthropic
Anthropicβs recommendations to OSTP for the U.S. AI action plan
Anthropic is an AI safety and research company that's working to build reliable, interpretable, and steerable AI systems.
π₯200β€187π165π157