Skills in action on Grok for iOS π
* not available yet
* not available yet
π5 3β€1
We will likely see a deeper integration between Codex and ChatGPT already very soon.
> Use the ChatGPT app on your phone to keep working with Codex whenever your computer is awake.
Additionally, this image from OpenAI sparked loads of speculations, including the one where OpenAI would be teasing their own mobile phone.
Even though it is quite unrealistic, this would be a huge steal of attention from the Google I/O event.
> Use the ChatGPT app on your phone to keep working with Codex whenever your computer is awake.
Additionally, this image from OpenAI sparked loads of speculations, including the one where OpenAI would be teasing their own mobile phone.
Even though it is quite unrealistic, this would be a huge steal of attention from the Google I/O event.
OpenAI set to add remote Codex control to ChatGPT mobile app
OpenAI teased a Thursday launch, prompting device rumors, but app evidence points to mobile control for Codex inside ChatGPT.
π #chatgpt @testingcatalog
OpenAI teased a Thursday launch, prompting device rumors, but app evidence points to mobile control for Codex inside ChatGPT.
π #chatgpt @testingcatalog
TestingCatalog AI News
OpenAI set to add remote Codex control to ChatGPT mobile app
OpenAI teases a new ChatGPT feature allowing direct control of Codex from mobile devices, aiming to unify coding workflows.
β€7π₯3π2
GOOGLE I/O π₯: New evidence of the upcoming Gemini Omni vide model has been spotted on the Gemini mobile app.
A video sample below π
> "Meet our new video model. Remix your videos, edit directly in chat, try a template, and more."
> Based on the description, we might be really talking about the true "Omni" model based on Gemini, rather than Veo.
> It also seems to be quickly consuming usage limits, based on early tests. "Usage" is a new tab that will be available on both the web and mobile.
A video sample below π
> "Meet our new video model. Remix your videos, edit directly in chat, try a template, and more."
> Based on the description, we might be really talking about the true "Omni" model based on Gemini, rather than Veo.
> It also seems to be quickly consuming usage limits, based on early tests. "Usage" is a new tab that will be available on both the web and mobile.
π¨ AI News | TestingCatalog
GOOGLE I/O π₯: New evidence of the upcoming Gemini Omni vide model has been spotted on the Gemini mobile app. A video sample below π > "Meet our new video model. Remix your videos, edit directly in chat, try a template, and more." > Based on the descriptionβ¦
This media is not supported in your browser
VIEW IN TELEGRAM
Sample video and early feedback (quotes from Reddit)
> I wonβt lie, this is one of the best video models I have seen, maybe not *the* best, but a really strong performance. I was particularly impressed by the prompt adherence (except for the one shot with the missing centerpiece), the model nailed all the constraints.
> Additionally, the voice quality is much better than the Veo models by quite a large margin. It even added some light background music, that would fit right in with an upscale dining experience.
> While there are some continuity issues if you look close enough, the ability to change camera angles on the fly so frequently and with good coherence is impressive to me. Overall this is definitely the new model and quite a step up from the Veo we are used to
> I wonβt lie, this is one of the best video models I have seen, maybe not *the* best, but a really strong performance. I was particularly impressed by the prompt adherence (except for the one shot with the missing centerpiece), the model nailed all the constraints.
> Additionally, the voice quality is much better than the Veo models by quite a large margin. It even added some light background music, that would fit right in with an upscale dining experience.
> While there are some continuity issues if you look close enough, the ability to change camera angles on the fly so frequently and with good coherence is impressive to me. Overall this is definitely the new model and quite a step up from the Veo we are used to
β€12 6π₯4π1
OPENAI π₯: A mention of a new Ultrafast mode appeared for some time on the Codex GitHub repository.
> "The fastest available responses for latency-sensitive work."
Seems like it was unintended push.
> "The fastest available responses for latency-sensitive work."
Seems like it was unintended push.
β€5π3π₯3π1
π¨ AI News | TestingCatalog
Sample video and early feedback (quotes from Reddit) > I wonβt lie, this is one of the best video models I have seen, maybe not *the* best, but a really strong performance. I was particularly impressed by the prompt adherence (except for the one shot withβ¦
This media is not supported in your browser
VIEW IN TELEGRAM
GOOGLE π₯: An upcoming Gemini Omni video model from Google is expected to be much more advanced in video editing, capable of completing tasks like removing watermarks, replacing objects in the video, and more.
It is also likely that Google will release 2 versions of this model, including a Pro variant.
And I assume what we see isn't Pro?
Anime sample π
h/t @QuantumFast
It is also likely that Google will release 2 versions of this model, including a Pro variant.
And I assume what we see isn't Pro?
Anime sample π
h/t @QuantumFast
Googleβs Gemini Omni video model surfaces ahead of I/O debut
Leaked Gemini Omni details point to Google unveiling a unified video model at I/O, with strong in-chat editing and remix tools but generation quality trailing Seedance 2. Credit-based limits and possible Flash/Pro tiers also surfaced.
π #gemini @testingcatalog
Leaked Gemini Omni details point to Google unveiling a unified video model at I/O, with strong in-chat editing and remix tools but generation quality trailing Seedance 2. Credit-based limits and possible Flash/Pro tiers also surfaced.
π #gemini @testingcatalog
TestingCatalog AI News
Googleβs Gemini Omni video model surfaces ahead of I/O debut
Googleβs upcoming Gemini Omni video model briefly surfaced, revealing new video editing features ahead of Google I/O 2026.
π4 2β€1
Google keeps preparing its upcoming Gemini Omni models for the release.
> Gemini Omni model will be available on APIs as well
> The model will be considered as Agent, similarly to Deep Research on AI Studio
Soon? π
> Gemini Omni model will be available on APIs as well
> The model will be considered as Agent, similarly to Deep Research on AI Studio
Soon? π
β€9π6π₯5 1
Anthropic adds Agent View to Claude Code CLI interface
Anthropicβs Agent View for Claude Code adds a CLI dashboard for managing parallel coding sessions in one place. It shows status, activity, and input needs, supports background jobs, and is available now in Research Preview.
π #claude @testingcatalog
Anthropicβs Agent View for Claude Code adds a CLI dashboard for managing parallel coding sessions in one place. It shows status, activity, and input needs, supports background jobs, and is available now in Research Preview.
π #claude @testingcatalog
TestingCatalog AI News
Anthropic adds Agent View to Claude Code CLI interface
Anthropic introduces Agent View for Claude Code, allowing developers to manage parallel coding sessions in a single command-line dashboard.
THINKING MACHINES π₯: Research preview of a new family of realtime voice models have been announced!
> Today, weβre announcing a research preview of interaction models: models that handle interaction natively rather than through external scaffolding.
> Our research preview demonstrates qualitatively new interaction capabilities, as well as state-of-the-art combined performance in intelligence and responsiveness.
A new SOTA?! π
> Today, weβre announcing a research preview of interaction models: models that handle interaction natively rather than through external scaffolding.
> Our research preview demonstrates qualitatively new interaction capabilities, as well as state-of-the-art combined performance in intelligence and responsiveness.
A new SOTA?! π
β€4π4 4
OpenAI announces Daybreak initiative around Codex Security
OpenAI launched Daybreak, a cybersecurity program that extends Codex into secure code review, threat modeling, patch validation, and detection support, with verified access, partner integrations, and rollout for defenders and enterprises.
π #chatgpt @testingcatalog
OpenAI launched Daybreak, a cybersecurity program that extends Codex into secure code review, threat modeling, patch validation, and detection support, with verified access, partner integrations, and rollout for defenders and enterprises.
π #chatgpt @testingcatalog
TestingCatalog AI News
OpenAI announces Daybreak initiative around Codex Security
OpenAI launches Daybreak, a cybersecurity initiative integrating AI models and Codex Security to help organizations patch vulnerabilities.
β€4π4π₯2
Thinking Machines announced new Interaction Voice Models
Thinking Machines unveiled a research preview of multimodal AI models built for real-time collaboration across audio, video, and text, using native time-aware processing, low-latency micro-turns, and background reasoning.
π #ai @testingcatalog
Thinking Machines unveiled a research preview of multimodal AI models built for real-time collaboration across audio, video, and text, using native time-aware processing, low-latency micro-turns, and background reasoning.
π #ai @testingcatalog
TestingCatalog AI News
Thinking Machines announced new Interaction Voice Models
What's new? Thinking Machines previewed its AI for real-time native exchange over audio, video, and text.
β€3π2
π¨ AI News | TestingCatalog
Anthropic adds Agent View to Claude Code CLI interface Anthropicβs Agent View for Claude Code adds a CLI dashboard for managing parallel coding sessions in one place. It shows status, activity, and input needs, supports background jobs, and is available nowβ¦
This media is not supported in your browser
VIEW IN TELEGRAM
Anthropic released Agent View in Claude Code CLI, from where users can observe and interact with parallel-running agents.
It looks like preparation for a future in which agents will pursue broader long-term goals. Claude's mobile app is being prepared for that as well.
It looks like preparation for a future in which agents will pursue broader long-term goals. Claude's mobile app is being prepared for that as well.
π4 4β€1
GOOGLE π₯: A new Gemini Omni banner has been added to the web build recently.
> Gemini Omni will be an Agent that can combine text, images, and videos.
> Users will be able to add themselves to different scenes. As we know, AI Avatars (Likeness) are coming to Gemini as well, and Gemini Omni will likely be connected to that.
> "Likeness" feature will likely be highly coupled to mobile apps (as it used to work on Sora).
What's the chance we will get it today during the Android show?
h/t @Thomasguka
> Gemini Omni will be an Agent that can combine text, images, and videos.
> Users will be able to add themselves to different scenes. As we know, AI Avatars (Likeness) are coming to Gemini as well, and Gemini Omni will likely be connected to that.
> "Likeness" feature will likely be highly coupled to mobile apps (as it used to work on Sora).
What's the chance we will get it today during the Android show?
h/t @Thomasguka
π5β€2