Early look at Imagine Agent Mode on Grok app for iOS!
Users will be able to use Imagine Agent via a mobile optimised native UI to generate images and videos that require more complex workflows.
SpaceXAI is getting quite ahead of everyone else on this front!
We just need Imagine v2 now π
Additionally, Skills are coming soon on mobile as well.
Users will be able to use Imagine Agent via a mobile optimised native UI to generate images and videos that require more complex workflows.
SpaceXAI is getting quite ahead of everyone else on this front!
We just need Imagine v2 now π
Additionally, Skills are coming soon on mobile as well.
π4β€1 1
Hermes vs OpenClaw π₯
Hermes Agent overtook OpenClaw on the Global OpenRouter token ranking and claimed the first spot.
Tokens are a new currency!
Hermes Agent overtook OpenClaw on the Global OpenRouter token ranking and claimed the first spot.
Tokens are a new currency!
β€9π₯5
Skills in action on Grok for iOS π
* not available yet
* not available yet
π5 3β€1
We will likely see a deeper integration between Codex and ChatGPT already very soon.
> Use the ChatGPT app on your phone to keep working with Codex whenever your computer is awake.
Additionally, this image from OpenAI sparked loads of speculations, including the one where OpenAI would be teasing their own mobile phone.
Even though it is quite unrealistic, this would be a huge steal of attention from the Google I/O event.
> Use the ChatGPT app on your phone to keep working with Codex whenever your computer is awake.
Additionally, this image from OpenAI sparked loads of speculations, including the one where OpenAI would be teasing their own mobile phone.
Even though it is quite unrealistic, this would be a huge steal of attention from the Google I/O event.
OpenAI set to add remote Codex control to ChatGPT mobile app
OpenAI teased a Thursday launch, prompting device rumors, but app evidence points to mobile control for Codex inside ChatGPT.
π #chatgpt @testingcatalog
OpenAI teased a Thursday launch, prompting device rumors, but app evidence points to mobile control for Codex inside ChatGPT.
π #chatgpt @testingcatalog
TestingCatalog AI News
OpenAI set to add remote Codex control to ChatGPT mobile app
OpenAI teases a new ChatGPT feature allowing direct control of Codex from mobile devices, aiming to unify coding workflows.
β€7π₯3π2
GOOGLE I/O π₯: New evidence of the upcoming Gemini Omni vide model has been spotted on the Gemini mobile app.
A video sample below π
> "Meet our new video model. Remix your videos, edit directly in chat, try a template, and more."
> Based on the description, we might be really talking about the true "Omni" model based on Gemini, rather than Veo.
> It also seems to be quickly consuming usage limits, based on early tests. "Usage" is a new tab that will be available on both the web and mobile.
A video sample below π
> "Meet our new video model. Remix your videos, edit directly in chat, try a template, and more."
> Based on the description, we might be really talking about the true "Omni" model based on Gemini, rather than Veo.
> It also seems to be quickly consuming usage limits, based on early tests. "Usage" is a new tab that will be available on both the web and mobile.
π¨ AI News | TestingCatalog
GOOGLE I/O π₯: New evidence of the upcoming Gemini Omni vide model has been spotted on the Gemini mobile app. A video sample below π > "Meet our new video model. Remix your videos, edit directly in chat, try a template, and more." > Based on the descriptionβ¦
This media is not supported in your browser
VIEW IN TELEGRAM
Sample video and early feedback (quotes from Reddit)
> I wonβt lie, this is one of the best video models I have seen, maybe not *the* best, but a really strong performance. I was particularly impressed by the prompt adherence (except for the one shot with the missing centerpiece), the model nailed all the constraints.
> Additionally, the voice quality is much better than the Veo models by quite a large margin. It even added some light background music, that would fit right in with an upscale dining experience.
> While there are some continuity issues if you look close enough, the ability to change camera angles on the fly so frequently and with good coherence is impressive to me. Overall this is definitely the new model and quite a step up from the Veo we are used to
> I wonβt lie, this is one of the best video models I have seen, maybe not *the* best, but a really strong performance. I was particularly impressed by the prompt adherence (except for the one shot with the missing centerpiece), the model nailed all the constraints.
> Additionally, the voice quality is much better than the Veo models by quite a large margin. It even added some light background music, that would fit right in with an upscale dining experience.
> While there are some continuity issues if you look close enough, the ability to change camera angles on the fly so frequently and with good coherence is impressive to me. Overall this is definitely the new model and quite a step up from the Veo we are used to
β€12 6π₯4π1
OPENAI π₯: A mention of a new Ultrafast mode appeared for some time on the Codex GitHub repository.
> "The fastest available responses for latency-sensitive work."
Seems like it was unintended push.
> "The fastest available responses for latency-sensitive work."
Seems like it was unintended push.
β€5π3π₯3π1
π¨ AI News | TestingCatalog
Sample video and early feedback (quotes from Reddit) > I wonβt lie, this is one of the best video models I have seen, maybe not *the* best, but a really strong performance. I was particularly impressed by the prompt adherence (except for the one shot withβ¦
This media is not supported in your browser
VIEW IN TELEGRAM
GOOGLE π₯: An upcoming Gemini Omni video model from Google is expected to be much more advanced in video editing, capable of completing tasks like removing watermarks, replacing objects in the video, and more.
It is also likely that Google will release 2 versions of this model, including a Pro variant.
And I assume what we see isn't Pro?
Anime sample π
h/t @QuantumFast
It is also likely that Google will release 2 versions of this model, including a Pro variant.
And I assume what we see isn't Pro?
Anime sample π
h/t @QuantumFast
Googleβs Gemini Omni video model surfaces ahead of I/O debut
Leaked Gemini Omni details point to Google unveiling a unified video model at I/O, with strong in-chat editing and remix tools but generation quality trailing Seedance 2. Credit-based limits and possible Flash/Pro tiers also surfaced.
π #gemini @testingcatalog
Leaked Gemini Omni details point to Google unveiling a unified video model at I/O, with strong in-chat editing and remix tools but generation quality trailing Seedance 2. Credit-based limits and possible Flash/Pro tiers also surfaced.
π #gemini @testingcatalog
TestingCatalog AI News
Googleβs Gemini Omni video model surfaces ahead of I/O debut
Googleβs upcoming Gemini Omni video model briefly surfaced, revealing new video editing features ahead of Google I/O 2026.
π4 2β€1
Google keeps preparing its upcoming Gemini Omni models for the release.
> Gemini Omni model will be available on APIs as well
> The model will be considered as Agent, similarly to Deep Research on AI Studio
Soon? π
> Gemini Omni model will be available on APIs as well
> The model will be considered as Agent, similarly to Deep Research on AI Studio
Soon? π
β€9π6π₯5 1
Anthropic adds Agent View to Claude Code CLI interface
Anthropicβs Agent View for Claude Code adds a CLI dashboard for managing parallel coding sessions in one place. It shows status, activity, and input needs, supports background jobs, and is available now in Research Preview.
π #claude @testingcatalog
Anthropicβs Agent View for Claude Code adds a CLI dashboard for managing parallel coding sessions in one place. It shows status, activity, and input needs, supports background jobs, and is available now in Research Preview.
π #claude @testingcatalog
TestingCatalog AI News
Anthropic adds Agent View to Claude Code CLI interface
Anthropic introduces Agent View for Claude Code, allowing developers to manage parallel coding sessions in a single command-line dashboard.
THINKING MACHINES π₯: Research preview of a new family of realtime voice models have been announced!
> Today, weβre announcing a research preview of interaction models: models that handle interaction natively rather than through external scaffolding.
> Our research preview demonstrates qualitatively new interaction capabilities, as well as state-of-the-art combined performance in intelligence and responsiveness.
A new SOTA?! π
> Today, weβre announcing a research preview of interaction models: models that handle interaction natively rather than through external scaffolding.
> Our research preview demonstrates qualitatively new interaction capabilities, as well as state-of-the-art combined performance in intelligence and responsiveness.
A new SOTA?! π
β€4π4 4
OpenAI announces Daybreak initiative around Codex Security
OpenAI launched Daybreak, a cybersecurity program that extends Codex into secure code review, threat modeling, patch validation, and detection support, with verified access, partner integrations, and rollout for defenders and enterprises.
π #chatgpt @testingcatalog
OpenAI launched Daybreak, a cybersecurity program that extends Codex into secure code review, threat modeling, patch validation, and detection support, with verified access, partner integrations, and rollout for defenders and enterprises.
π #chatgpt @testingcatalog
TestingCatalog AI News
OpenAI announces Daybreak initiative around Codex Security
OpenAI launches Daybreak, a cybersecurity initiative integrating AI models and Codex Security to help organizations patch vulnerabilities.
β€4π4π₯2