Please open Telegram to view this post
VIEW IN TELEGRAM
YouTube
Google Keynote (Google I/O ‘24)
It’s time to I/O! Tune in to learn the latest news, announcements, and AI updates from Google.
Update: As of September 2, if you’re 18 or over, your files are now used to improve Google AI but you can control this in Gemini Apps Activity. Learn more in the…
Update: As of September 2, if you’re 18 or over, your files are now used to improve Google AI but you can control this in Gemini Apps Activity. Learn more in the…
👍4
OpenAI is considering allowing generation of NSFW content
OpenAI is considering allowing NSFW content generation on DALL-E and more importantly, ChatGPT. Despite this, they will still have a ban on deepfake content, especially adult related. This suggestion was revealed in a recent document that talks about the development of the AI.
Now, the main point is: OpenAI MAY allow NSFW soon. And as it's the leader in AI space, consider how other AI services (ESPECIALLY Character.ai) react.
🔗 Link
🧑💻 @agamtechtricks
OpenAI is considering allowing NSFW content generation on DALL-E and more importantly, ChatGPT. Despite this, they will still have a ban on deepfake content, especially adult related. This suggestion was revealed in a recent document that talks about the development of the AI.
Now, the main point is: OpenAI MAY allow NSFW soon. And as it's the leader in AI space, consider how other AI services (ESPECIALLY Character.ai) react.
Please open Telegram to view this post
VIEW IN TELEGRAM
🌚29👍6🤣5❤2
Google Photos ‘Ask Photos’ will let you search your images with voice and text prompts
"Ask Photos" will make it easier to search across your photos and videos with the help of Gemini models. It goes beyond simple search to understand context and answer more complex questions
🔗 9to5Google
🧑💻 @agamtechtricks
"Ask Photos" will make it easier to search across your photos and videos with the help of Gemini models. It goes beyond simple search to understand context and answer more complex questions
Please open Telegram to view this post
VIEW IN TELEGRAM
🤯5👍2
NotebookLM gains podcast-like audio overviews with live Q&A
Google is upgrading NotebookLM, for asking questions about a set of documents, with new “audio overviews” that generate a podcast-like discussion between two speakers.
This feature was presented from the perspective of someone who learns best when listening rather than just reading. In the on-stage example, NotebookLM was given a set of simple physics lessons as its base context. By generating an audio overview, the tool generated a conversation between two speakers and voiced both participants in real time.
🔗 9to5Google
🧑💻 @agamtechtricks
Google is upgrading NotebookLM, for asking questions about a set of documents, with new “audio overviews” that generate a podcast-like discussion between two speakers.
This feature was presented from the perspective of someone who learns best when listening rather than just reading. In the on-stage example, NotebookLM was given a set of simple physics lessons as its base context. By generating an audio overview, the tool generated a conversation between two speakers and voiced both participants in real time.
Please open Telegram to view this post
VIEW IN TELEGRAM
🤯3❤1
Google is infusing AI into Android with Gemini app and on-device Nano
Gemini Nano is a new, 3.8B parameter that is designed to run on-device and can process not just text input but also audio and images. It’s coming later this year “starting with Pixel” and will be used for:
• Clearer descriptions with TalkBack: TalkBack will soon be able to automatically generate more useful image descriptions. This will help people with visual impairments who can’t see images, especially when those images don’t have alt text already.
• Scam detection:
A new feature that processes voice calls to detect and warn you when the person on the other end is trying to scam you. The feature will look for conversation patterns commonly associated with scams, like when a “bank representative” asks you to urgently transfer funds. If so, your phone will buzz and show a warning. This feature is opt-in and will be processed on-device by Gemini Nano.
ℹ️ Credits: @MishaalAndroidNews
🔗 9to5Google
🧑💻 @agamtechtricks
Gemini Nano is a new, 3.8B parameter that is designed to run on-device and can process not just text input but also audio and images. It’s coming later this year “starting with Pixel” and will be used for:
• Clearer descriptions with TalkBack: TalkBack will soon be able to automatically generate more useful image descriptions. This will help people with visual impairments who can’t see images, especially when those images don’t have alt text already.
• Scam detection:
A new feature that processes voice calls to detect and warn you when the person on the other end is trying to scam you. The feature will look for conversation patterns commonly associated with scams, like when a “bank representative” asks you to urgently transfer funds. If so, your phone will buzz and show a warning. This feature is opt-in and will be processed on-device by Gemini Nano.
Please open Telegram to view this post
VIEW IN TELEGRAM
Please open Telegram to view this post
VIEW IN TELEGRAM
👍1🤯1
Gemini for Android is getting upgraded with new contextual capabilities!
• A new “ask this video” chip will appear if you invoke Gemini while a YouTube video is open. Gemini will then use the YouTube video as context.
• A new “ask this PDF” chip will appear when you’re viewing a PDF file in supported apps. Tapping this will let you ask questions about/summarize the document. This requires Gemini Advanced, though, because of its longer context window.
• You’ll be able to drag and drop generated images directly into Gmail, Messages, and other supported apps.
• Finally, devices with the next-gen multimodal Gemini Nano will surface dynamic suggestions. The way this works is that a screenshot of the screen is fed to Gemini Nano, which can then generate chips of potential things you might want to ask about.
ℹ️ Credits: @MishaalAndroidNews
🧑💻 @agamtechtricks
• A new “ask this video” chip will appear if you invoke Gemini while a YouTube video is open. Gemini will then use the YouTube video as context.
• A new “ask this PDF” chip will appear when you’re viewing a PDF file in supported apps. Tapping this will let you ask questions about/summarize the document. This requires Gemini Advanced, though, because of its longer context window.
• You’ll be able to drag and drop generated images directly into Gmail, Messages, and other supported apps.
• Finally, devices with the next-gen multimodal Gemini Nano will surface dynamic suggestions. The way this works is that a screenshot of the screen is fed to Gemini Nano, which can then generate chips of potential things you might want to ask about.
Please open Telegram to view this post
VIEW IN TELEGRAM
Please open Telegram to view this post
VIEW IN TELEGRAM
Gemini Live
In Gemini Live, you’ll get a fullscreen experience with a cool audio waveform effect. This will let you have a 2-way dialogue, with Gemini returning concise responses.
You can speak at your own pace, with Google adapting, and interrupt Gemini as it’s replying to add new information or ask for clarification.
Available for Gemini Advanced subscribers, it’s launching in the coming months.
🔗 9to5Google
🧑💻 @agamtechtricks
In Gemini Live, you’ll get a fullscreen experience with a cool audio waveform effect. This will let you have a 2-way dialogue, with Gemini returning concise responses.
You can speak at your own pace, with Google adapting, and interrupt Gemini as it’s replying to add new information or ask for clarification.
Available for Gemini Advanced subscribers, it’s launching in the coming months.
Please open Telegram to view this post
VIEW IN TELEGRAM
Please open Telegram to view this post
VIEW IN TELEGRAM
🤯5
This media is not supported in your browser
VIEW IN TELEGRAM
Project Astra
The Astra demo Google showed — single-take in real-time — pointed a phone at objects as someone issued commands or questions, with Gemini recognizing what’s in front of it in near real-time. You can show it a cityscape and ask what neighborhood you’re in, or inquire about code.
This is built on the Gemini 1.5 Pro mode and “other task specific models.” Google says it’s “designed to process information faster by continuously encoding video frames,” with reducing the response times to “something conversational” a “difficult engineering challenge.”
🔗 9to5Google
🧑💻 @agamtechtricks
The Astra demo Google showed — single-take in real-time — pointed a phone at objects as someone issued commands or questions, with Gemini recognizing what’s in front of it in near real-time. You can show it a cityscape and ask what neighborhood you’re in, or inquire about code.
This is built on the Gemini 1.5 Pro mode and “other task specific models.” Google says it’s “designed to process information faster by continuously encoding video frames,” with reducing the response times to “something conversational” a “difficult engineering challenge.”
…combining the video and speech input into a timeline of events, and caching this information for efficient recall.”
Please open Telegram to view this post
VIEW IN TELEGRAM
🤯7
More Gemini Updates:
Gemini 1.5 Pro:
Google announced Gemini 1.5 Pro in February and is now launching it in the paid Gemini Advanced subscription.
New Extension:
The new YouTube Music extension lets you search for songs by "mentioning a favorite verse or a featured artist."
Gems:
In the coming months, Gemini Advanced users will be able to create customized versions of Gemini. Examples include a "gym buddy, sous chef, coding partner, etc." All Gemini users will have access to a number of pre-made Gems, like Learning Coach.
Immersive Planner:
Gemini Advanced on the web is getting an “immersive planner” that can create a custom, timeline-based itinerary. Google says this “new planning experience will go beyond showing a list of suggested activities.”
Gemini 1.5 Flash:
Google is introducing 1.5 Flash as its “fastest and most versatile multimodal AI model.” It has the same 1 million context window and is aimed at use cases where low latency and cost matters the most.
🔗 9to5Google
🧑💻 @agamtechtricks
Gemini 1.5 Pro:
Google announced Gemini 1.5 Pro in February and is now launching it in the paid Gemini Advanced subscription.
New Extension:
The new YouTube Music extension lets you search for songs by "mentioning a favorite verse or a featured artist."
Gems:
In the coming months, Gemini Advanced users will be able to create customized versions of Gemini. Examples include a "gym buddy, sous chef, coding partner, etc." All Gemini users will have access to a number of pre-made Gems, like Learning Coach.
Immersive Planner:
Gemini Advanced on the web is getting an “immersive planner” that can create a custom, timeline-based itinerary. Google says this “new planning experience will go beyond showing a list of suggested activities.”
Gemini 1.5 Flash:
Google is introducing 1.5 Flash as its “fastest and most versatile multimodal AI model.” It has the same 1 million context window and is aimed at use cases where low latency and cost matters the most.
Please open Telegram to view this post
VIEW IN TELEGRAM
Please open Telegram to view this post
VIEW IN TELEGRAM
🤯5👍1
Gmail on Android, iOS getting more Gemini: Q&A, better Smart Reply, Summarize
Gmail on Android and iOS is getting a “Summarize this email” feature for longer threads. Workspace Labs users will get Summarize this week, with a launch for Google One AI Premium subscribers and paying Gemini for Workspace customers in June.
Meanwhile, Gmail Q&A will let you go beyond summarizing which will let you enter full prompts. You can ask a question from the e-mail.
Google is building on 2017’s Smart Reply and Smart Compose in 2018 with Contextual Smart Replies. Appearing as a carousel of chips, each suggestion is quickly summarized, like “Proceed & confirm time” or “suggest new time.”
🔗 9to5Google
🧑💻 @agamtechtricks
Gmail on Android and iOS is getting a “Summarize this email” feature for longer threads. Workspace Labs users will get Summarize this week, with a launch for Google One AI Premium subscribers and paying Gemini for Workspace customers in June.
Meanwhile, Gmail Q&A will let you go beyond summarizing which will let you enter full prompts. You can ask a question from the e-mail.
Google is building on 2017’s Smart Reply and Smart Compose in 2018 with Contextual Smart Replies. Appearing as a carousel of chips, each suggestion is quickly summarized, like “Proceed & confirm time” or “suggest new time.”
Please open Telegram to view this post
VIEW IN TELEGRAM
Please open Telegram to view this post
VIEW IN TELEGRAM
🤯1
Gemini 1.5 Pro-powered side panel launching in Gmail, Google Docs, and more
Available in Gmail Docs, Sheets, Slides, and Drive, the side panel is now powered by Gemini 1.5 Pro. A larger context window allows for more information to be analyzed, while there’s also more advanced reasoning.
it’s coming to users enrolled in Workspace Labs and the Gemini for Workspace Alpha. All paying Gemini for Workspace customers and Google One AI Premium subscribers will get it next month.
🔗 9to5Google
🧑💻 @agamtechtricks
Available in Gmail Docs, Sheets, Slides, and Drive, the side panel is now powered by Gemini 1.5 Pro. A larger context window allows for more information to be analyzed, while there’s also more advanced reasoning.
For example you can ask it to summarize emails from your child’s school or even to highlight the main points from a recording of a PTA meeting you missed.
it’s coming to users enrolled in Workspace Labs and the Gemini for Workspace Alpha. All paying Gemini for Workspace customers and Google One AI Premium subscribers will get it next month.
Please open Telegram to view this post
VIEW IN TELEGRAM
🤯2
Gemini in Android Studio is getting some new features with Android Studio Koala!
• You can now provide custom prompts to generate a code suggestion that either adds new code or transforms selected code. You can ask Gemini to simplify complex code by rewriting it, perform specific code transformations like “make this code idiomatic”, or generate new functions you describe. Android Studio will show you Gemini’s code suggestion as a code diff you can review.
• It can now analyze your crash reports, generate insights that are shown in the Gemini tool window, provide a crash summary, and sometimes recommend next steps like sample code and links to relevant documentation.
• Later this year, the underlying Gemini model will be upgraded to Gemini 1.5 Pro, which offers a much larger context window and multimodal input.
ℹ️ Credits: @MishaalAndroidNews
🧑💻 @agamtechtricks
• You can now provide custom prompts to generate a code suggestion that either adds new code or transforms selected code. You can ask Gemini to simplify complex code by rewriting it, perform specific code transformations like “make this code idiomatic”, or generate new functions you describe. Android Studio will show you Gemini’s code suggestion as a code diff you can review.
• It can now analyze your crash reports, generate insights that are shown in the Gemini tool window, provide a crash summary, and sometimes recommend next steps like sample code and links to relevant documentation.
• Later this year, the underlying Gemini model will be upgraded to Gemini 1.5 Pro, which offers a much larger context window and multimodal input.
Please open Telegram to view this post
VIEW IN TELEGRAM
Please open Telegram to view this post
VIEW IN TELEGRAM
👍3👎1🤯1