<?xml version="1.0" encoding="UTF-8"?>
<rss version="2.0" xmlns:atom="http://www.w3.org/2005/Atom" xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:googleplay="http://www.google.com/schemas/play-podcasts/1.0" xmlns:itunes="http://www.itunes.com/dtds/podcast-1.0.dtd" xmlns:media="http://search.yahoo.com/mrss/" xmlns:podcast="https://podcastindex.org/namespace/1.0">
  <channel>
    <atom:link href="https://feeds.simplecast.com/niac4zV1" rel="self" title="MP3 Audio" type="application/atom+xml"/>
    <atom:link href="https://simplecast.superfeedr.com" rel="hub" xmlns="http://www.w3.org/2005/Atom"/>
    <generator>https://simplecast.com</generator>
    <title>Google AI: Release Notes</title>
    <description>Ever wondered what it&apos;s really like to build the future of AI? Join host Logan Kilpatrick for a deep dive into the world of Google AI, straight from the minds of the builders. We&apos;re pulling back the curtain on the latest breakthroughs, sharing the unfiltered stories behind the tech, and answering the questions you&apos;ve been dying to ask.

Whether you&apos;re a seasoned developer or an AI enthusiast, this podcast is your backstage pass to the cutting-edge of AI technology. Tune in for:

- Exclusive interviews with AI pioneers and industry leaders.
- In-depth discussions on the latest AI trends and developments.
- Behind-the-scenes stories and anecdotes from the world of AI.
- Unfiltered insights and opinions from the people shaping the future.

So, if you&apos;re ready to go beyond the headlines and get the real scoop on AI, join Logan Kilpatrick on Google AI: Release Notes.</description>
    <copyright>2024 Google</copyright>
    <language>en</language>
    <pubDate>Thu, 12 Mar 2026 21:48:45 +0000</pubDate>
    <lastBuildDate>Thu, 12 Mar 2026 21:48:55 +0000</lastBuildDate>
    <image>
      <link>https://google-ai-release-notes.simplecast.com</link>
      <title>Google AI: Release Notes</title>
      <url>https://image.simplecastcdn.com/images/95190ff2-3221-4589-bc23-6d49f11be41f/b9555480-c416-4b1b-a5fa-3cd3fd92e9c5/3000x3000/rnp-logo.jpg?aid=rss_feed</url>
    </image>
    <link>https://google-ai-release-notes.simplecast.com</link>
    <itunes:type>episodic</itunes:type>
    <itunes:summary>Ever wondered what it&apos;s really like to build the future of AI? Join host Logan Kilpatrick for a deep dive into the world of Google AI, straight from the minds of the builders. We&apos;re pulling back the curtain on the latest breakthroughs, sharing the unfiltered stories behind the tech, and answering the questions you&apos;ve been dying to ask.

Whether you&apos;re a seasoned developer or an AI enthusiast, this podcast is your backstage pass to the cutting-edge of AI technology. Tune in for:

- Exclusive interviews with AI pioneers and industry leaders.
- In-depth discussions on the latest AI trends and developments.
- Behind-the-scenes stories and anecdotes from the world of AI.
- Unfiltered insights and opinions from the people shaping the future.

So, if you&apos;re ready to go beyond the headlines and get the real scoop on AI, join Logan Kilpatrick on Google AI: Release Notes.</itunes:summary>
    <itunes:author>Google AI</itunes:author>
    <itunes:explicit>false</itunes:explicit>
    <itunes:image href="https://image.simplecastcdn.com/images/95190ff2-3221-4589-bc23-6d49f11be41f/b9555480-c416-4b1b-a5fa-3cd3fd92e9c5/3000x3000/rnp-logo.jpg?aid=rss_feed"/>
    <itunes:new-feed-url>https://feeds.simplecast.com/niac4zV1</itunes:new-feed-url>
    <itunes:owner>
      <itunes:name>Google</itunes:name>
      <itunes:email>google-ai-podcast-release-notes@google.com</itunes:email>
    </itunes:owner>
    <itunes:category text="Technology"/>
    <itunes:category text="Science"/>
    <item>
      <guid isPermaLink="false">269e2f7b-0b2e-4105-a58c-f629619e1f18</guid>
      <title>Google Maps Leaders Talk About Its Biggest Update in 10 Years</title>
      <description><![CDATA[<p>Miriam Daniel and David Cronin join host Logan Kilpatrick to unveil the biggest update in Google Maps history. This episode explores the launch of Ask Maps, a conversational experience powered by Gemini 3.0 Pro, and Immersive Navigation, a high-fidelity 3D driving experience. Learn how Gemini is making navigation more intuitive, personalized, and stress-free.</p>
<p>Watch on YouTube: https://www.youtube.com/watch?v=YzZFdHzx-Y4</p>
]]></description>
      <pubDate>Thu, 12 Mar 2026 21:48:45 +0000</pubDate>
      <author>google-ai-podcast-release-notes@google.com (Google)</author>
      <link>https://google-ai-release-notes.simplecast.com/episodes/google-maps-leaders-talk-about-its-biggest-update-in-10-years-M_I_dRoD</link>
      <content:encoded><![CDATA[<p>Miriam Daniel and David Cronin join host Logan Kilpatrick to unveil the biggest update in Google Maps history. This episode explores the launch of Ask Maps, a conversational experience powered by Gemini 3.0 Pro, and Immersive Navigation, a high-fidelity 3D driving experience. Learn how Gemini is making navigation more intuitive, personalized, and stress-free.</p>
<p>Watch on YouTube: https://www.youtube.com/watch?v=YzZFdHzx-Y4</p>
]]></content:encoded>
      <enclosure length="34615656" type="audio/mpeg" url="https://cdn.simplecast.com/media/audio/transcoded/8d1b89f9-c362-4126-8d8a-b028e80680d5/c68ffadd-1655-4420-a3c1-6df8dd507951/episodes/audio/group/b9637e84-6834-498d-bf98-81a198bd6e16/group-item/3ef0f1dc-a18b-46c8-bc25-16941eb6da37/128_default_tc.mp3?aid=rss_feed&amp;feed=niac4zV1"/>
      <itunes:title>Google Maps Leaders Talk About Its Biggest Update in 10 Years</itunes:title>
      <itunes:author>Google</itunes:author>
      <itunes:duration>00:36:03</itunes:duration>
      <itunes:summary></itunes:summary>
      <itunes:subtitle></itunes:subtitle>
      <itunes:explicit>false</itunes:explicit>
      <itunes:episodeType>full</itunes:episodeType>
      <itunes:episode>27</itunes:episode>
    </item>
    <item>
      <guid isPermaLink="false">e9644482-008a-4348-9e3d-583605e2de0e</guid>
      <title>Gemini in Workspace: New Ways to Create Faster</title>
      <description><![CDATA[<p>Chapters:<br><br>
 1:15 - Gemini in Docs<br>
 3:17 - Which models power Workspace<br>
 3:45 - AI Overviews in Drive<br>
 5:22 - Rollout and availability<br>
 6:33 - Reimagining every Workspace canvas<br>
 8:58 - Gemini in Calendar<br>
 9:50 - The future of the side panel<br>
 11:16 - A new way to work<br>
 13:18 - AI-powered slide generation<br>
 15:08 - User data and privacy<br>
 19:07 - Balancing AI innovation and user trust<br>
 22:04 - The power of Google Vids<br>
 24:42 - The journey to deep canvas integration<br>
 28:32 - Vibe coding meets Apps Script<br>
 32:42 - AI as a tool for thinking</p>
]]></description>
      <pubDate>Thu, 12 Mar 2026 21:38:45 +0000</pubDate>
      <author>google-ai-podcast-release-notes@google.com (Google)</author>
      <link>https://google-ai-release-notes.simplecast.com/episodes/gemini-in-workspace-new-ways-to-create-faster-qRvYkgg_</link>
      <content:encoded><![CDATA[<p>Chapters:<br><br>
 1:15 - Gemini in Docs<br>
 3:17 - Which models power Workspace<br>
 3:45 - AI Overviews in Drive<br>
 5:22 - Rollout and availability<br>
 6:33 - Reimagining every Workspace canvas<br>
 8:58 - Gemini in Calendar<br>
 9:50 - The future of the side panel<br>
 11:16 - A new way to work<br>
 13:18 - AI-powered slide generation<br>
 15:08 - User data and privacy<br>
 19:07 - Balancing AI innovation and user trust<br>
 22:04 - The power of Google Vids<br>
 24:42 - The journey to deep canvas integration<br>
 28:32 - Vibe coding meets Apps Script<br>
 32:42 - AI as a tool for thinking</p>
]]></content:encoded>
      <enclosure length="38600897" type="audio/mpeg" url="https://cdn.simplecast.com/media/audio/transcoded/8d1b89f9-c362-4126-8d8a-b028e80680d5/c68ffadd-1655-4420-a3c1-6df8dd507951/episodes/audio/group/7703263d-73a1-46ca-85b4-8b12f8ad4250/group-item/74cf9b16-36c0-41e6-bb77-130c45ff0190/128_default_tc.mp3?aid=rss_feed&amp;feed=niac4zV1"/>
      <itunes:title>Gemini in Workspace: New Ways to Create Faster</itunes:title>
      <itunes:author>Google</itunes:author>
      <itunes:duration>00:40:12</itunes:duration>
      <itunes:summary>Google Workspace VP of Engineering Chandu Thota joins host Logan Kilpatrick for a deep dive into the all-new Gemini experience in Docs, Sheets, Slides and Drive. Gemini can now pull relevant information from your files, emails and the web to help you go from a blank page to a final draft, build entire spreadsheets and create presentation-ready slides. Their conversation covers the engineering journey behind deep canvas integration, user data privacy, and how new personal semantic indexes power features like AI Overviews in Drive.

Watch on YouTube: https://www.youtube.com/watch?v=8wtKHwlKMBQ</itunes:summary>
      <itunes:subtitle>Google Workspace VP of Engineering Chandu Thota joins host Logan Kilpatrick for a deep dive into the all-new Gemini experience in Docs, Sheets, Slides and Drive. Gemini can now pull relevant information from your files, emails and the web to help you go from a blank page to a final draft, build entire spreadsheets and create presentation-ready slides. Their conversation covers the engineering journey behind deep canvas integration, user data privacy, and how new personal semantic indexes power features like AI Overviews in Drive.

Watch on YouTube: https://www.youtube.com/watch?v=8wtKHwlKMBQ</itunes:subtitle>
      <itunes:explicit>false</itunes:explicit>
      <itunes:episodeType>full</itunes:episodeType>
      <itunes:episode>26</itunes:episode>
    </item>
    <item>
      <guid isPermaLink="false">da9de9f0-f9fa-43ea-84af-2203de6fc7d4</guid>
      <title>Gemini in Chrome: Your agentic browsing assistant</title>
      <description><![CDATA[<p>Chapters:<br><br>
 0:00 - Introduction<br>
 2:49 - Evolution from web apps to integrated assistants<br>
 4:37 - Chrome as a platform for personal context<br>
 6:38 - Navigating the context overload problem<br>
 7:52 - Transforming media in-context with Nano Banana<br>
 9:10 - Solving tab overload with history recall<br>
 13:28 - The browser as an automated workflow system<br>
 15:50 - Demo: Nano Banana<br>
 17:20 - Demo: Auto browse<br>
 22:48 - Demo: Agentic research and guardrails<br>
 26:04 - Designing for billions<br>
 29:14 - Transitioning to agentic web actuation<br>
 30:37 - Standards and security in an AI-driven web<br>
 35:18 - Infrastructure and investment strategy<br>
 39:23 - Empowering knowledge workers<br>
 42:11 - Collaboration within Google<br>
 44:18 - Safety and the user alignment critic</p>
]]></description>
      <pubDate>Thu, 12 Mar 2026 21:36:09 +0000</pubDate>
      <author>google-ai-podcast-release-notes@google.com (Google)</author>
      <link>https://google-ai-release-notes.simplecast.com/episodes/gemini-in-chrome-your-agentic-browsing-assistant-L2qv6wiW</link>
      <content:encoded><![CDATA[<p>Chapters:<br><br>
 0:00 - Introduction<br>
 2:49 - Evolution from web apps to integrated assistants<br>
 4:37 - Chrome as a platform for personal context<br>
 6:38 - Navigating the context overload problem<br>
 7:52 - Transforming media in-context with Nano Banana<br>
 9:10 - Solving tab overload with history recall<br>
 13:28 - The browser as an automated workflow system<br>
 15:50 - Demo: Nano Banana<br>
 17:20 - Demo: Auto browse<br>
 22:48 - Demo: Agentic research and guardrails<br>
 26:04 - Designing for billions<br>
 29:14 - Transitioning to agentic web actuation<br>
 30:37 - Standards and security in an AI-driven web<br>
 35:18 - Infrastructure and investment strategy<br>
 39:23 - Empowering knowledge workers<br>
 42:11 - Collaboration within Google<br>
 44:18 - Safety and the user alignment critic</p>
]]></content:encoded>
      <enclosure length="46593112" type="audio/mpeg" url="https://cdn.simplecast.com/media/audio/transcoded/8d1b89f9-c362-4126-8d8a-b028e80680d5/c68ffadd-1655-4420-a3c1-6df8dd507951/episodes/audio/group/88cf3912-00a7-4610-ba94-c20c2d4df55e/group-item/d5688c41-92b7-4dbe-be9f-2baa0b824785/128_default_tc.mp3?aid=rss_feed&amp;feed=niac4zV1"/>
      <itunes:title>Gemini in Chrome: Your agentic browsing assistant</itunes:title>
      <itunes:author>Google</itunes:author>
      <itunes:duration>00:48:32</itunes:duration>
      <itunes:summary>Rick Osterloh, SVP of Platforms &amp; Devices, and Parisa Tabriz, VP of Chrome, join host Logan Kilpatrick for a deep dive into the launch of Gemini in Chrome. They discuss the browser&apos;s evolution from a passive window to an active agentic assistant. Learn more about Auto Browse for automating web tasks, on-device creativity with Nano Banana, solving tab overload with History Recall, and the safety guardrails built with the User Alignment Critic.

Watch on YouTube: https://www.youtube.com/watch?v=5OR4c87Xt-E</itunes:summary>
      <itunes:subtitle>Rick Osterloh, SVP of Platforms &amp; Devices, and Parisa Tabriz, VP of Chrome, join host Logan Kilpatrick for a deep dive into the launch of Gemini in Chrome. They discuss the browser&apos;s evolution from a passive window to an active agentic assistant. Learn more about Auto Browse for automating web tasks, on-device creativity with Nano Banana, solving tab overload with History Recall, and the safety guardrails built with the User Alignment Critic.

Watch on YouTube: https://www.youtube.com/watch?v=5OR4c87Xt-E</itunes:subtitle>
      <itunes:explicit>false</itunes:explicit>
      <itunes:episodeType>full</itunes:episodeType>
      <itunes:episode>26</itunes:episode>
    </item>
    <item>
      <guid isPermaLink="false">f2a23fd9-0a40-486f-a529-4b6a1830e168</guid>
      <title>Inside Lyria 3, Google&apos;s music generation model</title>
      <description><![CDATA[<p>1:00 - Defining music generation models<br />1:40 - Lyria as a new instrument<br />3:05 - Connecting language and creative intent<br />5:08 - Guest backgrounds and musical journeys<br />7:57 - Demo: Instrumental funk jam<br />8:29 - Bridging the gap for non-musicians<br />12:03 - Demo: Exploring lyrics and vocals<br />15:07 - The magic of iterative co-creation<br />15:40 - Meeting users across the expertise spectrum<br />17:01 - Empowering new musical expressions<br />18:29 - Emotional and communal impact of music<br />19:51 - Opportunities for developers and community<br />21:09 - Real-time vs. song generation models<br />23:23 - Creating experimental sonic landscapes<br />25:08 - Demo: Capturing unexpectedness and energy<br />28:33 - Evaluating music through taste and expertise<br />31:30 - The diligence of music evaluation<br />31:52 - The future of Lyria and AI-first workflows<br />35:07 - Articulating creative vision through languag</p>
]]></description>
      <pubDate>Wed, 18 Feb 2026 20:00:00 +0000</pubDate>
      <author>google-ai-podcast-release-notes@google.com (Google)</author>
      <link>https://google-ai-release-notes.simplecast.com/episodes/inside-lyria-3-googles-music-generation-model-xNnOLgIN</link>
      <content:encoded><![CDATA[<p>1:00 - Defining music generation models<br />1:40 - Lyria as a new instrument<br />3:05 - Connecting language and creative intent<br />5:08 - Guest backgrounds and musical journeys<br />7:57 - Demo: Instrumental funk jam<br />8:29 - Bridging the gap for non-musicians<br />12:03 - Demo: Exploring lyrics and vocals<br />15:07 - The magic of iterative co-creation<br />15:40 - Meeting users across the expertise spectrum<br />17:01 - Empowering new musical expressions<br />18:29 - Emotional and communal impact of music<br />19:51 - Opportunities for developers and community<br />21:09 - Real-time vs. song generation models<br />23:23 - Creating experimental sonic landscapes<br />25:08 - Demo: Capturing unexpectedness and energy<br />28:33 - Evaluating music through taste and expertise<br />31:30 - The diligence of music evaluation<br />31:52 - The future of Lyria and AI-first workflows<br />35:07 - Articulating creative vision through languag</p>
]]></content:encoded>
      <enclosure length="35173214" type="audio/mpeg" url="https://cdn.simplecast.com/audio/c68ffadd-1655-4420-a3c1-6df8dd507951/episodes/a97f1cb1-01c2-44e5-a473-cf08720564bc/audio/edefe902-b569-4cf5-a7f8-19dd11047af7/default_tc.mp3?aid=rss_feed&amp;feed=niac4zV1"/>
      <itunes:title>Inside Lyria 3, Google&apos;s music generation model</itunes:title>
      <itunes:author>Google</itunes:author>
      <itunes:duration>00:36:38</itunes:duration>
      <itunes:summary>Jeff Chang, Myriam Hamed Torres, and Jason Baldridge from the Google DeepMind team join host Logan Kilpatrick for a deep dive into Lyria 3, Google’s latest music generation model. Their conversation explores the transition from simple audio generation to a model that acts as a collaborative instrument, providing creators with fine-grained control over mood, instrumentation, and vocals. Learn more about the technical challenges of prompt adherence in music, the importance of &quot;vibe&quot; in human evaluations, and the future of layered, iterative music composition.

Watch on YouTube: https://www.youtube.com/watch?v=mQUpaaU02DM</itunes:summary>
      <itunes:subtitle>Jeff Chang, Myriam Hamed Torres, and Jason Baldridge from the Google DeepMind team join host Logan Kilpatrick for a deep dive into Lyria 3, Google’s latest music generation model. Their conversation explores the transition from simple audio generation to a model that acts as a collaborative instrument, providing creators with fine-grained control over mood, instrumentation, and vocals. Learn more about the technical challenges of prompt adherence in music, the importance of &quot;vibe&quot; in human evaluations, and the future of layered, iterative music composition.

Watch on YouTube: https://www.youtube.com/watch?v=mQUpaaU02DM</itunes:subtitle>
      <itunes:explicit>false</itunes:explicit>
      <itunes:episodeType>full</itunes:episodeType>
      <itunes:episode>25</itunes:episode>
    </item>
    <item>
      <guid isPermaLink="false">0c40e5ef-d79f-4cc5-b5c5-fb80c694965b</guid>
      <title>Project Genie: Create and explore worlds</title>
      <description><![CDATA[<p><strong>Chapters:</strong><br />00:00 - Intro and defining world models and RL roots<br />01:51 - Demo: Goldfish and shark in underwater world<br />04:59 - Project Genie gallery<br />06:31 - Physics, remixing, and UI prompts<br />11:00 - Demo: Nano Banana mascot “Bob”<br />13:20 - Constraints, generation limits, and infrastructure<br />17:04 - Trusted testers and robotics future<br />28:34 - Frontier prompting and universal simulation<br />29:27 - Cross-Google collaboration<br />31:16 - Adoption timelines and impact<br />34:16 - Model generalization and historical context<br />38:52 - Hardware limits and the slope of progress</p>
]]></description>
      <pubDate>Fri, 30 Jan 2026 17:45:09 +0000</pubDate>
      <author>google-ai-podcast-release-notes@google.com (Google)</author>
      <link>https://google-ai-release-notes.simplecast.com/episodes/project-genie-create-and-explore-worlds-_52W_tRH</link>
      <content:encoded><![CDATA[<p><strong>Chapters:</strong><br />00:00 - Intro and defining world models and RL roots<br />01:51 - Demo: Goldfish and shark in underwater world<br />04:59 - Project Genie gallery<br />06:31 - Physics, remixing, and UI prompts<br />11:00 - Demo: Nano Banana mascot “Bob”<br />13:20 - Constraints, generation limits, and infrastructure<br />17:04 - Trusted testers and robotics future<br />28:34 - Frontier prompting and universal simulation<br />29:27 - Cross-Google collaboration<br />31:16 - Adoption timelines and impact<br />34:16 - Model generalization and historical context<br />38:52 - Hardware limits and the slope of progress</p>
]]></content:encoded>
      <enclosure length="40829037" type="audio/mpeg" url="https://cdn.simplecast.com/audio/c68ffadd-1655-4420-a3c1-6df8dd507951/episodes/dfe2e6e9-dfdb-47a1-8021-4acf953c35db/audio/ed97fe8d-1c33-4535-8ebb-ced52ae6732a/default_tc.mp3?aid=rss_feed&amp;feed=niac4zV1"/>
      <itunes:title>Project Genie: Create and explore worlds</itunes:title>
      <itunes:author>Google</itunes:author>
      <itunes:duration>00:42:31</itunes:duration>
      <itunes:summary>Diego Rivas, Shlomi Fruchter, and Jack Parker-Holder from the Project Genie team join host Logan Kilpatrick for an in-depth discussion on Google DeepMind’s latest breakthrough in world models. Project Genie is an experimental research prototype that allows users to generate, explore, and interact with infinitely diverse, photorealistic worlds in real-time. Learn more about the shift from passive video generation to interactive media, the technical challenges of maintaining world consistency and memory, and how these models serve as an essential training ground for AI agents.

Watch on YouTube: https://www.youtube.com/watch?v=Ow0W3WlJxRY</itunes:summary>
      <itunes:subtitle>Diego Rivas, Shlomi Fruchter, and Jack Parker-Holder from the Project Genie team join host Logan Kilpatrick for an in-depth discussion on Google DeepMind’s latest breakthrough in world models. Project Genie is an experimental research prototype that allows users to generate, explore, and interact with infinitely diverse, photorealistic worlds in real-time. Learn more about the shift from passive video generation to interactive media, the technical challenges of maintaining world consistency and memory, and how these models serve as an essential training ground for AI agents.

Watch on YouTube: https://www.youtube.com/watch?v=Ow0W3WlJxRY</itunes:subtitle>
      <itunes:explicit>false</itunes:explicit>
      <itunes:episodeType>full</itunes:episodeType>
      <itunes:episode>24</itunes:episode>
    </item>
    <item>
      <guid isPermaLink="false">91b818e6-b496-4f04-a9d3-403b05210b4b</guid>
      <title>Gemini 3 and Gen UI in Google Search</title>
      <description><![CDATA[<p>Rhiannon Bell and Robby Stein, Product and Design leads for Google Search, join host Logan Kilpatrick for a deep dive into the integration of Gemini 3 into Search. Their conversation explores the evolution of Generative UI, where models act as designers to create bespoke, interactive simulations on the fly. Learn more about the role of Gemini 3 Flash in delivering speed at scale, the development of Search's new "persona," and how models like Nano Banana are powering next-generation data visualization.</p><p>Watch on YouTube: https://www.youtube.com/watch?v=AqyclkRBSe4</p><p><strong>Chapters:</strong><br />0:00 - Introduction<br />1:24 - What is Generative UI?<br />2:23 - From static to generative design<br />6:37 - Interactive simulations<br />8:47 - Latency and visual QA<br />10:48 - Gemini 3 Flash in Search<br />12:08 - Fusing AI Mode and AI Overviews<br />14:24 - The Search persona<br />17:12 - Agentic system understanding<br />18:22 - Visualizing data with Nano Banana</p><p> </p>
]]></description>
      <pubDate>Thu, 18 Dec 2025 21:44:43 +0000</pubDate>
      <author>google-ai-podcast-release-notes@google.com (Google)</author>
      <link>https://google-ai-release-notes.simplecast.com/episodes/gemini-3-and-gen-ui-in-google-search-PQhbsQbN</link>
      <content:encoded><![CDATA[<p>Rhiannon Bell and Robby Stein, Product and Design leads for Google Search, join host Logan Kilpatrick for a deep dive into the integration of Gemini 3 into Search. Their conversation explores the evolution of Generative UI, where models act as designers to create bespoke, interactive simulations on the fly. Learn more about the role of Gemini 3 Flash in delivering speed at scale, the development of Search's new "persona," and how models like Nano Banana are powering next-generation data visualization.</p><p>Watch on YouTube: https://www.youtube.com/watch?v=AqyclkRBSe4</p><p><strong>Chapters:</strong><br />0:00 - Introduction<br />1:24 - What is Generative UI?<br />2:23 - From static to generative design<br />6:37 - Interactive simulations<br />8:47 - Latency and visual QA<br />10:48 - Gemini 3 Flash in Search<br />12:08 - Fusing AI Mode and AI Overviews<br />14:24 - The Search persona<br />17:12 - Agentic system understanding<br />18:22 - Visualizing data with Nano Banana</p><p> </p>
]]></content:encoded>
      <enclosure length="20868561" type="audio/mpeg" url="https://cdn.simplecast.com/audio/c68ffadd-1655-4420-a3c1-6df8dd507951/episodes/ef5c3ebf-469f-4e77-a6d0-0cee88182cde/audio/c5fe31bd-5841-411d-bffd-04f937c9bd3b/default_tc.mp3?aid=rss_feed&amp;feed=niac4zV1"/>
      <itunes:title>Gemini 3 and Gen UI in Google Search</itunes:title>
      <itunes:author>Google</itunes:author>
      <itunes:duration>00:21:44</itunes:duration>
      <itunes:summary>Rhiannon Bell and Robby Stein, Product and Design leads for Google Search, join host Logan Kilpatrick for a deep dive into the integration of Gemini 3 into Search. Their conversation explores the evolution of Generative UI, where models act as designers to create bespoke, interactive simulations on the fly. Learn more about the role of Gemini 3 Flash in delivering speed at scale, the development of Search&apos;s new &quot;persona,&quot; and how models like Nano Banana are powering next-generation data visualization.</itunes:summary>
      <itunes:subtitle>Rhiannon Bell and Robby Stein, Product and Design leads for Google Search, join host Logan Kilpatrick for a deep dive into the integration of Gemini 3 into Search. Their conversation explores the evolution of Generative UI, where models act as designers to create bespoke, interactive simulations on the fly. Learn more about the role of Gemini 3 Flash in delivering speed at scale, the development of Search&apos;s new &quot;persona,&quot; and how models like Nano Banana are powering next-generation data visualization.</itunes:subtitle>
      <itunes:explicit>false</itunes:explicit>
      <itunes:episodeType>full</itunes:episodeType>
      <itunes:episode>22</itunes:episode>
    </item>
    <item>
      <guid isPermaLink="false">6942404f-d92b-4cf5-94f6-40e45310bbb3</guid>
      <title>Sundar Pichai: Gemini 3, Vibe Coding and Google&apos;s Full Stack Strategy</title>
      <description><![CDATA[<p>Logan Kilpatrick from Google DeepMind sits down with Sundar Pichai, CEO of Google and Alphabet to discuss the launch of Gemini 3, Nano Banana Pro and Google's overall AI momentum. They talk about Google’s long-term bets on infrastructure, what it’s actually like to ship SOTA models, and the rise of vibe coding. Sundar also shares his personal launch day rituals and thoughts on future moonshots like putting data centers in space.</p><p>Watch on YouTube: https://www.youtube.com/watch?v=iFqDyWFuw1c<br /><br /><strong>Chapters:</strong><br />0:00 - Intro<br />0:51 - Shipping Gemini 3<br />2:44 - Google's decade-long investment in AI<br />4:27 - The full stack advantage<br />5:43 - Scaling up compute and capacity<br />7:32 - Sim-shipping Gemini across products<br />9:35 - Nano Banana Pro<br />12:13 - Monitoring launch day<br />14:13 - Future model roadmap<br />16:05 - Launch day rituals<br />18:02 - The Blue Micro Kitchen<br />21:57 - Future moonshots<br />23:26 - The rise of vibe coding<br />26:50 - What’s next</p>
]]></description>
      <pubDate>Wed, 26 Nov 2025 22:26:20 +0000</pubDate>
      <author>google-ai-podcast-release-notes@google.com (Google)</author>
      <link>https://google-ai-release-notes.simplecast.com/episodes/sundar-pichai-gemini-3-vibe-coding-and-googles-full-stack-strategy-OFkDXDVG</link>
      <content:encoded><![CDATA[<p>Logan Kilpatrick from Google DeepMind sits down with Sundar Pichai, CEO of Google and Alphabet to discuss the launch of Gemini 3, Nano Banana Pro and Google's overall AI momentum. They talk about Google’s long-term bets on infrastructure, what it’s actually like to ship SOTA models, and the rise of vibe coding. Sundar also shares his personal launch day rituals and thoughts on future moonshots like putting data centers in space.</p><p>Watch on YouTube: https://www.youtube.com/watch?v=iFqDyWFuw1c<br /><br /><strong>Chapters:</strong><br />0:00 - Intro<br />0:51 - Shipping Gemini 3<br />2:44 - Google's decade-long investment in AI<br />4:27 - The full stack advantage<br />5:43 - Scaling up compute and capacity<br />7:32 - Sim-shipping Gemini across products<br />9:35 - Nano Banana Pro<br />12:13 - Monitoring launch day<br />14:13 - Future model roadmap<br />16:05 - Launch day rituals<br />18:02 - The Blue Micro Kitchen<br />21:57 - Future moonshots<br />23:26 - The rise of vibe coding<br />26:50 - What’s next</p>
]]></content:encoded>
      <enclosure length="26478827" type="audio/mpeg" url="https://cdn.simplecast.com/audio/c68ffadd-1655-4420-a3c1-6df8dd507951/episodes/49d9e29b-f4b0-46c0-9fac-80781c13b1ab/audio/1b4cd561-6c5d-4267-a7e9-841208c95a7e/default_tc.mp3?aid=rss_feed&amp;feed=niac4zV1"/>
      <itunes:title>Sundar Pichai: Gemini 3, Vibe Coding and Google&apos;s Full Stack Strategy</itunes:title>
      <itunes:author>Google</itunes:author>
      <itunes:image href="https://image.simplecastcdn.com/images/95190ff2-3221-4589-bc23-6d49f11be41f/a75abd17-40e5-4475-b4fc-b2261c41d976/3000x3000/rnp-logo.jpg?aid=rss_feed"/>
      <itunes:duration>00:27:34</itunes:duration>
      <itunes:summary>Logan Kilpatrick from Google DeepMind sits down with Sundar Pichai, CEO of Google and Alphabet to discuss the launch of Gemini 3, Nano Banana Pro and Google&apos;s overall AI momentum. They talk about Google’s long-term bets on infrastructure, what it’s actually like to ship SOTA models, and the rise of vibe coding. Sundar also shares his personal launch day rituals and thoughts on future moonshots like putting data centers in space.</itunes:summary>
      <itunes:subtitle>Logan Kilpatrick from Google DeepMind sits down with Sundar Pichai, CEO of Google and Alphabet to discuss the launch of Gemini 3, Nano Banana Pro and Google&apos;s overall AI momentum. They talk about Google’s long-term bets on infrastructure, what it’s actually like to ship SOTA models, and the rise of vibe coding. Sundar also shares his personal launch day rituals and thoughts on future moonshots like putting data centers in space.</itunes:subtitle>
      <itunes:explicit>false</itunes:explicit>
      <itunes:episodeType>full</itunes:episodeType>
      <itunes:episode>21</itunes:episode>
    </item>
    <item>
      <guid isPermaLink="false">13c99014-8008-4025-95e3-5e21fe92a516</guid>
      <title>Nano Banana Pro: Hands-on with the World’s Most Powerful Image Model</title>
      <description><![CDATA[<p>Introducing Nano Banana Pro, a powerful model built on Gemini 3 Pro, designed to enhance text rendering, infographics, and structured content generation. Tune in to learn about Nano Banana Pro’s advanced visual reasoning and multi-turn generation capabilities, and how this next-gen tool enables complex image edits and real-world applications. In this episode, we discuss how user feedback and continuous benchmarking drive model improvements, ensuring a superior experience for developers.</p><p>Watch on YouTube: https://www.youtube.com/watch?v=hk6gwiZmSWA<br /><br /><strong>Chapters:</strong><br />00:00 - Introducing Nano Banana Pro<br />02:00 - Enhanced world understanding<br />04:59 - Advanced text rendering<br />05:49 - Gemini 3 Pro's influence<br />09:30 - Multi-turn & infographics<br />14:04 - Text rendering comparison<br />16:26 - Multilingual text support<br />18:22 - Infographics for learning<br />24:00 - Multi-image input<br />26:38 - Resolution & fidelity<br />30:07 - Advanced editing & style<br />32:09 - Practical use cases<br />35:26 - Future outlook & thanks</p>
]]></description>
      <pubDate>Wed, 26 Nov 2025 22:19:51 +0000</pubDate>
      <author>google-ai-podcast-release-notes@google.com (Google)</author>
      <link>https://google-ai-release-notes.simplecast.com/episodes/nano-banana-pro-hands-on-with-the-worlds-most-powerful-image-model-EZxl7d1V</link>
      <content:encoded><![CDATA[<p>Introducing Nano Banana Pro, a powerful model built on Gemini 3 Pro, designed to enhance text rendering, infographics, and structured content generation. Tune in to learn about Nano Banana Pro’s advanced visual reasoning and multi-turn generation capabilities, and how this next-gen tool enables complex image edits and real-world applications. In this episode, we discuss how user feedback and continuous benchmarking drive model improvements, ensuring a superior experience for developers.</p><p>Watch on YouTube: https://www.youtube.com/watch?v=hk6gwiZmSWA<br /><br /><strong>Chapters:</strong><br />00:00 - Introducing Nano Banana Pro<br />02:00 - Enhanced world understanding<br />04:59 - Advanced text rendering<br />05:49 - Gemini 3 Pro's influence<br />09:30 - Multi-turn & infographics<br />14:04 - Text rendering comparison<br />16:26 - Multilingual text support<br />18:22 - Infographics for learning<br />24:00 - Multi-image input<br />26:38 - Resolution & fidelity<br />30:07 - Advanced editing & style<br />32:09 - Practical use cases<br />35:26 - Future outlook & thanks</p>
]]></content:encoded>
      <enclosure length="34948770" type="audio/mpeg" url="https://cdn.simplecast.com/audio/c68ffadd-1655-4420-a3c1-6df8dd507951/episodes/c868b2b7-a544-4fe8-be61-4afdb7e04490/audio/2903b099-82c0-4250-9747-3268e03f12da/default_tc.mp3?aid=rss_feed&amp;feed=niac4zV1"/>
      <itunes:title>Nano Banana Pro: Hands-on with the World’s Most Powerful Image Model</itunes:title>
      <itunes:author>Google</itunes:author>
      <itunes:image href="https://image.simplecastcdn.com/images/95190ff2-3221-4589-bc23-6d49f11be41f/f9c63fc1-a049-428d-82eb-c0552adc0a92/3000x3000/rnp-logo.jpg?aid=rss_feed"/>
      <itunes:duration>00:36:24</itunes:duration>
      <itunes:summary>Introducing Nano Banana Pro, a powerful model built on Gemini 3 Pro, designed to enhance text rendering, infographics, and structured content generation. Tune in to learn about Nano Banana Pro’s advanced visual reasoning and multi-turn generation capabilities, and how this next-gen tool enables complex image edits and real-world applications. In this episode, we discuss how user feedback and continuous benchmarking drive model improvements, ensuring a superior experience for developers.</itunes:summary>
      <itunes:subtitle>Introducing Nano Banana Pro, a powerful model built on Gemini 3 Pro, designed to enhance text rendering, infographics, and structured content generation. Tune in to learn about Nano Banana Pro’s advanced visual reasoning and multi-turn generation capabilities, and how this next-gen tool enables complex image edits and real-world applications. In this episode, we discuss how user feedback and continuous benchmarking drive model improvements, ensuring a superior experience for developers.</itunes:subtitle>
      <itunes:explicit>false</itunes:explicit>
      <itunes:episodeType>full</itunes:episodeType>
      <itunes:episode>20</itunes:episode>
    </item>
    <item>
      <guid isPermaLink="false">3f1cadf7-b4ee-4199-9b8a-b1fe0d12d485</guid>
      <title>Koray Kavukcuoglu: “This Is How We Are Going to Build AGI”</title>
      <description><![CDATA[<p>Join Logan Kilpatrick and Koray Kavukcuoglu, CTO of Google DeepMind and Chief AI Architect of Google, as they discuss Gemini 3 and the state of AI!</p><p>Their conversation includes the reception of Gemini 3, the ongoing advancements in AI research, and the role of benchmarks in pushing new frontiers. They explore critical areas for Gemini's focus, emphasizing instruction following, tool calls, and internationalization, alongside Google's collaborative approach to AI development.</p><p>Watch on YouTube: https://www.youtube.com/watch?v=fXtna7UrL44</p><p><strong>Chapters:</strong><br />0:00 - Intro<br />2:00 - Gemini 3 launch reception<br />4:16 - Continuous progress and innovation<br />6:47 - Key areas for Gemini improvement<br />11:45 - Product scaffolding for model improvement<br />13:56 - Chief AI architect role<br />17:04 - Engineering mindset and collaboration<br />18:37 - Future growth areas for Gemini<br />20:33 - From research to engineering mindset<br />23:22 - The rise of generative media<br />27:22 - Nano Banana Pro capabilities<br />29:31 - Towards unified model checkpoints<br />36:26 - Organizing for AI success<br />38:26 - Balancing exploration and scaling<br />41:40 - DeepMind's collaborative culture<br />45:21 - Innovating at Google<br />48:37 - Closing</p>
]]></description>
      <pubDate>Tue, 25 Nov 2025 00:45:59 +0000</pubDate>
      <author>google-ai-podcast-release-notes@google.com (Google)</author>
      <link>https://google-ai-release-notes.simplecast.com/episodes/koray-kavukcuoglu-this-is-how-we-are-going-to-build-agi-VEzpyLmh</link>
      <content:encoded><![CDATA[<p>Join Logan Kilpatrick and Koray Kavukcuoglu, CTO of Google DeepMind and Chief AI Architect of Google, as they discuss Gemini 3 and the state of AI!</p><p>Their conversation includes the reception of Gemini 3, the ongoing advancements in AI research, and the role of benchmarks in pushing new frontiers. They explore critical areas for Gemini's focus, emphasizing instruction following, tool calls, and internationalization, alongside Google's collaborative approach to AI development.</p><p>Watch on YouTube: https://www.youtube.com/watch?v=fXtna7UrL44</p><p><strong>Chapters:</strong><br />0:00 - Intro<br />2:00 - Gemini 3 launch reception<br />4:16 - Continuous progress and innovation<br />6:47 - Key areas for Gemini improvement<br />11:45 - Product scaffolding for model improvement<br />13:56 - Chief AI architect role<br />17:04 - Engineering mindset and collaboration<br />18:37 - Future growth areas for Gemini<br />20:33 - From research to engineering mindset<br />23:22 - The rise of generative media<br />27:22 - Nano Banana Pro capabilities<br />29:31 - Towards unified model checkpoints<br />36:26 - Organizing for AI success<br />38:26 - Balancing exploration and scaling<br />41:40 - DeepMind's collaborative culture<br />45:21 - Innovating at Google<br />48:37 - Closing</p>
]]></content:encoded>
      <enclosure length="46796241" type="audio/mpeg" url="https://cdn.simplecast.com/audio/c68ffadd-1655-4420-a3c1-6df8dd507951/episodes/c0b9db3e-d0cb-49fc-a7b1-210bff72db1c/audio/09f25ef5-6e77-4adf-95eb-ee21f2e52891/default_tc.mp3?aid=rss_feed&amp;feed=niac4zV1"/>
      <itunes:title>Koray Kavukcuoglu: “This Is How We Are Going to Build AGI”</itunes:title>
      <itunes:author>Google</itunes:author>
      <itunes:image href="https://image.simplecastcdn.com/images/95190ff2-3221-4589-bc23-6d49f11be41f/d0e25761-ae5e-45ec-a570-842e56bc6eb0/3000x3000/1x1-rnp.jpg?aid=rss_feed"/>
      <itunes:duration>00:48:44</itunes:duration>
      <itunes:summary>Join Logan Kilpatrick and Koray Kavukcuoglu, CTO of Google DeepMind and Chief AI Architect of Google, as they discuss Gemini 3 and the state of AI! 

Their conversation includes the reception of Gemini 3, the ongoing advancements in AI research, and the role of benchmarks in pushing new frontiers. They explore critical areas for Gemini&apos;s focus, emphasizing instruction following, tool calls, and internationalization, alongside Google&apos;s collaborative approach to AI development.</itunes:summary>
      <itunes:subtitle>Join Logan Kilpatrick and Koray Kavukcuoglu, CTO of Google DeepMind and Chief AI Architect of Google, as they discuss Gemini 3 and the state of AI! 

Their conversation includes the reception of Gemini 3, the ongoing advancements in AI research, and the role of benchmarks in pushing new frontiers. They explore critical areas for Gemini&apos;s focus, emphasizing instruction following, tool calls, and internationalization, alongside Google&apos;s collaborative approach to AI development.</itunes:subtitle>
      <itunes:explicit>false</itunes:explicit>
      <itunes:episodeType>full</itunes:episodeType>
      <itunes:episode>19</itunes:episode>
    </item>
    <item>
      <guid isPermaLink="false">7fb5e1bd-2751-4b5a-bfba-2f389aee39a4</guid>
      <title>Google Antigravity: Hands on with our new agentic development platform</title>
      <description><![CDATA[<p>Explore Antigravity, Google DeepMind’s innovative new AI developer coding product, with Varun Mohan on Release Notes. This episode dives into Antigravity as a powerful agent development platform, integrating a familiar IDE experience with browser verification and Gemini 3.0 capabilities. Discover how developers can orchestrate complex agentic workflows, leverage artifacts for task communication, and balance AI automation with human collaboration. Learn about the philosophy behind building next-gen agentic experiences, the platform's multimodal strengths, and its role in accelerating software development at scale.</p><p>Watch on YouTube: https://www.youtube.com/watch?v=uzFOhkORVfk</p><p><strong>Chapters</strong><br />00:00 - Introducing Google Antigravity<br />04:02 - Evolution of AI in coding<br />04:53 - Beyond writing code<br />06:21 - Ideal Google Antigravity user<br />09:48 - Evolving user personas<br />11:46 - Agents versus the IDE<br />14:46 - Human-agent collaboration<br />16:43 - Local versus server-side<br />18:50 - Self-improvement and knowledge<br />21:29 - Generalizing agent capabilities<br />24:20 - Naming Google Antigravity<br />27:04 - Integrating Google's AI models<br />27:59 - Demo: Airbnb for dogs<br />28:48 - Understanding artifacts<br />29:51 - Asynchronous user feedback<br />32:16 - Agent manager workflow<br />33:17 - Browser actuation demo<br />34:36 - Browser for research and testing<br />36:45 - Parallel agent conversations<br />41:04 - Agent task best practices<br />42:51 - Future of Google Antigravity</p><p> </p>
]]></description>
      <pubDate>Tue, 25 Nov 2025 00:45:48 +0000</pubDate>
      <author>google-ai-podcast-release-notes@google.com (Google)</author>
      <link>https://google-ai-release-notes.simplecast.com/episodes/google-antigravity-hands-on-with-our-new-agentic-development-platform-Dd6xd5WJ</link>
      <content:encoded><![CDATA[<p>Explore Antigravity, Google DeepMind’s innovative new AI developer coding product, with Varun Mohan on Release Notes. This episode dives into Antigravity as a powerful agent development platform, integrating a familiar IDE experience with browser verification and Gemini 3.0 capabilities. Discover how developers can orchestrate complex agentic workflows, leverage artifacts for task communication, and balance AI automation with human collaboration. Learn about the philosophy behind building next-gen agentic experiences, the platform's multimodal strengths, and its role in accelerating software development at scale.</p><p>Watch on YouTube: https://www.youtube.com/watch?v=uzFOhkORVfk</p><p><strong>Chapters</strong><br />00:00 - Introducing Google Antigravity<br />04:02 - Evolution of AI in coding<br />04:53 - Beyond writing code<br />06:21 - Ideal Google Antigravity user<br />09:48 - Evolving user personas<br />11:46 - Agents versus the IDE<br />14:46 - Human-agent collaboration<br />16:43 - Local versus server-side<br />18:50 - Self-improvement and knowledge<br />21:29 - Generalizing agent capabilities<br />24:20 - Naming Google Antigravity<br />27:04 - Integrating Google's AI models<br />27:59 - Demo: Airbnb for dogs<br />28:48 - Understanding artifacts<br />29:51 - Asynchronous user feedback<br />32:16 - Agent manager workflow<br />33:17 - Browser actuation demo<br />34:36 - Browser for research and testing<br />36:45 - Parallel agent conversations<br />41:04 - Agent task best practices<br />42:51 - Future of Google Antigravity</p><p> </p>
]]></content:encoded>
      <enclosure length="43037116" type="audio/mpeg" url="https://cdn.simplecast.com/audio/c68ffadd-1655-4420-a3c1-6df8dd507951/episodes/6cb83c74-8a4f-4f92-834e-606bd54874d6/audio/bedece0e-22c6-48ba-af2d-4c83a16afc70/default_tc.mp3?aid=rss_feed&amp;feed=niac4zV1"/>
      <itunes:title>Google Antigravity: Hands on with our new agentic development platform</itunes:title>
      <itunes:author>Google</itunes:author>
      <itunes:image href="https://image.simplecastcdn.com/images/95190ff2-3221-4589-bc23-6d49f11be41f/ca643431-dd27-44ff-be16-464538ee8cdb/3000x3000/1x1-rnp.jpg?aid=rss_feed"/>
      <itunes:duration>00:44:49</itunes:duration>
      <itunes:summary>Explore Antigravity, Google DeepMind’s innovative new AI developer coding product, with Varun Mohan on Release Notes. This episode dives into Antigravity as a powerful agent development platform, integrating a familiar IDE experience with browser verification and Gemini 3.0 capabilities. Discover how developers can orchestrate complex agentic workflows, leverage artifacts for task communication, and balance AI automation with human collaboration. Learn about the philosophy behind building next-gen agentic experiences, the platform&apos;s multimodal strengths, and its role in accelerating software development at scale.</itunes:summary>
      <itunes:subtitle>Explore Antigravity, Google DeepMind’s innovative new AI developer coding product, with Varun Mohan on Release Notes. This episode dives into Antigravity as a powerful agent development platform, integrating a familiar IDE experience with browser verification and Gemini 3.0 capabilities. Discover how developers can orchestrate complex agentic workflows, leverage artifacts for task communication, and balance AI automation with human collaboration. Learn about the philosophy behind building next-gen agentic experiences, the platform&apos;s multimodal strengths, and its role in accelerating software development at scale.</itunes:subtitle>
      <itunes:explicit>false</itunes:explicit>
      <itunes:episodeType>full</itunes:episodeType>
      <itunes:episode>18</itunes:episode>
    </item>
    <item>
      <guid isPermaLink="false">e117ad94-703e-4b3b-a545-fc8700897bbc</guid>
      <title>Gemini 3: Launch day reactions</title>
      <description><![CDATA[<p>Join us for a special episode of Release Notes as we unpack Gemini 3, Google’s latest AI model with key team members. Learn how Gemini 3 empowers developers with enhanced multimodal understanding, agentic capabilities for complex tasks, and generative interfaces that transform prompts into interactive applications. We discuss real-world use cases, the iterative development process driven by user feedback, and the strategic balance between model performance and broad accessibility across various Google platforms.</p><p>Watch on YouTube: https://www.youtube.com/watch?v=mci0f2dy7G0</p><p><strong>Chapters:</strong><br />00:00 - Introducing Gemini 3<br />03:08 - Gemini 3 everywhere<br />04:13 - The product-model partnership<br />08:20 - Balancing speed and quality<br />11:40 - Gemini 3 'wow' moments<br />27:47 - Generative interfaces and UI<br />31:44 - Gemini's agentic capabilities<br />33:55 - Proactive AI and future<br />34:55 - Managing compute demand<br />39:32 - The Gemini 3 family<br />41:45 - Conclusion</p>
]]></description>
      <pubDate>Tue, 25 Nov 2025 00:44:47 +0000</pubDate>
      <author>google-ai-podcast-release-notes@google.com (Google)</author>
      <link>https://google-ai-release-notes.simplecast.com/episodes/gemini-3-launch-day-reactions-WCnzR2Y8</link>
      <content:encoded><![CDATA[<p>Join us for a special episode of Release Notes as we unpack Gemini 3, Google’s latest AI model with key team members. Learn how Gemini 3 empowers developers with enhanced multimodal understanding, agentic capabilities for complex tasks, and generative interfaces that transform prompts into interactive applications. We discuss real-world use cases, the iterative development process driven by user feedback, and the strategic balance between model performance and broad accessibility across various Google platforms.</p><p>Watch on YouTube: https://www.youtube.com/watch?v=mci0f2dy7G0</p><p><strong>Chapters:</strong><br />00:00 - Introducing Gemini 3<br />03:08 - Gemini 3 everywhere<br />04:13 - The product-model partnership<br />08:20 - Balancing speed and quality<br />11:40 - Gemini 3 'wow' moments<br />27:47 - Generative interfaces and UI<br />31:44 - Gemini's agentic capabilities<br />33:55 - Proactive AI and future<br />34:55 - Managing compute demand<br />39:32 - The Gemini 3 family<br />41:45 - Conclusion</p>
]]></content:encoded>
      <enclosure length="40580352" type="audio/mpeg" url="https://cdn.simplecast.com/audio/c68ffadd-1655-4420-a3c1-6df8dd507951/episodes/aaf441cf-a3ee-43c0-ab65-4419d01380a4/audio/6939ef3f-ccef-4f71-91dd-cfdac82889b0/default_tc.mp3?aid=rss_feed&amp;feed=niac4zV1"/>
      <itunes:title>Gemini 3: Launch day reactions</itunes:title>
      <itunes:author>Google</itunes:author>
      <itunes:image href="https://image.simplecastcdn.com/images/95190ff2-3221-4589-bc23-6d49f11be41f/2eb3c4b7-d9b2-463c-bc54-a5e7a29be28d/3000x3000/1x1-rnp.jpg?aid=rss_feed"/>
      <itunes:duration>00:42:16</itunes:duration>
      <itunes:summary>Join us for a special episode of Release Notes as we unpack Gemini 3, Google’s latest AI model with key team members. Learn how Gemini 3 empowers developers with enhanced multimodal understanding, agentic capabilities for complex tasks, and generative interfaces that transform prompts into interactive applications. We discuss real-world use cases, the iterative development process driven by user feedback, and the strategic balance between model performance and broad accessibility across various Google platforms.</itunes:summary>
      <itunes:subtitle>Join us for a special episode of Release Notes as we unpack Gemini 3, Google’s latest AI model with key team members. Learn how Gemini 3 empowers developers with enhanced multimodal understanding, agentic capabilities for complex tasks, and generative interfaces that transform prompts into interactive applications. We discuss real-world use cases, the iterative development process driven by user feedback, and the strategic balance between model performance and broad accessibility across various Google platforms.</itunes:subtitle>
      <itunes:explicit>false</itunes:explicit>
      <itunes:episodeType>full</itunes:episodeType>
      <itunes:episode>17</itunes:episode>
    </item>
    <item>
      <guid isPermaLink="false">ff188317-1bab-4375-ae9c-02913bee662a</guid>
      <title>How a Moonshot Led to Google DeepMind&apos;s Veo 3</title>
      <description><![CDATA[<p>Dumi Erhan, co-lead of the Veo project at Google DeepMind, joins host Logan Kilpatrick for a deep dive into the evolution of generative video models. They discuss the journey from early research in 2018 to the launch of state-of-the-art Veo 3 model with native audio generation. Learn about the technical hurdles in evaluating and scaling video models, the challenges of long-duration video coherence and how user feedback is shaping the future of AI-powered video creation.<br /><br /><strong>Chapter: </strong><br />0:00 - Intro<br />0:47 - Veo project's beginnings<br />3:02 - Veo's origins in Google Brain<br />5:07 - Video prediction and robotics applications<br />7:45 - Early progress and evaluation challenges<br />10:30 - Physics-based evaluations and their limitations<br />12:18 - The launch of the original Veo model<br />14:06 - Scaling challenges for video models<br />16:02 - The leap from Veo1 to Veo2<br />19:40 - Veo 3’s viral audio moment<br />21:17 - User trends shaping Veo's roadmap<br />23:49 - Image-to-video vs. text-to-video complexity<br />26:00 - New prompting methods and user control<br />27:55 - Coherence in long video generation<br />31:03 - Genie 3 and world models<br />35:54 - The steerability challenge<br />41:59 - Capability transfer and image data's role<br />47:25 - Closing</p><p> </p>
]]></description>
      <pubDate>Thu, 16 Oct 2025 01:26:13 +0000</pubDate>
      <author>google-ai-podcast-release-notes@google.com (Google)</author>
      <link>https://google-ai-release-notes.simplecast.com/episodes/how-a-moonshot-led-to-google-deepminds-veo-3-pcCVL5KH</link>
      <content:encoded><![CDATA[<p>Dumi Erhan, co-lead of the Veo project at Google DeepMind, joins host Logan Kilpatrick for a deep dive into the evolution of generative video models. They discuss the journey from early research in 2018 to the launch of state-of-the-art Veo 3 model with native audio generation. Learn about the technical hurdles in evaluating and scaling video models, the challenges of long-duration video coherence and how user feedback is shaping the future of AI-powered video creation.<br /><br /><strong>Chapter: </strong><br />0:00 - Intro<br />0:47 - Veo project's beginnings<br />3:02 - Veo's origins in Google Brain<br />5:07 - Video prediction and robotics applications<br />7:45 - Early progress and evaluation challenges<br />10:30 - Physics-based evaluations and their limitations<br />12:18 - The launch of the original Veo model<br />14:06 - Scaling challenges for video models<br />16:02 - The leap from Veo1 to Veo2<br />19:40 - Veo 3’s viral audio moment<br />21:17 - User trends shaping Veo's roadmap<br />23:49 - Image-to-video vs. text-to-video complexity<br />26:00 - New prompting methods and user control<br />27:55 - Coherence in long video generation<br />31:03 - Genie 3 and world models<br />35:54 - The steerability challenge<br />41:59 - Capability transfer and image data's role<br />47:25 - Closing</p><p> </p>
]]></content:encoded>
      <enclosure length="46242863" type="audio/mpeg" url="https://cdn.simplecast.com/audio/c68ffadd-1655-4420-a3c1-6df8dd507951/episodes/6a37974a-5a0f-40ce-a90a-babff358ec9e/audio/d5660c23-224b-45b2-a4c0-9f7a5ec44e4f/default_tc.mp3?aid=rss_feed&amp;feed=niac4zV1"/>
      <itunes:title>How a Moonshot Led to Google DeepMind&apos;s Veo 3</itunes:title>
      <itunes:author>Google</itunes:author>
      <itunes:image href="https://image.simplecastcdn.com/images/95190ff2-3221-4589-bc23-6d49f11be41f/9c731a14-1742-4658-97a2-ab3a5dfd63cc/3000x3000/ep17-1x1.jpg?aid=rss_feed"/>
      <itunes:duration>00:48:10</itunes:duration>
      <itunes:summary>Dumi Erhan, co-lead of the Veo project at Google DeepMind, joins host Logan Kilpatrick for a deep dive into the evolution of generative video models. They discuss the journey from early research in 2018 to the launch of state-of-the-art Veo 3 model with native audio generation. Learn about the technical hurdles in evaluating and scaling video models, the challenges of long-duration video coherence and how user feedback is shaping the future of AI-powered video creation.</itunes:summary>
      <itunes:subtitle>Dumi Erhan, co-lead of the Veo project at Google DeepMind, joins host Logan Kilpatrick for a deep dive into the evolution of generative video models. They discuss the journey from early research in 2018 to the launch of state-of-the-art Veo 3 model with native audio generation. Learn about the technical hurdles in evaluating and scaling video models, the challenges of long-duration video coherence and how user feedback is shaping the future of AI-powered video creation.</itunes:subtitle>
      <itunes:explicit>false</itunes:explicit>
      <itunes:episodeType>full</itunes:episodeType>
      <itunes:episode>16</itunes:episode>
    </item>
    <item>
      <guid isPermaLink="false">13b03199-0804-4398-8ec9-e23bdda4f558</guid>
      <title>GDM’s Pushmeet Kohli on solving science&apos;s biggest challenges with AI</title>
      <description><![CDATA[<p>Pushmeet Kohli, Head of Science and Strategic Initiatives at Google DeepMind, joins host Logan Kilpatrick to explore the intersection of AI and scientific discovery. Learn how the team's unique problem-solving framework led to innovations like AlphaFold and AlphaEvolve, and how new tools like AI Co-scientist aim to democratize these types of breakthroughs for everyone. </p><p>Watch on YouTube: https://www.youtube.com/watch?v=o7mdsL6BHsk</p><p><strong>Chapters: </strong><br />0:00 - Intro<br />1:04 - Recent Alpha launches<br />02:15 - Framework for selecting research domains<br />06:21  - Scientific, commercial and social impact<br />15:00 - Wielding AGI for breakthroughs<br />16:48 - Tech transfer and team collaboration<br />19:46  -  IMO Gold Medal<br />21:42  - Evaluating math proofs<br />22:55 - From specialized models to Deep Think<br />24:22 -  Do math skills generalize?<br />25:53 - Generalizing the IMO model<br />27:43 - Democratizing AI science tools<br />30:09 - AI Co-scientist<br />35:17 - An API for science?</p>
]]></description>
      <pubDate>Mon, 15 Sep 2025 16:31:39 +0000</pubDate>
      <author>google-ai-podcast-release-notes@google.com (Google)</author>
      <link>https://google-ai-release-notes.simplecast.com/episodes/gdms-pushmeet-kohli-on-solving-sciences-biggest-challenges-with-ai-U3BB6k5G</link>
      <content:encoded><![CDATA[<p>Pushmeet Kohli, Head of Science and Strategic Initiatives at Google DeepMind, joins host Logan Kilpatrick to explore the intersection of AI and scientific discovery. Learn how the team's unique problem-solving framework led to innovations like AlphaFold and AlphaEvolve, and how new tools like AI Co-scientist aim to democratize these types of breakthroughs for everyone. </p><p>Watch on YouTube: https://www.youtube.com/watch?v=o7mdsL6BHsk</p><p><strong>Chapters: </strong><br />0:00 - Intro<br />1:04 - Recent Alpha launches<br />02:15 - Framework for selecting research domains<br />06:21  - Scientific, commercial and social impact<br />15:00 - Wielding AGI for breakthroughs<br />16:48 - Tech transfer and team collaboration<br />19:46  -  IMO Gold Medal<br />21:42  - Evaluating math proofs<br />22:55 - From specialized models to Deep Think<br />24:22 -  Do math skills generalize?<br />25:53 - Generalizing the IMO model<br />27:43 - Democratizing AI science tools<br />30:09 - AI Co-scientist<br />35:17 - An API for science?</p>
]]></content:encoded>
      <enclosure length="35969844" type="audio/mpeg" url="https://cdn.simplecast.com/audio/c68ffadd-1655-4420-a3c1-6df8dd507951/episodes/70884187-1185-4658-a007-fb05e691b983/audio/ba6c7525-afd5-49f8-bb9c-77c96383d3a2/default_tc.mp3?aid=rss_feed&amp;feed=niac4zV1"/>
      <itunes:title>GDM’s Pushmeet Kohli on solving science&apos;s biggest challenges with AI</itunes:title>
      <itunes:author>Google</itunes:author>
      <itunes:image href="https://image.simplecastcdn.com/images/95190ff2-3221-4589-bc23-6d49f11be41f/34d78ae4-84cb-4cd8-93fa-a51b6620336c/3000x3000/release-notes-pushmeet.jpg?aid=rss_feed"/>
      <itunes:duration>00:37:28</itunes:duration>
      <itunes:summary>Pushmeet Kohli, Head of Science and Strategic Initiatives at Google DeepMind, joins host Logan Kilpatrick to explore the intersection of AI and scientific discovery. Learn how the team&apos;s unique problem-solving framework led to innovations like AlphaFold and AlphaEvolve, and how new tools like AI Co-scientist aim to democratize these types of breakthroughs for everyone.

Watch on YouTube: https://www.youtube.com/watch?v=o7mdsL6BHsk</itunes:summary>
      <itunes:subtitle>Pushmeet Kohli, Head of Science and Strategic Initiatives at Google DeepMind, joins host Logan Kilpatrick to explore the intersection of AI and scientific discovery. Learn how the team&apos;s unique problem-solving framework led to innovations like AlphaFold and AlphaEvolve, and how new tools like AI Co-scientist aim to democratize these types of breakthroughs for everyone.

Watch on YouTube: https://www.youtube.com/watch?v=o7mdsL6BHsk</itunes:subtitle>
      <itunes:explicit>false</itunes:explicit>
      <itunes:episodeType>full</itunes:episodeType>
      <itunes:episode>15</itunes:episode>
    </item>
    <item>
      <guid isPermaLink="false">04cbfa67-b6b8-4d02-8403-cb57bc4ed5d2</guid>
      <title>Behind the scenes of Google&apos;s state-of-the-art &quot;nano-banana&quot; image model</title>
      <description><![CDATA[<p>Join host Logan Kilpatrick in discussion with some of the minds behind Google's new state-of-the-art image model, Gemini 2.5 Flash. Product and research leads from the Gemini team break down the technology behind its key capabilities, including interleaved generation for complex edits and new approaches to achieving character consistency and pixel-perfect control. With Nicole Brichtova, Kaushik Shivakumar, Mostafa Dehghani and Robert Riachi. <br /><br />Watch on YouTube: <br /><br />Chapters:<br />0:37 - New model introduction<br />1:21 -Demo - Image Editing<br />3:44 - Text rendering capabilities<br />4:44 Beyond human preference evals<br />6:44 - Text rendering as a proxy for quality<br />8:38 - Positive transfer between modalities<br />11:25 - Demo - Multi-turn, context aware image generation<br />13:54 - Pixel-perfect editing and character consistency<br />15:51 - Interleaved image generation<br />17:59 - Specialized vs. native models<br />19:52 - Understanding nuanced prompts<br />20:59 - User feedback shaping model development<br />22:37 - Improvements in character consistency<br />24:17 - More natural looking images from team collaboration<br />26:41 - What’s next for image generation models</p>
]]></description>
      <pubDate>Wed, 27 Aug 2025 00:45:00 +0000</pubDate>
      <author>google-ai-podcast-release-notes@google.com (Google)</author>
      <link>https://google-ai-release-notes.simplecast.com/episodes/behind-the-scenes-of-googles-state-of-the-art-nano-banana-image-model-_Jtmi3CW</link>
      <content:encoded><![CDATA[<p>Join host Logan Kilpatrick in discussion with some of the minds behind Google's new state-of-the-art image model, Gemini 2.5 Flash. Product and research leads from the Gemini team break down the technology behind its key capabilities, including interleaved generation for complex edits and new approaches to achieving character consistency and pixel-perfect control. With Nicole Brichtova, Kaushik Shivakumar, Mostafa Dehghani and Robert Riachi. <br /><br />Watch on YouTube: <br /><br />Chapters:<br />0:37 - New model introduction<br />1:21 -Demo - Image Editing<br />3:44 - Text rendering capabilities<br />4:44 Beyond human preference evals<br />6:44 - Text rendering as a proxy for quality<br />8:38 - Positive transfer between modalities<br />11:25 - Demo - Multi-turn, context aware image generation<br />13:54 - Pixel-perfect editing and character consistency<br />15:51 - Interleaved image generation<br />17:59 - Specialized vs. native models<br />19:52 - Understanding nuanced prompts<br />20:59 - User feedback shaping model development<br />22:37 - Improvements in character consistency<br />24:17 - More natural looking images from team collaboration<br />26:41 - What’s next for image generation models</p>
]]></content:encoded>
      <enclosure length="29320949" type="audio/mpeg" url="https://cdn.simplecast.com/audio/c68ffadd-1655-4420-a3c1-6df8dd507951/episodes/ef9d25a5-d83d-4376-9c6f-72982ca5a1f5/audio/ca1fa06c-8943-4d96-a998-8370da173931/default_tc.mp3?aid=rss_feed&amp;feed=niac4zV1"/>
      <itunes:title>Behind the scenes of Google&apos;s state-of-the-art &quot;nano-banana&quot; image model</itunes:title>
      <itunes:author>Google</itunes:author>
      <itunes:image href="https://image.simplecastcdn.com/images/95190ff2-3221-4589-bc23-6d49f11be41f/5bcfc763-c8c3-4df4-b6e1-ec34d18990ee/3000x3000/ep-16-500x500.jpg?aid=rss_feed"/>
      <itunes:duration>00:30:32</itunes:duration>
      <itunes:summary>Join host Logan Kilpatrick in discussion with some of the minds behind Google&apos;s new state-of-the-art image model, Gemini 2.5 Flash. Product and research leads from the Gemini team break down the technology behind its key capabilities, including interleaved generation for complex edits and new approaches to achieving character consistency and pixel-perfect control. With Nicole Brichtova, Kaushik Shivakumar, Mostafa Dehghani and Robert Riachi. </itunes:summary>
      <itunes:subtitle>Join host Logan Kilpatrick in discussion with some of the minds behind Google&apos;s new state-of-the-art image model, Gemini 2.5 Flash. Product and research leads from the Gemini team break down the technology behind its key capabilities, including interleaved generation for complex edits and new approaches to achieving character consistency and pixel-perfect control. With Nicole Brichtova, Kaushik Shivakumar, Mostafa Dehghani and Robert Riachi. </itunes:subtitle>
      <itunes:explicit>false</itunes:explicit>
      <itunes:episodeType>full</itunes:episodeType>
      <itunes:episode>14</itunes:episode>
    </item>
    <item>
      <guid isPermaLink="false">3aaac72e-3b3c-45b1-8e85-1ac01aefffcf</guid>
      <title>Demis Hassabis on shipping momentum, better evals and world models</title>
      <description><![CDATA[<p>Demis Hassabis, CEO of Google DeepMind, sits down with host Logan Kilpatrick. In this episode, learn about the evolution from game-playing AI to today's thinking models, how projects like Genie 3 are building world models to help AI understand reality and why new testing grounds like Kaggle’s Game Arena are needed to evaluate progress on the path to AGI.</p><p>Watch on YouTube: https://www.youtube.com/watch?v=njDochQ2zHs<br /><br /><strong>Chapters:</strong><br />00:00 - Intro<br />01:16 - Recent GDM momentum<br />02:07 - Deep Think and agent systems<br />04:11 - Jagged intelligence<br />07:02 - Genie 3 and world models<br />10:21 - Future applications of Genie 3<br />13:01 - The need for better benchmarks and Kaggle Game Arena<br />19:03 - Evals beyond games<br />21:47 - Tool use for expanding AI capabilities<br />24:52 - Shift from models to systems<br />27:38 - Roadmap for Genie 3 and the omni model<br />29:25 - The quadrillion token club</p><p> </p>
]]></description>
      <pubDate>Mon, 11 Aug 2025 17:03:31 +0000</pubDate>
      <author>google-ai-podcast-release-notes@google.com (Google)</author>
      <link>https://google-ai-release-notes.simplecast.com/episodes/demis-hassabis-on-shipping-momentum-better-evals-and-world-models-AkZcL0Oe</link>
      <content:encoded><![CDATA[<p>Demis Hassabis, CEO of Google DeepMind, sits down with host Logan Kilpatrick. In this episode, learn about the evolution from game-playing AI to today's thinking models, how projects like Genie 3 are building world models to help AI understand reality and why new testing grounds like Kaggle’s Game Arena are needed to evaluate progress on the path to AGI.</p><p>Watch on YouTube: https://www.youtube.com/watch?v=njDochQ2zHs<br /><br /><strong>Chapters:</strong><br />00:00 - Intro<br />01:16 - Recent GDM momentum<br />02:07 - Deep Think and agent systems<br />04:11 - Jagged intelligence<br />07:02 - Genie 3 and world models<br />10:21 - Future applications of Genie 3<br />13:01 - The need for better benchmarks and Kaggle Game Arena<br />19:03 - Evals beyond games<br />21:47 - Tool use for expanding AI capabilities<br />24:52 - Shift from models to systems<br />27:38 - Roadmap for Genie 3 and the omni model<br />29:25 - The quadrillion token club</p><p> </p>
]]></content:encoded>
      <enclosure length="29906787" type="audio/mpeg" url="https://cdn.simplecast.com/audio/c68ffadd-1655-4420-a3c1-6df8dd507951/episodes/28c1f7fe-ee23-40b5-a937-c6b120005ead/audio/be22d7c4-9400-43fc-8687-487447f0c3d7/default_tc.mp3?aid=rss_feed&amp;feed=niac4zV1"/>
      <itunes:title>Demis Hassabis on shipping momentum, better evals and world models</itunes:title>
      <itunes:author>Google</itunes:author>
      <itunes:image href="https://image.simplecastcdn.com/images/95190ff2-3221-4589-bc23-6d49f11be41f/6f14fed7-94fa-4f57-a380-504b574c6ebc/3000x3000/ep14-1x1.jpg?aid=rss_feed"/>
      <itunes:duration>00:31:09</itunes:duration>
      <itunes:summary>Demis Hassabis, CEO of Google DeepMind, sits down with host Logan Kilpatrick. In this episode, learn about the evolution from game-playing AI to today&apos;s thinking models, how projects like Genie 3 are building world models to help AI understand reality and why new testing grounds like Kaggle’s Game Arena are needed to evaluate progress on the path to AGI.

Watch on YouTube: https://www.youtube.com/watch?v=njDochQ2zHs</itunes:summary>
      <itunes:subtitle>Demis Hassabis, CEO of Google DeepMind, sits down with host Logan Kilpatrick. In this episode, learn about the evolution from game-playing AI to today&apos;s thinking models, how projects like Genie 3 are building world models to help AI understand reality and why new testing grounds like Kaggle’s Game Arena are needed to evaluate progress on the path to AGI.

Watch on YouTube: https://www.youtube.com/watch?v=njDochQ2zHs</itunes:subtitle>
      <itunes:explicit>false</itunes:explicit>
      <itunes:episodeType>full</itunes:episodeType>
      <itunes:episode>13</itunes:episode>
    </item>
    <item>
      <guid isPermaLink="false">fdc2a6f5-50dc-4c4a-957d-4e44beea84bc</guid>
      <title>Building real-time voice applications with Live API</title>
      <description><![CDATA[<p>Shrestha Basu Mallick, one of the product leads for the Gemini API, joins host Logan Kilpatrick for a deep dive of Gemini Live API, Google’s real-time, multimodal interface for developers. Learn about how native audio alongside new capabilities like proactive audio and async function calling unlocks the unique power of audio as an interface.</p><p>Watch on YouTube: <a href="https://www.youtube.com/watch?v=4xlwlU6h-wM">https://www.youtube.com/watch?v=4xlwlU6h-wM</a><br /><br />0:00 - Intro<br />1:18 - Live API Overview<br />3:36 - Why audio is a special modality<br />5:07 - Speed vs. precision in audio<br />6:17 - Controllable and promptable TTS<br />8:31 - What developers are building with the Live API<br />11:14 - URL context and async calling features<br />15:02 - Proactive audio and affective dialog<br />16:55 - Addressing developer feedback<br />21:54 - Live API roadmap<br />23:49 - The role of long context<br />24:57 - What’s next for the Live API<br />26:41 - State of the AI audio market<br />30:10 - Advice for developers getting started with the Live API<br />31:16 - Live API demo<br />38:10 - Demo wrap up and closing</p><p> </p>
]]></description>
      <pubDate>Wed, 6 Aug 2025 07:22:26 +0000</pubDate>
      <author>google-ai-podcast-release-notes@google.com (Google)</author>
      <link>https://google-ai-release-notes.simplecast.com/episodes/building-real-time-voice-applications-with-live-api-brFoE9Wa</link>
      <content:encoded><![CDATA[<p>Shrestha Basu Mallick, one of the product leads for the Gemini API, joins host Logan Kilpatrick for a deep dive of Gemini Live API, Google’s real-time, multimodal interface for developers. Learn about how native audio alongside new capabilities like proactive audio and async function calling unlocks the unique power of audio as an interface.</p><p>Watch on YouTube: <a href="https://www.youtube.com/watch?v=4xlwlU6h-wM">https://www.youtube.com/watch?v=4xlwlU6h-wM</a><br /><br />0:00 - Intro<br />1:18 - Live API Overview<br />3:36 - Why audio is a special modality<br />5:07 - Speed vs. precision in audio<br />6:17 - Controllable and promptable TTS<br />8:31 - What developers are building with the Live API<br />11:14 - URL context and async calling features<br />15:02 - Proactive audio and affective dialog<br />16:55 - Addressing developer feedback<br />21:54 - Live API roadmap<br />23:49 - The role of long context<br />24:57 - What’s next for the Live API<br />26:41 - State of the AI audio market<br />30:10 - Advice for developers getting started with the Live API<br />31:16 - Live API demo<br />38:10 - Demo wrap up and closing</p><p> </p>
]]></content:encoded>
      <enclosure length="38624998" type="audio/mpeg" url="https://cdn.simplecast.com/audio/c68ffadd-1655-4420-a3c1-6df8dd507951/episodes/1196d0a1-db50-4a87-ae0a-fdf90ef149ad/audio/4bbcb459-3d5e-4807-82a1-51ed6fd7e3ed/default_tc.mp3?aid=rss_feed&amp;feed=niac4zV1"/>
      <itunes:title>Building real-time voice applications with Live API</itunes:title>
      <itunes:author>Google</itunes:author>
      <itunes:image href="https://image.simplecastcdn.com/images/95190ff2-3221-4589-bc23-6d49f11be41f/a3befa62-36ad-4d21-b2d3-999e68a0eb63/3000x3000/live-api-thumbnail-square.jpg?aid=rss_feed"/>
      <itunes:duration>00:40:14</itunes:duration>
      <itunes:summary>Shrestha Basu Mallick, one of the product leads for the Gemini API, joins host Logan Kilpatrick for a deep dive of Gemini Live API, Google’s real-time, multimodal interface for developers. Learn about how native audio alongside new capabilities like proactive audio and async function calling unlocks the unique power of audio as an interface. 

Watch on YouTube: https://www.youtube.com/watch?v=4xlwlU6h-wM

</itunes:summary>
      <itunes:subtitle>Shrestha Basu Mallick, one of the product leads for the Gemini API, joins host Logan Kilpatrick for a deep dive of Gemini Live API, Google’s real-time, multimodal interface for developers. Learn about how native audio alongside new capabilities like proactive audio and async function calling unlocks the unique power of audio as an interface. 

Watch on YouTube: https://www.youtube.com/watch?v=4xlwlU6h-wM

</itunes:subtitle>
      <itunes:explicit>false</itunes:explicit>
      <itunes:episodeType>full</itunes:episodeType>
      <itunes:episode>12</itunes:episode>
    </item>
    <item>
      <guid isPermaLink="false">3f38dca8-1795-4206-921f-5e69fc32e8d7</guid>
      <title>Building a frontier AI search experience</title>
      <description><![CDATA[<p>Robby Stein, VP of Product for Google Search, joins host Logan Kilpatrick to explore how Search is evolving into a frontier AI product. Their conversation covers the shift from simple keywords to complex, conversational queries, the rise of agentic capabilities that can take action on your behalf, and the vision to help billions of users truly "ask anything." Learn more about the technology behind AI Overviews, AI Mode, Deep Search, and the future of multimodal interaction.<br /><br />Watch on YouTube: <a href="https://youtu.be/zUB5A_ezIOU">https://youtu.be/zUB5A_ezIOU</a></p><p><strong>Chapters</strong><br />01:07 Search as a Frontier AI Product<br />02:38 Reaching 1.5 Billion Users<br />03:37 What Is AI Mode?<br />04:17 Understanding Query Fan-Out<br />05:18 Balancing Latency and performance with Gemini 2.5 Pro<br />06:51 How Deep Search works<br />09:08 Fine-tuning models for product experience<br />11:24 Shifting user behaviors<br />14:07 The rise of visual search<br />16:52 Speech and conversational AI in Search<br />18:36 Comparing Gemini and Search<br />20:04 Real-time tool use in Search<br />22:52 Evolving the Search interface<br />26:03 Making Search more personal<br />29:15 The agentic future of Search<br />31:15 Agents beyond booking tickets<br />37:11 On-the-fly software creation<br />38:06 Google DeepMind and Search collaboration<br />40:08 What's next for Search</p><p><br /> </p>
]]></description>
      <pubDate>Wed, 23 Jul 2025 19:05:00 +0000</pubDate>
      <author>google-ai-podcast-release-notes@google.com (Google)</author>
      <link>https://google-ai-release-notes.simplecast.com/episodes/building-a-frontier-ai-search-experience-mrw0fgtZ</link>
      <content:encoded><![CDATA[<p>Robby Stein, VP of Product for Google Search, joins host Logan Kilpatrick to explore how Search is evolving into a frontier AI product. Their conversation covers the shift from simple keywords to complex, conversational queries, the rise of agentic capabilities that can take action on your behalf, and the vision to help billions of users truly "ask anything." Learn more about the technology behind AI Overviews, AI Mode, Deep Search, and the future of multimodal interaction.<br /><br />Watch on YouTube: <a href="https://youtu.be/zUB5A_ezIOU">https://youtu.be/zUB5A_ezIOU</a></p><p><strong>Chapters</strong><br />01:07 Search as a Frontier AI Product<br />02:38 Reaching 1.5 Billion Users<br />03:37 What Is AI Mode?<br />04:17 Understanding Query Fan-Out<br />05:18 Balancing Latency and performance with Gemini 2.5 Pro<br />06:51 How Deep Search works<br />09:08 Fine-tuning models for product experience<br />11:24 Shifting user behaviors<br />14:07 The rise of visual search<br />16:52 Speech and conversational AI in Search<br />18:36 Comparing Gemini and Search<br />20:04 Real-time tool use in Search<br />22:52 Evolving the Search interface<br />26:03 Making Search more personal<br />29:15 The agentic future of Search<br />31:15 Agents beyond booking tickets<br />37:11 On-the-fly software creation<br />38:06 Google DeepMind and Search collaboration<br />40:08 What's next for Search</p><p><br /> </p>
]]></content:encoded>
      <enclosure length="41545837" type="audio/mpeg" url="https://cdn.simplecast.com/audio/c68ffadd-1655-4420-a3c1-6df8dd507951/episodes/c91f6af2-539b-4577-8bd5-f5344316d20e/audio/2e7ed2b7-8b2f-426c-9693-17a4ee5a3007/default_tc.mp3?aid=rss_feed&amp;feed=niac4zV1"/>
      <itunes:title>Building a frontier AI search experience</itunes:title>
      <itunes:author>Google</itunes:author>
      <itunes:image href="https://image.simplecastcdn.com/images/95190ff2-3221-4589-bc23-6d49f11be41f/57fb173d-f075-4c93-9b13-5afdbfeb88c7/3000x3000/release-notes-ep12-500x500.jpg?aid=rss_feed"/>
      <itunes:duration>00:43:16</itunes:duration>
      <itunes:summary>Robby Stein, VP of Product for Google Search, joins host Logan Kilpatrick to explore how Search is evolving into a frontier AI product. Their conversation covers the shift from simple keywords to complex, conversational queries, the rise of agentic capabilities that can take action on your behalf, and the vision to help billions of users truly &quot;ask anything.&quot; Learn more about the technology behind AI Overviews, AI Mode, Deep Search, and the future of multimodal interaction.</itunes:summary>
      <itunes:subtitle>Robby Stein, VP of Product for Google Search, joins host Logan Kilpatrick to explore how Search is evolving into a frontier AI product. Their conversation covers the shift from simple keywords to complex, conversational queries, the rise of agentic capabilities that can take action on your behalf, and the vision to help billions of users truly &quot;ask anything.&quot; Learn more about the technology behind AI Overviews, AI Mode, Deep Search, and the future of multimodal interaction.</itunes:subtitle>
      <itunes:explicit>false</itunes:explicit>
      <itunes:episodeType>full</itunes:episodeType>
      <itunes:episode>11</itunes:episode>
    </item>
    <item>
      <guid isPermaLink="false">dc287686-de49-4e52-9f8b-55e3fd021431</guid>
      <title>Gemini&apos;s Multimodality</title>
      <description><![CDATA[<p>Ani Baddepudi, Gemini Model Behavior Product Lead, joins host Logan Kilpatrick for a deep dive into Gemini's multimodal capabilities. Their conversation explores why Gemini was built as a natively multimodal model from day one, the future of proactive AI assistants, and how we are moving towards a world where "everything is vision." Learn about the differences between video and image understanding and token representations, higher FPS video sampling, and more.</p><p> </p><p><strong>Chapters:</strong></p><p>0:00 - Intro<br />1:12 - Why Gemini is natively multimodal<br />2:23 - The technology behind multimodal models<br />5:15 - Video understanding with Gemini 2.5<br />9:25 - Deciding what to build next<br />13:23 - Building new product experiences with multimodal AI<br />17:15 - The vision for proactive assistants<br />24:13 - Improving video usability with variable FPS and frame tokenization<br />27:35 - What’s next for Gemini’s multimodal development<br />31:47 - Deep dive on Gemini’s document understanding capabilities<br />37:56 - The teamwork and collaboration behind Gemini<br />40:56 - What’s next with model behavior</p><p><br />Watch on YouTube: https://www.youtube.com/watch?v=K4vXvaRV0dw</p>
]]></description>
      <pubDate>Wed, 2 Jul 2025 22:27:37 +0000</pubDate>
      <author>google-ai-podcast-release-notes@google.com (Google)</author>
      <link>https://google-ai-release-notes.simplecast.com/episodes/geminis-multimodality-VctGtJJi</link>
      <content:encoded><![CDATA[<p>Ani Baddepudi, Gemini Model Behavior Product Lead, joins host Logan Kilpatrick for a deep dive into Gemini's multimodal capabilities. Their conversation explores why Gemini was built as a natively multimodal model from day one, the future of proactive AI assistants, and how we are moving towards a world where "everything is vision." Learn about the differences between video and image understanding and token representations, higher FPS video sampling, and more.</p><p> </p><p><strong>Chapters:</strong></p><p>0:00 - Intro<br />1:12 - Why Gemini is natively multimodal<br />2:23 - The technology behind multimodal models<br />5:15 - Video understanding with Gemini 2.5<br />9:25 - Deciding what to build next<br />13:23 - Building new product experiences with multimodal AI<br />17:15 - The vision for proactive assistants<br />24:13 - Improving video usability with variable FPS and frame tokenization<br />27:35 - What’s next for Gemini’s multimodal development<br />31:47 - Deep dive on Gemini’s document understanding capabilities<br />37:56 - The teamwork and collaboration behind Gemini<br />40:56 - What’s next with model behavior</p><p><br />Watch on YouTube: https://www.youtube.com/watch?v=K4vXvaRV0dw</p>
]]></content:encoded>
      <enclosure length="42519682" type="audio/mpeg" url="https://cdn.simplecast.com/audio/c68ffadd-1655-4420-a3c1-6df8dd507951/episodes/7d5d0c57-f399-42f4-9dc6-42212886fa05/audio/0abe674e-0782-4637-9edf-a792e707e2fd/default_tc.mp3?aid=rss_feed&amp;feed=niac4zV1"/>
      <itunes:title>Gemini&apos;s Multimodality</itunes:title>
      <itunes:author>Google</itunes:author>
      <itunes:duration>00:44:17</itunes:duration>
      <itunes:summary>Ani Baddepudi, Gemini Model Behavior Product Lead, joins host Logan Kilpatrick for a deep dive into Gemini&apos;s multimodal capabilities. Their conversation explores why Gemini was built as a natively multimodal model from day one, the future of proactive AI assistants, and how we are moving towards a world where &quot;everything is vision.&quot; Learn about the differences between video and image understanding and token representations, higher FPS video sampling, and more.

Chapters:
0:00 - Intro 
1:12 - Why Gemini is natively multimodal
2:23 - The technology behind multimodal models
5:15 - Video understanding with Gemini 2.5
9:25 - Deciding what to build next
13:23 - Building new product experiences with multimodal AI
17:15 - The vision for proactive assistants
24:13 - Improving video usability with variable FPS and frame tokenization
27:35 - What’s next for Gemini’s multimodal development
31:47 - Deep dive on Gemini’s document understanding capabilities
37:56 - The teamwork and collaboration behind Gemini
40:56 - What’s next with model behavior

</itunes:summary>
      <itunes:subtitle>Ani Baddepudi, Gemini Model Behavior Product Lead, joins host Logan Kilpatrick for a deep dive into Gemini&apos;s multimodal capabilities. Their conversation explores why Gemini was built as a natively multimodal model from day one, the future of proactive AI assistants, and how we are moving towards a world where &quot;everything is vision.&quot; Learn about the differences between video and image understanding and token representations, higher FPS video sampling, and more.

Chapters:
0:00 - Intro 
1:12 - Why Gemini is natively multimodal
2:23 - The technology behind multimodal models
5:15 - Video understanding with Gemini 2.5
9:25 - Deciding what to build next
13:23 - Building new product experiences with multimodal AI
17:15 - The vision for proactive assistants
24:13 - Improving video usability with variable FPS and frame tokenization
27:35 - What’s next for Gemini’s multimodal development
31:47 - Deep dive on Gemini’s document understanding capabilities
37:56 - The teamwork and collaboration behind Gemini
40:56 - What’s next with model behavior

</itunes:subtitle>
      <itunes:explicit>false</itunes:explicit>
      <itunes:episodeType>full</itunes:episodeType>
      <itunes:episode>10</itunes:episode>
    </item>
    <item>
      <guid isPermaLink="false">0a196801-c8f1-4a67-bb64-b50a1803df99</guid>
      <title>Building Gemini&apos;s Coding Capabilities</title>
      <description><![CDATA[<p>Connie Fan, Product Lead for Gemini's coding capabilities, and Danny Tarlow, Research Lead for Gemini's coding capabilities, join host Logan Kilpatrick for an in-depth discussion on how the team built one of the world's leading AI coding models. Learn more about the early goals that shaped Gemini's approach to code, the rise of 'vibe coding' and its impact on development, strategies for tackling large codebases with long context and agents, and the future of programming languages in the age of AI.</p><p>Watch on YouTube: ⁠<a href="https://www.youtube.com/watch?v=jwbG_m-X-gE⁠">https://www.youtube.com/watch?v=jwbG_m-X-gE⁠</a></p><p><strong>Chapters:</strong></p><p>0:00 - Intro<br />1:10 - Defining Early Coding Goals<br />6:23 - Ingredients of a Great Coding Model<br />9:28 - Adapting to Developer Workflows<br />11:40 - The Rise of Vibe Coding<br />14:43 - Code as a Reasoning Tool<br />17:20 - Code as a Universal Solver<br />20:47 - Evaluating Coding Models<br />24:30 - Leveraging Internal Googler Feedback<br />26:52 - Winning Over AI Skeptics<br />28:04 - Performance Across Programming Languages<br />33:05 - The Future of Programming Languages<br />36:16 - Strategies for Large Codebases<br />41:06 - Hill Climbing New Benchmarks<br />42:46 - Short-Term Improvements<br />44:42 - Model Style and Taste<br />47:43 - 2.5 Pro’s Breakthrough<br />51:06 - Early AI Coding Experiences<br />56:19 - Specialist vs. Generalist Models</p><p>⁠</p><p>⁠</p><p> </p>
]]></description>
      <pubDate>Mon, 16 Jun 2025 19:44:55 +0000</pubDate>
      <author>google-ai-podcast-release-notes@google.com (Google)</author>
      <link>https://google-ai-release-notes.simplecast.com/episodes/building-geminis-coding-capabilities-epaqJzYi</link>
      <content:encoded><![CDATA[<p>Connie Fan, Product Lead for Gemini's coding capabilities, and Danny Tarlow, Research Lead for Gemini's coding capabilities, join host Logan Kilpatrick for an in-depth discussion on how the team built one of the world's leading AI coding models. Learn more about the early goals that shaped Gemini's approach to code, the rise of 'vibe coding' and its impact on development, strategies for tackling large codebases with long context and agents, and the future of programming languages in the age of AI.</p><p>Watch on YouTube: ⁠<a href="https://www.youtube.com/watch?v=jwbG_m-X-gE⁠">https://www.youtube.com/watch?v=jwbG_m-X-gE⁠</a></p><p><strong>Chapters:</strong></p><p>0:00 - Intro<br />1:10 - Defining Early Coding Goals<br />6:23 - Ingredients of a Great Coding Model<br />9:28 - Adapting to Developer Workflows<br />11:40 - The Rise of Vibe Coding<br />14:43 - Code as a Reasoning Tool<br />17:20 - Code as a Universal Solver<br />20:47 - Evaluating Coding Models<br />24:30 - Leveraging Internal Googler Feedback<br />26:52 - Winning Over AI Skeptics<br />28:04 - Performance Across Programming Languages<br />33:05 - The Future of Programming Languages<br />36:16 - Strategies for Large Codebases<br />41:06 - Hill Climbing New Benchmarks<br />42:46 - Short-Term Improvements<br />44:42 - Model Style and Taste<br />47:43 - 2.5 Pro’s Breakthrough<br />51:06 - Early AI Coding Experiences<br />56:19 - Specialist vs. Generalist Models</p><p>⁠</p><p>⁠</p><p> </p>
]]></content:encoded>
      <enclosure length="58509291" type="audio/mpeg" url="https://cdn.simplecast.com/audio/c68ffadd-1655-4420-a3c1-6df8dd507951/episodes/487caaf8-2fa3-4df9-9fc2-637b4a3fc0ed/audio/9a9fec6a-abab-4490-a3dd-ea96ddb7b4bf/default_tc.mp3?aid=rss_feed&amp;feed=niac4zV1"/>
      <itunes:title>Building Gemini&apos;s Coding Capabilities</itunes:title>
      <itunes:author>Google</itunes:author>
      <itunes:duration>01:00:27</itunes:duration>
      <itunes:summary>Connie Fan, Product Lead for Gemini&apos;s coding capabilities, and Danny Tarlow, Research Lead for Gemini&apos;s coding capabilities, join host Logan Kilpatrick for an in-depth discussion on how the team built one of the world&apos;s leading AI coding models. Learn more about the early goals that shaped Gemini&apos;s approach to code, the rise of &apos;vibe coding&apos; and its impact on development, strategies for tackling large codebases with long context and agents, and the future of programming languages in the age of AI.</itunes:summary>
      <itunes:subtitle>Connie Fan, Product Lead for Gemini&apos;s coding capabilities, and Danny Tarlow, Research Lead for Gemini&apos;s coding capabilities, join host Logan Kilpatrick for an in-depth discussion on how the team built one of the world&apos;s leading AI coding models. Learn more about the early goals that shaped Gemini&apos;s approach to code, the rise of &apos;vibe coding&apos; and its impact on development, strategies for tackling large codebases with long context and agents, and the future of programming languages in the age of AI.</itunes:subtitle>
      <itunes:explicit>false</itunes:explicit>
      <itunes:episodeType>full</itunes:episodeType>
      <itunes:episode>9</itunes:episode>
    </item>
    <item>
      <guid isPermaLink="false">dc38a14e-2ebc-44de-9b88-6aa37b58717f</guid>
      <title>Sergey Brin on the Future of AI &amp; Gemini</title>
      <description><![CDATA[<p>A conversation with Sergey Brin, co-founder of Google and computer scientist working on Gemini, in reaction to a year of progress with Gemini.</p><p>Watch on YouTube: <a href="https://www.youtube.com/watch?v=o7U4DV9Fkc0">https://www.youtube.com/watch?v=o7U4DV9Fkc0</a><br /><br /><strong>Chapters</strong></p><p>0:20 - Initial reactions to I/O<br />2:00 - Focus on Gemini’s core text model<br />4:29 - Native audio in Gemini and Veo 3<br />8:34 - Insights from model training runs<br />10:07 - Surprises in current AI developments vs. past expectations<br />14:20 - Evolution of model training<br />16:40 - The future of reasoning and Deep Think<br />20:19 - Google’s startup culture and accelerating AI innovation<br />24:51 - Closing</p><p> </p>
]]></description>
      <pubDate>Mon, 16 Jun 2025 19:38:55 +0000</pubDate>
      <author>google-ai-podcast-release-notes@google.com (Google)</author>
      <link>https://google-ai-release-notes.simplecast.com/episodes/sergey-brin-on-the-future-of-ai-gemini-iZX8C2km</link>
      <content:encoded><![CDATA[<p>A conversation with Sergey Brin, co-founder of Google and computer scientist working on Gemini, in reaction to a year of progress with Gemini.</p><p>Watch on YouTube: <a href="https://www.youtube.com/watch?v=o7U4DV9Fkc0">https://www.youtube.com/watch?v=o7U4DV9Fkc0</a><br /><br /><strong>Chapters</strong></p><p>0:20 - Initial reactions to I/O<br />2:00 - Focus on Gemini’s core text model<br />4:29 - Native audio in Gemini and Veo 3<br />8:34 - Insights from model training runs<br />10:07 - Surprises in current AI developments vs. past expectations<br />14:20 - Evolution of model training<br />16:40 - The future of reasoning and Deep Think<br />20:19 - Google’s startup culture and accelerating AI innovation<br />24:51 - Closing</p><p> </p>
]]></content:encoded>
      <enclosure length="26227215" type="audio/mpeg" url="https://cdn.simplecast.com/audio/c68ffadd-1655-4420-a3c1-6df8dd507951/episodes/0418899f-549c-4c3e-8f71-c8a56f273560/audio/33ac83dc-0739-49d7-9720-ebf4a28c6a95/default_tc.mp3?aid=rss_feed&amp;feed=niac4zV1"/>
      <itunes:title>Sergey Brin on the Future of AI &amp; Gemini</itunes:title>
      <itunes:author>Google</itunes:author>
      <itunes:duration>00:27:19</itunes:duration>
      <itunes:summary>A conversation with Sergey Brin, co-founder of Google and computer scientist working on Gemini, in reaction to a year of progress with Gemini.</itunes:summary>
      <itunes:subtitle>A conversation with Sergey Brin, co-founder of Google and computer scientist working on Gemini, in reaction to a year of progress with Gemini.</itunes:subtitle>
      <itunes:explicit>false</itunes:explicit>
      <itunes:episodeType>full</itunes:episodeType>
      <itunes:episode>8</itunes:episode>
    </item>
    <item>
      <guid isPermaLink="false">ecb94ef7-3a48-4fe2-803f-3498a6fdb742</guid>
      <title>Google I/O 2025 Recap with Josh Woodward and Tulsee Doshi</title>
      <description><![CDATA[<p><strong>Learn more</strong></p><ul><li>AI Studio: https://aistudio.google.com/</li><li>Gemini Canvas: https://gemini.google.com/canvas</li><li>Mariner: https://labs.google.com/mariner/</li><li>Gemini Ultra: https://one.google.com/about/google-a...</li><li>Jules: https://jules.google/</li><li>Gemini Diffusion: https://deepmind.google/models/gemini...</li><li>Flow: https://labs.google/flow/about</li><li>Notebook LM: https://notebooklm.google.com/</li><li>Stitch: https://stitch.withgoogle.com/</li></ul><p><strong>Chapters</strong></p><ul><li>0:59 - I/O Day 1 Recap</li><li>02:48 - Envisioning I/O 2030</li><li>08:11 - AI for Scientific Breakthroughs</li><li>09:20 - Veo 3 & Flow</li><li>7:35 - Gemini Live & the Future of Proactive Assistants</li><li>20:30 - Gemini in Chrome & Future Apps</li><li>22:28 - New Gemini Models: DeepThink, Diffusion & 2.5 Flash/Pro Updates</li><li>27:19 - Developer Momentum & Feedback Loop</li><li>31:50 - New Developer Products: Jules, Stitch & CodeGen in AI Studio</li><li>37:44 - Evolving Product Development Process with AI</li><li>39:23 - Closing</li></ul><p> </p><p> </p><p> </p><p> </p><p> </p>
]]></description>
      <pubDate>Thu, 22 May 2025 17:39:16 +0000</pubDate>
      <author>google-ai-podcast-release-notes@google.com (Google)</author>
      <link>https://google-ai-release-notes.simplecast.com/episodes/google-i-o-2025-recap-with-josh-woodward-and-tulsee-doshi-4sL_YsLo</link>
      <content:encoded><![CDATA[<p><strong>Learn more</strong></p><ul><li>AI Studio: https://aistudio.google.com/</li><li>Gemini Canvas: https://gemini.google.com/canvas</li><li>Mariner: https://labs.google.com/mariner/</li><li>Gemini Ultra: https://one.google.com/about/google-a...</li><li>Jules: https://jules.google/</li><li>Gemini Diffusion: https://deepmind.google/models/gemini...</li><li>Flow: https://labs.google/flow/about</li><li>Notebook LM: https://notebooklm.google.com/</li><li>Stitch: https://stitch.withgoogle.com/</li></ul><p><strong>Chapters</strong></p><ul><li>0:59 - I/O Day 1 Recap</li><li>02:48 - Envisioning I/O 2030</li><li>08:11 - AI for Scientific Breakthroughs</li><li>09:20 - Veo 3 & Flow</li><li>7:35 - Gemini Live & the Future of Proactive Assistants</li><li>20:30 - Gemini in Chrome & Future Apps</li><li>22:28 - New Gemini Models: DeepThink, Diffusion & 2.5 Flash/Pro Updates</li><li>27:19 - Developer Momentum & Feedback Loop</li><li>31:50 - New Developer Products: Jules, Stitch & CodeGen in AI Studio</li><li>37:44 - Evolving Product Development Process with AI</li><li>39:23 - Closing</li></ul><p> </p><p> </p><p> </p><p> </p><p> </p>
]]></content:encoded>
      <enclosure length="38645201" type="audio/mpeg" url="https://cdn.simplecast.com/audio/c68ffadd-1655-4420-a3c1-6df8dd507951/episodes/b45a48fe-f192-4477-a9ae-760e2552f72f/audio/5d635fae-d1ff-4f30-af68-420cd295f0c0/default_tc.mp3?aid=rss_feed&amp;feed=niac4zV1"/>
      <itunes:title>Google I/O 2025 Recap with Josh Woodward and Tulsee Doshi</itunes:title>
      <itunes:author>Google</itunes:author>
      <itunes:duration>00:40:15</itunes:duration>
      <itunes:summary>Tulsi Doshi (Head of Product, Gemini) and Josh Woodward (VP of Google Labs and Google Gemini) join host Logan Kilpatrick from Google I/O 2025 for a special edition of Release Notes. They unpack the latest launches, new models and tools for developers. Listen to learn more about 2.5 Pro Deep Think, Veo 3 and AI coding agents like Jules and Stitch.

Watch on YouTube: https://www.youtube.com/watch?v=DPQF_6o4G_E
</itunes:summary>
      <itunes:subtitle>Tulsi Doshi (Head of Product, Gemini) and Josh Woodward (VP of Google Labs and Google Gemini) join host Logan Kilpatrick from Google I/O 2025 for a special edition of Release Notes. They unpack the latest launches, new models and tools for developers. Listen to learn more about 2.5 Pro Deep Think, Veo 3 and AI coding agents like Jules and Stitch.

Watch on YouTube: https://www.youtube.com/watch?v=DPQF_6o4G_E
</itunes:subtitle>
      <itunes:explicit>false</itunes:explicit>
      <itunes:episodeType>full</itunes:episodeType>
      <itunes:episode>7</itunes:episode>
    </item>
    <item>
      <guid isPermaLink="false">3eb40761-aad7-4180-a4d5-86b8c1272ac8</guid>
      <title>Deep Dive into Long Context</title>
      <description><![CDATA[<p>Explore the synergy between long context models and Retrieval Augmented Generation (RAG) in this episode of Release Notes. Join Google DeepMind's Nikolay Savinov as he discusses the importance of large context windows, how they enable Al agents, and what's next in the field.</p><p>Chapters:<br />0:52 Introduction & defining tokens<br />5:27 Context window importance<br />9:53 RAG vs. Long Context<br />14:19 Scaling beyond 2 million tokens<br />18:41 Long context improvements since 1.5 Pro release<br />23:26 Difficulty of attending to the whole context<br />28:37 Evaluating long context: beyond needle-in-a-haystack<br />33:41 Integrating long context research<br />34:57 Reasoning and long outputs<br />40:54 Tips for using long context<br />48:51 The future of long context: near-perfect recall and cost reduction<br />54:42 The role of infrastructure<br />56:15 Long-context and agents</p>
]]></description>
      <pubDate>Fri, 2 May 2025 15:24:03 +0000</pubDate>
      <author>google-ai-podcast-release-notes@google.com (Google)</author>
      <link>https://google-ai-release-notes.simplecast.com/episodes/deep-dive-into-long-context-QTFUO7xf</link>
      <content:encoded><![CDATA[<p>Explore the synergy between long context models and Retrieval Augmented Generation (RAG) in this episode of Release Notes. Join Google DeepMind's Nikolay Savinov as he discusses the importance of large context windows, how they enable Al agents, and what's next in the field.</p><p>Chapters:<br />0:52 Introduction & defining tokens<br />5:27 Context window importance<br />9:53 RAG vs. Long Context<br />14:19 Scaling beyond 2 million tokens<br />18:41 Long context improvements since 1.5 Pro release<br />23:26 Difficulty of attending to the whole context<br />28:37 Evaluating long context: beyond needle-in-a-haystack<br />33:41 Integrating long context research<br />34:57 Reasoning and long outputs<br />40:54 Tips for using long context<br />48:51 The future of long context: near-perfect recall and cost reduction<br />54:42 The role of infrastructure<br />56:15 Long-context and agents</p>
]]></content:encoded>
      <enclosure length="57159539" type="audio/mpeg" url="https://cdn.simplecast.com/audio/c68ffadd-1655-4420-a3c1-6df8dd507951/episodes/65c9ced4-9f3a-40a5-89fe-2ff0e5124c82/audio/7a5dab51-34de-4405-b4f2-3bc9ccc4844b/default_tc.mp3?aid=rss_feed&amp;feed=niac4zV1"/>
      <itunes:title>Deep Dive into Long Context</itunes:title>
      <itunes:author>Google</itunes:author>
      <itunes:duration>00:59:32</itunes:duration>
      <itunes:summary>Join Google DeepMind&apos;s Nikolay Savinov with host Logan Kilpatrick as they discuss scaling context windows into the millions, recent quality improvements, RAG versus long context, and what&apos;s next in the field.

Watch on YouTube: https://youtu.be/NHMJ9mqKeMQ
</itunes:summary>
      <itunes:subtitle>Join Google DeepMind&apos;s Nikolay Savinov with host Logan Kilpatrick as they discuss scaling context windows into the millions, recent quality improvements, RAG versus long context, and what&apos;s next in the field.

Watch on YouTube: https://youtu.be/NHMJ9mqKeMQ
</itunes:subtitle>
      <itunes:explicit>false</itunes:explicit>
      <itunes:episodeType>full</itunes:episodeType>
      <itunes:episode>6</itunes:episode>
    </item>
    <item>
      <guid isPermaLink="false">a39707a5-1662-4da7-8f8f-a1f0ad69db12</guid>
      <title>Launching Gemini 2.5</title>
      <description><![CDATA[<p>Tulsee Doshi, Head of Product for Gemini Models joins host Logan Kilpatrick for an in-depth discussion on the latest Gemini 2.5 Pro experimental launch. Gemini 2.5 is a well-rounded, multimodal thinking model, designed to tackle increasingly complex problems. From enhanced reasoning to advanced coding, Gemini 2.5 can create impressive web applications and agentic code applications. Learn about the process of building Gemini 2.5 Pro experimental, the improvements made across the stack, and what’s next for Gemini 2.5.</p><p> </p><p><strong>Chapters:</strong></p><p>0:00 - Introduction<br />1:05 - Gemini 2.5 launch overview<br />3:19 - Academic evals vs. vibe checks<br />6:19 - The jump to 2.5<br />7:51 - Coordinating cross-stack improvements<br />11:48 - Role of pre/post-training vs. test-time compute<br />13:21 - Shipping Gemini 2.5<br />15:29 - Embedded safety process<br />17:28 - Multimodal reasoning with Gemini 2.5<br />18:55 - Benchmark deep dive<br />22:07 - What’s next for Gemini<br />24:49 - Dynamic thinking in Gemini 2.5<br />25:37 - The team effort behind the launch</p><p> </p><p><strong>Resources:</strong></p><ul><li>Gemini → https://goo.gle/41Yf72b</li><li>Gemini 2.5 blog post → https://goo.gle/441SHiV</li><li>Example of Gemini’s 2.5 Pro’s game design skills →  https://goo.gle/43vxkq1</li><li>Demo: Gemini 2.5 Pro Experimental in Google AI Studio → https://goo.gle/4c5RbhE</li></ul>
]]></description>
      <pubDate>Fri, 28 Mar 2025 21:12:22 +0000</pubDate>
      <author>google-ai-podcast-release-notes@google.com (Google)</author>
      <link>https://google-ai-release-notes.simplecast.com/episodes/launching-gemini-25-CsDOk7Xw</link>
      <content:encoded><![CDATA[<p>Tulsee Doshi, Head of Product for Gemini Models joins host Logan Kilpatrick for an in-depth discussion on the latest Gemini 2.5 Pro experimental launch. Gemini 2.5 is a well-rounded, multimodal thinking model, designed to tackle increasingly complex problems. From enhanced reasoning to advanced coding, Gemini 2.5 can create impressive web applications and agentic code applications. Learn about the process of building Gemini 2.5 Pro experimental, the improvements made across the stack, and what’s next for Gemini 2.5.</p><p> </p><p><strong>Chapters:</strong></p><p>0:00 - Introduction<br />1:05 - Gemini 2.5 launch overview<br />3:19 - Academic evals vs. vibe checks<br />6:19 - The jump to 2.5<br />7:51 - Coordinating cross-stack improvements<br />11:48 - Role of pre/post-training vs. test-time compute<br />13:21 - Shipping Gemini 2.5<br />15:29 - Embedded safety process<br />17:28 - Multimodal reasoning with Gemini 2.5<br />18:55 - Benchmark deep dive<br />22:07 - What’s next for Gemini<br />24:49 - Dynamic thinking in Gemini 2.5<br />25:37 - The team effort behind the launch</p><p> </p><p><strong>Resources:</strong></p><ul><li>Gemini → https://goo.gle/41Yf72b</li><li>Gemini 2.5 blog post → https://goo.gle/441SHiV</li><li>Example of Gemini’s 2.5 Pro’s game design skills →  https://goo.gle/43vxkq1</li><li>Demo: Gemini 2.5 Pro Experimental in Google AI Studio → https://goo.gle/4c5RbhE</li></ul>
]]></content:encoded>
      <enclosure length="26810686" type="audio/mpeg" url="https://cdn.simplecast.com/audio/c68ffadd-1655-4420-a3c1-6df8dd507951/episodes/5f339b6e-00d1-4b39-9883-bce0fc931f15/audio/7c741811-586d-44a1-a552-3d4afda68c09/default_tc.mp3?aid=rss_feed&amp;feed=niac4zV1"/>
      <itunes:title>Launching Gemini 2.5</itunes:title>
      <itunes:author>Google</itunes:author>
      <itunes:duration>00:27:55</itunes:duration>
      <itunes:summary>Tulsee Doshi, Head of Product for Gemini Models joins host Logan Kilpatrick for an in-depth discussion on the latest Gemini 2.5 Pro experimental launch. Gemini 2.5 is a well-rounded, multimodal thinking model, designed to tackle increasingly complex problems. From enhanced reasoning to advanced coding, Gemini 2.5 can create impressive web applications and agentic code applications. Learn about the process of building Gemini 2.5 Pro experimental, the improvements made across the stack, and what’s next for Gemini 2.5.</itunes:summary>
      <itunes:subtitle>Tulsee Doshi, Head of Product for Gemini Models joins host Logan Kilpatrick for an in-depth discussion on the latest Gemini 2.5 Pro experimental launch. Gemini 2.5 is a well-rounded, multimodal thinking model, designed to tackle increasingly complex problems. From enhanced reasoning to advanced coding, Gemini 2.5 can create impressive web applications and agentic code applications. Learn about the process of building Gemini 2.5 Pro experimental, the improvements made across the stack, and what’s next for Gemini 2.5.</itunes:subtitle>
      <itunes:explicit>false</itunes:explicit>
      <itunes:episodeType>full</itunes:episodeType>
      <itunes:episode>5</itunes:episode>
    </item>
    <item>
      <guid isPermaLink="false">58ae1e4a-83ce-40d7-b401-f6eb41011c99</guid>
      <title>Gemini app: Canvas, Deep Research and Personalization</title>
      <description><![CDATA[<p>Dave Citron, Senior Director Product Management, joins host Logan Kilpatrick for an in-depth discussion on the latest Gemini updates and demos. Learn more about Canvas for collaborative content creation, enhanced Deep Research with Thinking Models and Audio Overview and a new personalization feature.</p><p>0:00 - Introduction<br />0:59 - Recent Gemini app launches<br />2:00 - Introducing Canvas<br />5:12 - Canvas in action<br />8:46 - More Canvas examples<br />12:02 - Enhanced capabilities with Thinking Models<br />15:12 - Deep Research in action<br />20:27 - The future of agentic experiences<br />22:12 Deep Research and Audio Overviews<br />24:11 - Personalization in Gemini app<br />27:50 - Personalization in action<br />29:58 - How personalization works: user data and privacy<br />32:30 -The future of personalization</p>
]]></description>
      <pubDate>Thu, 20 Mar 2025 16:00:00 +0000</pubDate>
      <author>google-ai-podcast-release-notes@google.com (Google)</author>
      <link>https://google-ai-release-notes.simplecast.com/episodes/gemini-app-canvas-deep-research-and-personalization-2_rBtEHE</link>
      <content:encoded><![CDATA[<p>Dave Citron, Senior Director Product Management, joins host Logan Kilpatrick for an in-depth discussion on the latest Gemini updates and demos. Learn more about Canvas for collaborative content creation, enhanced Deep Research with Thinking Models and Audio Overview and a new personalization feature.</p><p>0:00 - Introduction<br />0:59 - Recent Gemini app launches<br />2:00 - Introducing Canvas<br />5:12 - Canvas in action<br />8:46 - More Canvas examples<br />12:02 - Enhanced capabilities with Thinking Models<br />15:12 - Deep Research in action<br />20:27 - The future of agentic experiences<br />22:12 Deep Research and Audio Overviews<br />24:11 - Personalization in Gemini app<br />27:50 - Personalization in action<br />29:58 - How personalization works: user data and privacy<br />32:30 -The future of personalization</p>
]]></content:encoded>
      <enclosure length="35413958" type="audio/mpeg" url="https://cdn.simplecast.com/audio/c68ffadd-1655-4420-a3c1-6df8dd507951/episodes/7e5e8c65-c258-4523-b922-66cc2ae1a953/audio/049c55a5-bdd0-41e1-aca7-e222070200f9/default_tc.mp3?aid=rss_feed&amp;feed=niac4zV1"/>
      <itunes:title>Gemini app: Canvas, Deep Research and Personalization</itunes:title>
      <itunes:author>Google</itunes:author>
      <itunes:duration>00:36:53</itunes:duration>
      <itunes:summary>Dave Citron, Senior Director Product Management, joins host Logan Kilpatrick for an in-depth discussion on the latest Gemini updates and demos. Learn more about Canvas for collaborative content creation, enhanced Deep Research with Thinking Models and Audio Overview and a new personalization feature.


0:00 - Introduction
0:59 - Recent Gemini app launches
2:00 - Introducing Canvas 
5:12 - Canvas in action
8:46 - Advanced Canvas examples
12:02 - Enhanced capabilities with Thinking Models
15:12 - Deep Research in action
20:27 - The future of agentic experiences 
24:11 - Personalization in Gemini app
27:50 - Personalization in action
29:58 - How personalization works: user data and privacy
32:30 -The future of personalization</itunes:summary>
      <itunes:subtitle>Dave Citron, Senior Director Product Management, joins host Logan Kilpatrick for an in-depth discussion on the latest Gemini updates and demos. Learn more about Canvas for collaborative content creation, enhanced Deep Research with Thinking Models and Audio Overview and a new personalization feature.


0:00 - Introduction
0:59 - Recent Gemini app launches
2:00 - Introducing Canvas 
5:12 - Canvas in action
8:46 - Advanced Canvas examples
12:02 - Enhanced capabilities with Thinking Models
15:12 - Deep Research in action
20:27 - The future of agentic experiences 
24:11 - Personalization in Gemini app
27:50 - Personalization in action
29:58 - How personalization works: user data and privacy
32:30 -The future of personalization</itunes:subtitle>
      <itunes:explicit>false</itunes:explicit>
      <itunes:episodeType>full</itunes:episodeType>
      <itunes:episode>4</itunes:episode>
    </item>
    <item>
      <guid isPermaLink="false">8be68878-54fb-4ad3-b020-60232049c7f5</guid>
      <title>Developing Google DeepMind&apos;s Thinking Models</title>
      <description><![CDATA[<p>Jack Rae, Principal Scientist at Google DeepMind, joins host Logan Kilpatrick for an in-depth discussion on the development of Google’s thinking models. Learn more about practical applications of thinking models, the impact of increased 'thinking time' on model performance and the key role of long context.</p><p>01:14 - Defining Thinking Models<br />03:40 - Use Cases for Thinking Models<br />07:52 - Thinking Time Improves Answers<br />09:57 - Rapid Thinking Progress<br />20:11 - Long Context Is Key<br />27:41 - Tools for Thinking Models<br />29:44 - Incorporating Developer Feedback<br />35:11 - The Strawberry Counting Problem<br />39:15 - Thinking Model Development Timeline<br />42:30 - Towards a GA Thinking Model<br />49:24 - Thinking Models Powering AI Agents<br />54:14 - The Future of AI Model Evals</p>
]]></description>
      <pubDate>Mon, 24 Feb 2025 20:19:25 +0000</pubDate>
      <author>google-ai-podcast-release-notes@google.com (Google)</author>
      <link>https://google-ai-release-notes.simplecast.com/episodes/developing-google-deepminds-thinking-models-n2Z_GX_J</link>
      <content:encoded><![CDATA[<p>Jack Rae, Principal Scientist at Google DeepMind, joins host Logan Kilpatrick for an in-depth discussion on the development of Google’s thinking models. Learn more about practical applications of thinking models, the impact of increased 'thinking time' on model performance and the key role of long context.</p><p>01:14 - Defining Thinking Models<br />03:40 - Use Cases for Thinking Models<br />07:52 - Thinking Time Improves Answers<br />09:57 - Rapid Thinking Progress<br />20:11 - Long Context Is Key<br />27:41 - Tools for Thinking Models<br />29:44 - Incorporating Developer Feedback<br />35:11 - The Strawberry Counting Problem<br />39:15 - Thinking Model Development Timeline<br />42:30 - Towards a GA Thinking Model<br />49:24 - Thinking Models Powering AI Agents<br />54:14 - The Future of AI Model Evals</p>
]]></content:encoded>
      <enclosure length="61458445" type="audio/mpeg" url="https://cdn.simplecast.com/audio/c68ffadd-1655-4420-a3c1-6df8dd507951/episodes/533d25a8-8c4f-46a8-be45-d6fb9b33a358/audio/d960b18f-193f-4cd0-9d38-855c0f1aa3b9/default_tc.mp3?aid=rss_feed&amp;feed=niac4zV1"/>
      <itunes:title>Developing Google DeepMind&apos;s Thinking Models</itunes:title>
      <itunes:author>Google</itunes:author>
      <itunes:duration>01:03:32</itunes:duration>
      <itunes:summary>Jack Rae, Principal Scientist at Google DeepMind, joins host Logan Kilpatrick for an in-depth discussion on the development of Google’s thinking models. Learn more about practical applications of thinking models, the impact of increased &apos;thinking time&apos; on model performance and the key role of long context.</itunes:summary>
      <itunes:subtitle>Jack Rae, Principal Scientist at Google DeepMind, joins host Logan Kilpatrick for an in-depth discussion on the development of Google’s thinking models. Learn more about practical applications of thinking models, the impact of increased &apos;thinking time&apos; on model performance and the key role of long context.</itunes:subtitle>
      <itunes:explicit>false</itunes:explicit>
      <itunes:episodeType>full</itunes:episodeType>
      <itunes:episode>3</itunes:episode>
    </item>
    <item>
      <guid isPermaLink="false">20396b74-ef68-4bf4-ac89-1332d518915e</guid>
      <title>Behind the Scenes of Gemini 2.0</title>
      <description><![CDATA[Tulsee Doshi, Gemini model product lead, joins host Logan Kilpatrick to go behind the scenes of Gemini 2.0, taking a deep dive into the model's multimodal capabilities and native tool use, and Google's approach to shipping experimental models.

Watch on YouTube: https://www.youtube.com/watch?v=L7dw799vu5o
Chapters: 
Meet Tulsee Doshi
Gemini's Progress Over the Past Year
Introducing Gemini 2.0
Shipping Experimental Models
Gemini 2.0’s Native Tool Use
Function Calling
Multimodal Agents
Rapid Fire Questions 
]]></description>
      <pubDate>Wed, 11 Dec 2024 16:01:54 +0000</pubDate>
      <author>google-ai-podcast-release-notes@google.com (Google)</author>
      <link>https://google-ai-release-notes.simplecast.com/episodes/behind-the-scenes-of-gemini-20-kVTO2UT5</link>
      <enclosure length="34199056" type="audio/mpeg" url="https://cdn.simplecast.com/audio/c68ffadd-1655-4420-a3c1-6df8dd507951/episodes/059bb440-65d9-4cca-931a-afc1a2d4cd47/audio/93e65703-8c2c-40ed-aff8-a7adf3d1c4a9/default_tc.mp3?aid=rss_feed&amp;feed=niac4zV1"/>
      <itunes:title>Behind the Scenes of Gemini 2.0</itunes:title>
      <itunes:author>Google</itunes:author>
      <itunes:duration>00:35:18</itunes:duration>
      <itunes:summary>Tulsee Doshi, Gemini model product lead, joins host Logan Kilpatrick to go behind the scenes of Gemini 2.0, taking a deep dive into the model&apos;s multimodal capabilities and native tool use, and Google&apos;s approach to shipping experimental models.

Watch on YouTube: https://www.youtube.com/watch?v=L7dw799vu5o
Chapters: 
Meet Tulsee Doshi
Gemini&apos;s Progress Over the Past Year
Introducing Gemini 2.0
Shipping Experimental Models
Gemini 2.0’s Native Tool Use
Function Calling
Multimodal Agents
Rapid Fire Questions</itunes:summary>
      <itunes:subtitle>Tulsee Doshi, Gemini model product lead, joins host Logan Kilpatrick to go behind the scenes of Gemini 2.0, taking a deep dive into the model&apos;s multimodal capabilities and native tool use, and Google&apos;s approach to shipping experimental models.

Watch on YouTube: https://www.youtube.com/watch?v=L7dw799vu5o
Chapters: 
Meet Tulsee Doshi
Gemini&apos;s Progress Over the Past Year
Introducing Gemini 2.0
Shipping Experimental Models
Gemini 2.0’s Native Tool Use
Function Calling
Multimodal Agents
Rapid Fire Questions</itunes:subtitle>
      <itunes:explicit>false</itunes:explicit>
      <itunes:episodeType>full</itunes:episodeType>
      <itunes:episode>2</itunes:episode>
    </item>
    <item>
      <guid isPermaLink="false">f96ca290-889e-4f12-824c-40c92508380e</guid>
      <title>Smaller, Faster, Cheaper &amp; The Story of Flash 8B</title>
      <description><![CDATA[Logan Kilpatrick sits down with Emanuel Taropa, a key figure in the development of Gemini to delve into the cutting edge of AI. Taropa provides insights into the technical challenges and triumphs of building and deploying large language models, focusing on the recent release of the Flash 8B Gemini model.

Their conversation covers everything from the intricacies of model architecture and training to the practical challenges of shipping AI models at scale, and even speculates on the future of AI.  
]]></description>
      <pubDate>Thu, 5 Dec 2024 22:55:01 +0000</pubDate>
      <author>google-ai-podcast-release-notes@google.com (Logan Kilpatrick, Emanuel Taropa)</author>
      <link>https://google-ai-release-notes.simplecast.com/episodes/smaller-faster-cheaper-the-story-of-flash-8b-61azQD2P</link>
      <enclosure length="42853976" type="audio/mpeg" url="https://cdn.simplecast.com/audio/c68ffadd-1655-4420-a3c1-6df8dd507951/episodes/010cc18e-36df-488a-97f2-d2d085c6637d/audio/1acc344f-c58f-4293-974d-2ab2f12f9289/default_tc.mp3?aid=rss_feed&amp;feed=niac4zV1"/>
      <itunes:title>Smaller, Faster, Cheaper &amp; The Story of Flash 8B</itunes:title>
      <itunes:author>Logan Kilpatrick, Emanuel Taropa</itunes:author>
      <itunes:duration>00:43:20</itunes:duration>
      <itunes:summary>Logan Kilpatrick sits down with Emanuel Taropa, a key figure in the development of Gemini to delve into the cutting edge of AI. Taropa provides insights into the technical challenges and triumphs of building and deploying large language models, focusing on the recent release of the Flash 8B Gemini model.

Their conversation covers everything from the intricacies of model architecture and training to the practical challenges of shipping AI models at scale, and even speculates on the future of AI. </itunes:summary>
      <itunes:subtitle>Logan Kilpatrick sits down with Emanuel Taropa, a key figure in the development of Gemini to delve into the cutting edge of AI. Taropa provides insights into the technical challenges and triumphs of building and deploying large language models, focusing on the recent release of the Flash 8B Gemini model.

Their conversation covers everything from the intricacies of model architecture and training to the practical challenges of shipping AI models at scale, and even speculates on the future of AI. </itunes:subtitle>
      <itunes:explicit>false</itunes:explicit>
      <itunes:episodeType>full</itunes:episodeType>
      <itunes:episode>1</itunes:episode>
    </item>
  </channel>
</rss>