Transcript for:
Google Cloud Next's AI Innovations Overview

this is the Everyday AI show The everyday podcast where we simplify AI and bring its power to your fingertips Listen daily for practical advice to boost your career business and everyday life There's so much new that was just announced at Google Cloud Next I'm having a hard time wrapping my head around it It seems like there was dozens of new AI updates So I said what better than to bring in one of their leaders to help us make sense of it all So uh we're going to talk a little bit today about what's new inside Google Gemini and Google uh Gemini AI Studio like everything uh with Logan Kilpatrick the senior product manager at Google Deep Mind Logan thank you for joining us a second time Yeah round two This is going to be I don't even remember what we were talking about for round one It feels like it was it was super recently but um there's a ton of new stuff to talk about so I'm happy to be back Yeah absolutely So you know top to bottom I mean we saw new updates with uh Google Gemini 2.5 Pro being rolled out in other places a new model in Google Gemini 2.5 Flash right So many things for developers but where do you start or maybe like what are you most excited about that was just announced here at Google Cloud Next Yeah that's a great question So I've been continually excited about 2.5 Pro I think like we're we're seeing 2.5 Pro sort of rolling out across our developer products our consumer products It just landed in deep research yesterday which like folks have been super jazzed about I think if you're a if you're an advanced uh Gemini advance us you get like 20 deep research queries uh you know our customers prefer the sort of Gemini advance to uh with 2.5 Pro sort of 2:1 versus sort of other products in the market which I think is just sort of a nice proxy of like actually this model unlocks new stuff from a deep research perspective from a canvas perspective getting to see if folks haven't tried canvas yet um in the Gemini app being able to sort of vibe code and agentically sort of write code for you without having to be a developer is like such a such a cool and special experience um so That's what I've historically been most excited about I think today now at at Cloud Next we're you know tons of new stuff launched Vo is available for developers Um which if folks haven't seen is our sort of state-of-the-art uh video generation model which has been awesome Uh we just announced the live API which I think a lot of folks and actually like in parallel to that the live mode is rolling out to some customers and in Android I think as well So like there's everything happen like one of the things that I've been most happy about is it feels like more and more we're getting to the place where as these new capabilities come online they end up sort of ubiquitously across the Google ecosystem which is really cool because like you know some people are a Gemini you know user some people are a Google AI studio user we've got enterprise users we've got people in search and like I think it's awesome to sort of get to a place where new thing launches available everywhere for for the world to use So I want to quickly dive into two of those things that you mentioned there So the deep research I've been blown away not just by you know I I think there was a couple times you guys updated it first to 2.0 O and now I think to 2.5 Pro and yeah you talked about some of the benchmarks that came out in terms of uh you know the preference I'll say it's against open AIS right which I thought was a great uh you know a great offering but you know now seemingly your guys' is way way better uh you know what are you even using the deep research tools for like I I love asking the people that build it like what are you using it for because I think people can learn from what you're using it for Yeah that's a great question I think some of the stuff um like not uh the two use cases that have been top of mind for me one I was looking up like what the general sentiment is about MCP If folks haven't been following there's this we won't we won't dive into the MCP thread in this conversation but if you haven't done MCP before haven't looked into it use deep research It gives it an actually like pretty robust answer and like gave me a bunch of supporting materials of like not just how people on Twitter are thinking about MC people but like it's an agentic way of interacting with um with tools Uh so that was one of the use cases cuz I was just very intrigued to know like what happens when you do that Um the other one is I've been doing a bunch of like sort of competitive analysis of just like as we think about you know how we're showing up in the market what do we look like comparatively against other providers Um and this is a really interesting like I think for me the deep research conduit has been really interesting because what deep research is actually able to capture is sort of like the information that's available on the internet and I think it's like you know could I go and talk to customers and like get this perspective yes that's actually a really interesting and useful perspective but it's actually also interesting to capture like what is sort of the codified perspective on the internet of like how people think about you know the Gemini API or AI studio or stuff like So it's been really interesting just to like have that experience and like it it actually like diverges from in some interesting ways like what what people tell me in person about like how they think of the product is used and all this stuff So re really interesting if folks haven't done that exercise if you have like a product that you've built or you have like a favorite thing like just like ask the deep research um functionality in in the Gemini app like to put together a report and like see how that differs from your point of view or from yeah your perspective Yeah And another thing you just talked about there is kind of like you know vibe coding in canvas So canvas has been out what like two weeks two weeks something like that right like I use it so much already for a tool that's only been out two weeks but maybe walk people through some of the uh you know practical applications I think a lot of people are you know like oh like let's you know create a game and like that's fun to get started But you know in terms of you know business utility Uh what are you all seeing as as some of the more uh impressive or useful applications for the new canvas mode inside Gemini Yeah I still think we're in the era of um and and I think like chat bots in general were in that place for a long time and I think they've sort of just in the last like six to eight months like gotten out of just like being sort of a novelty item Um and I and I think like canvas is sort of still in that realm where it's like it can do interesting things To your point like building games from scratch Like I it would take me a long time to program a game from scratch right now You know be AI being able to do that is awesome Um but like where does the practical business value come in I think for a lot of people the practical business value comes in like when you're connecting this thing to your company's data and like all and like that's the kind of stuff that doesn't exist yet today at least in the sort of canvas environment that we have Um and it's what I'm most excited about because I think like ultimately for these tools to be useful like you need to connect a bunch of your stuff to them and sort of let them you know have access to your email and then I can sort of build a tool around my email to do it So I'm I'm really excited about that And I think from my like and I'm I'm not I'm not a product manager on the Gemini app but I'm a I'm a consumer of the Gemini app and I and I love it and I think it's a it's a great product Um one of the things that I'm most excited about is like this trend of the Gemini app sort of becoming this AI interface and this AI conduit to like all of the things that are happening inside of the Google ecosystem And like the the sort of pertinent example of this is the Gemini app also in addition to all the other canvas stuff and deep research and everything else going on Um it has a personalization feature and the personalization feature is actually built based on your Google search history So you can opt in to be like hey you know basically personalize the answers that the model's giving based on Google search And like that sounds very like uninteresting at the surface level but it starts to get to a world where like AI is this interface to like connect to this like vast set of data And I think about this for myself in the work context in the personal context like I'm on YouTube all the time I'm in Gmail all the time I'm searching stuff all the time Um I'm in docs all the time So like it's really wonderful to be able to sort of bring all that experience together And I think Canvas is like the first step of that with Docs specifically with cone now So I'm I'm super excited Yeah And and you know if you haven't had the time to use uh canvas yet I highly recommend it right It's literally being able to you know run and render code You don't have to be able to even know coding Uh it's it's so easy So uh another thing Logan there that you mentioned is is uh V2 Um and and you know some of the new capabilities uh in that that are available uh in in Vertex as well uh you know adjusting camera angles right What does this do for creatives right There are so many new things I wasn't even expecting that were announced today you know the the the text to text to music right The updated chirp like what is this going to do for creatives and how does this all unlock you know both in Vertex and AI Studio Yeah I think my I think the general trend that gets me excited is like and I was just having a conversation actually with the folks in the Vert.ex team about this and they and they sort of agree which is like this general upleveling of people to be able to like go to the next level like I'm not a creative like I'm also not a game designer like in the game design use case I couldn't build a video game I've tried before it's horribly difficult it's not it's not fun to sort of bash your head against the wall trying to do that I think there's a lot of cases where that's true in the video use case like you and me like you we were talking off camera like editing video is you know tough and there's a lot of great tools out there that help do it but like it's still kind of a pain in a lot of cases and to be able to have all these AI tools start to take those steps and like uplevel the people who are really excited and like take out the stuff that I'm not interested in doing Um I'm super excited I think like VO specifically has been the one that folks have been like losing their mind for a long time and this is actually like today is the first time that this with the exception of YouTube where it's like set up in a very specific product experience It's the first time the like raw model is like generally available to the world to actually get their hands on which feels like a crazy uh I don't think it's been as crazy of like a public moment yet as I think it actually is in reality but like the world's best video generation model is like now available for people to actually use and start building with So I think we're going to see the technology start showing up in lots of new interesting ways Yeah And it it was impressive and we'll share in the newsletter today uh the demo that they did I'm sure that Google's going to be posting that online right But being able to you know kind of do the the live shots of Las Vegas and animate them and put them to music Super impressive Uh something another new uh you know update here Fire Firebase is is is that what it is right Like did Google just release like an IDE out of nowhere like like tell us tell us what Firebase is How does it work Yeah Yeah This is a great question So there this is some slight developer context So if you're not a developer some of this stuff might not um it might not be relevant or might not make too much sense But so the original incarnation of that product which today became Firebase Studio was a was something called Project idx which we announced last year at um at Google IO And the intent of project idx was like how can we build a next generation um IDE integrated development environment for developers to actually use um in the browser which I think was the unique like today developers like download a local uh ID onto their computer and they do their development locally This was bringing the IDE to the develop to the to the browser Um and sort of the next iteration of that product suite uh and this was being created by the Firebase team which is why it ended up as Firebase Studio Um the next iteration is that product is how do you actually not just like do the basic developer environment but how do you infuse AI into that and how do you sort of help developers bootstrap actually going and creating apps and and products and stuff like that Um so I'm super excited for Firebase Studio I think it's like the um for folks who aren't close to Firebase like Firebase has a lot of like uh street credit as being like an incredibly like developer centric team and product surface Um so I think I I haven't spent a bunch of time with um I spent a much time with idx haven't spent a bunch of time with Firebase Studio yet Uh but I I have full conviction that that team is is going to knock it out of the park and uh hopefully we'll see like more of these tools that enable um folks who aren't developers actually to like start coming in and and and building stuff like they couldn't before Yeah You know one thing just getting back to uh you know 2.5 Pro I like I think it's worth gushing about it a little bit and I love that in the keynotes you know it was mentioned that the LM arena and you know I think it came in with like a 39 point uh lead over the second models when it was released How good is Gemini 2.5 Pro It's it's mindboggling to me Like when I use it inside AI Studio like I feel like I'm stealing something because it's so good It can handle so much data like and it's free and and and and it's free inside AI studio like like talk talk about like maybe some of the the best use cases that you're seeing for Gemini 2.5 Pro Yeah that's a great great example and I actually think one of the interesting thing and I had this conversation with some of the folks on the deep my team is like sometimes actually you see like a 40 point jump on some benchmark somewhere and like it actually doesn't even tell the story to the completeness of like just how much better it is Um there's also this like other and then I'll I'll answer your question directly There's also this other thread which is like every time a new model comes there's like an entire class of new companies that weren't possible before that like just become possible and like it feels like that's true when you get this like massive jump in capabilities I think 2.5 Pro is actually one of those models where like there's a bunch of new companies now that are possible I think there's a lot of coding stuff Um it is interesting that it's uh one of the things that makes me most excited is that as you see these like general purpose frontier models like take a step function change in capability it's like across every use case So like you like I think the one that's like didn't work really well before that now works really well is coding So like lots of people are like very excited about the model's ability to do code But like I've seen tons of creative writing examples I've seen tons of people using 2.5 Pro is like a harness to build a gentic products which is a little bit like in the weeds behind the scenes Um yeah So and and I think this actually we haven't even gotten to like a bunch of the yet yet yet to be released like a bunch of the multimodal stuff that I think we're seeing with 2.0 Flash which was another thing I think that happened since the last time we caught up There's too much stuff going on Uh it's hard it is certainly hard to keep up with All right so it it is hard to keep up Uh you know Logan I know you're a busy guy you have you have to go speak to you know thousands of people Uh but you know as we wrap up today's expedited uh conversation because you know maybe we'll have to get you on a third time Uh but you know what are some of the you know even speaking of kind of like a new class of companies right um which is great with like a great way to think of it with Gemini 2.5 Pro but you know what are you most excited about from this weekend and you know um or maybe for the average you know everyday business leader what are you most excited for them to get their hands on and how do you think you know kind of like there's oh a new class of companies now uh is there going to be a a new class or a new way that we do our everyday work because of what was announced here I think the live API is that and we we hadn't talked about it yet but the live API is is basically this and I don't remember if we did demos I don't think so before but if folks haven't tried this out AI studio uh.google.com/live google.com/live has this experience where you can come in and you can talk to the model you can share your screen the model can actually look at your camera if you give it permission and like it creates this really um what what I think is this like future of how people are going to work which is the models can actually see the stuff that you see which I think unlocks like a it it takes it takes the drudgery out of having to use AI tools which like my personal perspective is today the challenge with using AI is that you as the user of the AI product have to go and do a bunch of work to bring all the context to the model And like oftentimes like for me as the person who wants to use AI I'm like the context is already there I'm looking at it on the screen Like why is this so much work to like take that information and go bring it over to whatever AI product and it's like a very simple thing that all of these new back to the thread of like new classes of companies to be built All of these new companies and products to be built You just flip that switch and then all of a sudden like you know whatever the random product is that you're using can see your screen and like help you reason through whatever the problem is that you're trying to solve It can you know bring in real information in real time from Google search It can execute code on your behalf Like all of this like really really interesting stuff that I don't I don't actually think we've seen products built yet this technology which gets me excited because I think it's it's going to be wicked All right It's it's an exciting one uh a fast but furious interview just like what we've seen so far out of this this conference fast and furious updates So uh Logan thank you so much for taking time out of your day to join Everyday AI We really appreciate it Yeah And for folks who aren't watching on video Jordan has a sweet Everyday AI uh Nike shirt which looks awesome Yeah now you're crushing it Now now I'm going to be uh in like just drowning out with requests for it All right Well hey thanks uh thanks again Logan And if you want more we talked about a lot in a very short amount of time Uh it's all going to be in the newsletter So if you haven't already please go to your everydayai.com Sign up for the free daily newsletter Let me know Should we bring Logan on for the third time The first person ever after 500 episodes to be on a third time All right Thanks for tuning in We'll see you back tomorrow and everyday for more Everyday AI Thanks y'all And that's a wrap for today's edition of Everyday AI Thanks for joining us If you enjoyed this episode please subscribe and leave us a rating It helps keep us going For a little more AI magic visit your everyday.com and sign up to our daily newsletter so you don't get left behind Go break some barriers and we'll see you next