there has been a ton of AI news this week and I don't want to waste your time so let's just get right into it starting with the fact that anthropics Claud rolled out even more cool features you can now create what they call projects and a project is kind of similar to a custom GPT over on chat GPT you can upload a whole bunch of files for context and then every time you log into that project and have a chat with it it's going to use that same uploaded information every single time you can even give it custom instructions so it knows how to handle any prompts that you enter into this specific project they added the ability to share projects with other team members and it also got a slight user interface update we can see our standard chat box here select our model here use any of the projects we created by clicking on this little projects drop down here and we can see all of our recent chats here you'll notice your profile button is no longer up here in the top right of the screen now instead if you move your mouse to the left of the screen the sidebar automatically pops in we can see all of our recent chats here you have the ability to Star chats that you come back to often we can access our projects through here as well and if you need to change any of your settings on your account your profile link is now in the bottom left I can click here and access settings change the appearance go to the feature preview and turn the new artifacts feature on or off we talked about that last week it's a new sort of right side window of your screen that shows the sort of code output and things like that that you do with clad and if you want to use clad collaboratively with team members you can also upgrade to a team plan over here but let's go ahead and play with projects real quick and see what it's capable of we can create a project up here in the top right name the project I'll name it research papers and I'll give it the description analyze and respond to questions about uploaded research papers and create our project now we have our project window here and we can set some custom instructions I'll tell it to answer any question I have as concisely as possible about the research papers that are within the project knowledge start every response with no problem mat so we'll go ahead and save the instructions here I have a handful of research papers that I've downloaded and saved let's go ahead and grab a few of them dump them into our knowledge base like this you can see I've uploaded six PDFs here of various research papers and I've used 80% of the knowledge size they give you a maximum of 30 megabytes it's not based on how many files but the size of the overall knowledge base you're uploading let's go ahead and ask what papers are available inside my project knowledge it starts with no problem Matt the papers are available inside your project knowledge are generative agents interactive simulacra of human behavior game based platforms etc etc and we can say what correlations can you find between these papers no problem Matt the main correlations between these papers are Ai and machine learning all papers are focused on Advanced Ai and machine learning techniques generative models papers four and five deal with generative AI models particularly diffusion models papers 1 and two emphasize interactive AI models and it found the relationships between these five documents that I uploaded now what's cool is if I come back to my homepage on Claud and I want to use that same set of knowledge that I've already updated I can just click use project click on the research papers and now you can see that it is set to the research papers project and any prompt that I give it here will reference that knowledge that I already uploaded as well as the custom instructions that I gave it this to me seems really handy especially combined with the artifacts feature which I showed off in last Friday's video but just as a quick reminder of this new artifacts feature I can type something like create the HTML for a sales page about surfboards and you'll notice it opens up a new right hand side of my screen where it actually writes all of the code I can continue to have conversations on the left while the code is generated on the right and I can actually preview the website it generated right here on the right flip between the code and the website preview without ever actually leaving Claude I don't know about you but I really feel like these new features in clad make it really really useful Google also made a bunch of updates and announcements this week including the fact that Gemini 1.5 pro has a 2 million context window that's a combined 1.5 million words between the amount of text input it and output it back to you the entire set of Harry Potter books is about 1.1 million words and the entire set of Game of Thrones books A Song of Fire and Ice are roughly 1.7 million words so you can almost upload that entire series of books and have conversations with it that's how large this model is it's basically to a point where context windows are no longer really an issue if you're willing to use Gemini 1.5 Pro right now now Gemini already let us have a 1 million token context window and I would say probably 99.9% of people didn't even need that large of a context window a 2 million context window while a huge leap and insane that we can do it probably still not going to be applicable to most people but still a really killer feat Gemini 1.5 Pro can also execute code now as well and the Gemma to open source model from Google was also made available this week and based on the Benchmark tests here we can see that this new Gemma 2 model both the 9 billion parameter model and the 27 billion parameter model outperform llama 3 in most benchmarks as well as grock in most benchmarks this week Google also announced that 110 new languages are coming to Google Translate they're using AI to add these new languages including Cantonese andco I don't know how to pronounce that and tamit who's also announced this week that Google is rolling out their sidebar inside of Gmail we can see their example here the Gmail window is on the left and on the right Gemini pops out and offers the ability to summarize conversations show me emails that are unread draft an email or get more suggestions but basically it opens up a chat window that can interact directly with your Gmail emails it looks like in order to use it you've got to be a Gemini business or Enterprise Gemini education or have the Google one AI premium and we can see that the roll outs already started I checked my own Gmail I don't have it yet and I am a Google one subscriber so as soon as I have it available I'll show it off moving on to open AI news it was announced this week that that advanced voice assistant you know the one they demoed several weeks back that they got in trouble for because it sounded a little too much like Scarlet Johansson but everybody thought the feature was really cool well they've decided to push it out even further they say they had planned to start rolling this out in Alpha to a small group of chat GPT plus users in late June but need one more month to reach our bar to launch they're improving the model's ability to detect and refuse certain content they're also working on improving the user experience and preparing their infrastructure to scale to Millions while still being able to maintain real-time responses however it does appear that some people have been starting to get access to it I came across this xost here that shows that it even generates background sound effects while narrating so it appears that somebody asked it to tell it a story and in the story actually included sound effects one night a storm Rose in thunder lightning the whole shebang a couple of teens trying to be brave or just playing stupid decide to check out the house during the storm they sneak in through a broken window as St exploring with nothing but a flashlight and their phones for light pretty crazy as it's telling the story you hear like thunder going on in the background open AI also announced that they trained a new model based on gp4 called critic GPT it's basically designed to catch errors in chat gpt's code output we can see in their example here that they're asking it to write a python function it writes the python function and then critic GPT then reviews The Python function and points out any issues with it you can see that in this specific function it points out that checking if an absolute path of a file is within a directory by using starts with is insecure users could exploit this with Sim links or by naming directories similarly now it does say that critic GPT suggestions are not always correct but we find that they can help trainers to catch many more problems with model written answers than they would without AI help now I'm not sure exactly when this is going to be available for everybody to use in their blog post here they say we're planning to scale this work further and put it into practice the Mac desktop app is now available for all users I did actually make a tutorial on how to use this desktop app if you want to learn how to use it yourself if you were a plus user you've probably already had access to it and been able to use it but now if you're a free chat GPT user you can now download the Mac App and use it as well if you press option plus spacebar it opens up a little chat box kind of like the spotlight window that you already have in your Mac except whatever you type into it will ask the question or give the prompt to chat GPT and speaking of chat GPT and their desktop app open aai actually purchased another company this week they bought a remote collaboration platform called multi if you're not familiar with multi it's a collaboration software where somebody can actually share their screen and work collaboratively on the same desktop you can see it's got shared cursors and drawing simultaneous screen sharing you can click and type into somebody else's app and now this is owned by open aai so it's not too hard to see where open AI is likely going with this we might see a future where when we run into a problem on our computer we handle Over Control to chat GPT it views your screen and then collaboratively solves the problem with you imagine this if you're a coder as well and you're having chat GPT write the code and then you're opening up the output of the code and it's not working properly well chat GPT will likely in the future be able to see your screen offer suggestions and point out Solutions and likely even iteratively update the code based on what it's seeing on the screen now I'm just speculating here but to me this seems like the logical conclusion of where open AI is probably going with this Toys R Us this week which I didn't even know still existed leveraged open ai's Sora to make a brand new Toys R Us commercial but here's a peek at what the actual ad looks like did you ever wonder how Toys R Us and Jeffrey the giraffe came to be the son of a bike shop owner charles Lazarus had a vision that would go on to change toy stores forever now that's just a snippet of the commercial it's not the whole thing but opinions have been very mixed on it a lot of people hate the fact that companies are using AI to create commercials and a lot of people are talking about how maybe this is the future of commercials also it hasn't been claimed that 100% of this commercial was made with AI but we do know that a large majority of it was created with Sora and then I'm sure somebody came back in and did some editing and cleanup on it today's video is sponsored by LTX Studio which is perfect because they recently rolled out a bunch of new updates this new pack of updates is called Visions this update gives you a ton more flexibility over creating your videos like the ability to completely start from scratch with a new video or paste in a script and let LTX Studio make the video based on your script with my script broken down into scenes I can simply click Start and in less than a minute I have a complete storyboard created from the script that I just uploaded but check this out maybe I want to turn my whole video into a completely different style as part of the New Visions update I have the ability to drag and drop an image like this into the style reference in LTX studio and then watch as it reworks my entire storyboard to match that new style that I just uploaded once I'm happy with every single shot in every single scene of my video I can generate each video clip individually or I can come up here to export and click generate missing shots to generate the entire video all at once they also rolled out a pitch deck feature with this latest update where you can export your entire storyboard and it will turn it into a beautiful pitch deck that you can use to sell the story that you're creating here and this is just scratching the surface of what you can do with LTX studio it'll also create voiceovers create a Musical soundtrack allow you to use a consistent character throughout your video edit the location the lighting and the weather of every single scene and so much more it's actually a pretty impressive Suite of AI video editing tools you can join the wait list for LTX studio today by clicking the link in the description below they are accepting users now and thank you so much to LTX studio for sponsoring this video and even more open AI news they partnered with Time Magazine on a strategic content partnership time and open AI announced a multi-year Content deal and strategic partnership to bring time trusted journalism to open ai's products including chat GPT open AI will gain access to current and historic content from times extensive archives from the last 101 years the partnership will also enable time to gain access to open AI technology to develop new products for its audiences it is very interesting to me how the sort of media narrative has really shifted lately obviously the New York Times are still in the middle of a lawsuit with open AI but almost all of the other media companies are coming out and saying we're partnering with open AI we're partnering with open AI but in the early days most of these media companies were really really resistant to this AI technology saying why would people come and read our websites anymore if they can just go talk to chat GPT we know that chat GPT most likely trained on all of this data from all of these sites and now chat GPT is getting ahead of even more potential lawsuits by partnering up with a lot of these companies that create this content so that there is still some sort of in incentive for all of these companies to continue to create content report on the news do Deep dive research and even if people find that information through openai it's still in some ways going to benefit the media house because open AI is apparently licensing this content in some way and that wasn't the only partnership that time made this week time also partnered with 11 Labs the company that does really realistic text to speech apparently time is trying to go audio native basically it looks like if you go to the time website they're going to have embedded audio players with realistic voiceovers that you can just click play on and it's going to sound like a real human is reading that time article back to you here's an actual example that's live right now on the time website if I scroll down we can see this player here and if I press play headline We are the world power how Joe Biden leads by author Massimo calibra Sal Washington you can read the trans scpt of the interview here and the fact check here and we can see in the corner it's got the 11 Labs link and in the future I think most websites are probably going to go this direction not just time I think even like WordPress blogs and things like that you're going to see a lot more and more of them integrate Tech from companies like 11 Labs so you have the option to either read the article or blog post or listen to it in a realistic voice and 11 Labs even allows you to do things like train your own voice into the system so when I create blog posts you might be able to go to that blog post press play on it and hear the blog post as if it was read by me which is pretty cool that's not the only news we got out of 11 labs this week they also introduced a new reader app where you can listen to any text on the go with the highest quality voices this 11 Labs reader app lets you listen to any text content on the go so now any article PDF UB newsletter or any text on your phone can now be turned into into an audio that you can listen to with a realistic voice the app is available right now on iOS and coming to Android really soon I haven't played around with it too much yet but I am excited to use this to listen to a lot of the news articles and research papers that I've been meaning to get around to reading also this week character AI rolled out a new feature that allows you to talk to the AI avatars now character AI is a very interesting company they are huge especially among the younger demographic character AI is probably the most popular AI tool for like younger Generations but they just rolled out this feature where you can have phone calls essentially with fictional AI characters if for example I wanted to have a conversation with Spiderman I can try with voice ask it something like what's your favorite food oh man that's tough I love a good pizza that's pretty much the standard but I'll also never say no to dumplings or just generally spicy stuff how about you now this feature isn't the new feature the new feature is that you can actually call and have this conversation over the phone you can see here when I hover over this little phone icon it says calls available on the app I'm not going to download the app to demonstrate it because I'll probably never use it again but you get the idea same concept just like you're on a phone call instead of in a text chat and while we're on the topic of AI and audio all those fun AI music tools that we've been playing with lately you know Sony and yudoh and things like that well the record labels are starting to sue those companies claiming that they were trained on copyrighted music that they didn't have the right to use now this is sort of like a muddy area because they're going to claim that that was fair use the record companies are going to fight that it was a massive infringement on copyright but there really is no precedent for this yet so it's going to be interesting to see how it plays out if you want my opinion on it I feel like the record labels are going to really really be fighting for relevance in a world where anybody can kind of produce their own music the days and record labels actually do less and less of the marketing for musicians I feel like record labels are losing relevance really really quickly and some of these lawsuits are kind of them trying to stay relevant and grasp for the power and the control that they currently have but it's going to be something fascinating to watch and see how it plays out because it is another area where I can kind of see both sides and understand the arguments on either side YouTube on the other hand is trying to avoid this exact scenario this week it was announced that YouTube is trying to make AI music deals with the major record labels and get in front of this according to the Financial Times YouTube is offering to pay Universal Music Group Sony Music Entertainment and Warner Records lump sums of cash in exchange for licensing their songs to legally train its AI music tools this apparently isn't for their existing dream track feature but it's going to be used to train new AI tools that YouTube is planning to to launch later this year again this is a flip side of the coin from the last article which is going to be interesting to watch to see how this one plays out because if these record labels are willing to sign some of these licensing deals to allow music to be trained well then that could be the solution for companies like sunno and yudo go and agree to terms and get the licensing and make all this go away we we'll see how it plays out it's going to be interesting to watch nonetheless and I'll do my best to stay looped in and tell you what's going on as this plays out this week figma had a big event called the config conference and they made a whole bunch of AI announcements around what you can now do inside of figma it also got a complete redesign and overhaul figma's now going to be able to take text prompts and create designs from those text prompts they showed an example of how figma could create an app design for a new restaurant a few seconds after they typed the prompt figma mocked up an app with menu listings a tab bar and even buttons for delivery Partners like uber eats and door Dash they claim it was a pretty generic mobile mockup but it was better than starting from a blank slate and you were now able to edit and reconfigure the design based on the template that it created for you to start with it can also add autogenerated text instead of using the generic laurum ipsum placeholder text that people tend to use on websites and it has a new AI enhanced asset search to better help you find the various assets for designing your site now right now they're rolling this out as a beta feature and it's free for figma users in the beta mode but eventually it is going to be a paid feature the company Synthesia announced a handful of new updates this week they're a company that makes these like AI avatars that are kind of designed for having you know AI sales Bots and things like that supposedly the avatars are more expressive showing excitement and doubt and fear and anger and things like that they've also got what they call personal avatars which can actually look like you and use your voice and sound like you but also speak in other languages they're also adding the ability to do full body avatars instead of just you know from the neck up and a handful of other features I'll link the entire YouTube video up if you want to learn more about what Synthesia is up to a new AI video editing tool rolled out this week from the company captions you can use this tool and feed it a standard video and it will edit the video with different backgrounds make it look like you've got a green screen add text like you're on the screen right now add stuff that looks like after effects animations it looks like a pretty cool tool now I haven't played with this one myself yet but I think this one might deserve another video in the future if it turns out to be a pretty cool tool look out for that video later on and we'll see how it does a few months back I talked about a tool called magnific which is this really good upscaler tool but it kind of hallucinates extra stuff into the image and sort of adds this creative flare to the image well they just rolled out this new relight feature and it lets you take your images and sort of reimagine like for example this thumbnail here of me holding up all these glasses I reimagined it with me in outer space simply upload an image right here you have the option to transfer from a prompt another reference image or a light map but let's just go ahead and do from a prompt I uploaded an image of people dancing and let's just add people dancing underwater I'll leave everything the same although we do have some other style options and things and if I click relight you can see it relit it like this here's the original image here's the new image that it made it does a really good job of completely restyling the image this week's stability AI got a new CEO who was the former CEO of WETA digital if you're not familiar with weda digital WETA is a visual effects company that's worked on films that you might have heard of you know things like Avatar Planet of the Apes the Marvel movies Transformer Game of Thrones and is probably best known for a little series of movies called The Lord of the Rings things keep getting worse for the rabbit R1 this company called rabbit tude which is essentially a company who's trying to poke holes and sort of find flaws in the rabbit R1 well they uncovered a huge data breach the team gained access to the Rabbit Code base and found hard-coded API keys in the code which would allow anyone to read every response every R1 has ever given including ones containing personal information brick all the r1s alter the responses of all the r1s and replace every r1's voice they found the API keys for 11 Labs azir Yelp and Google Maps in there on top of all of the bad press and all of the bad reviews the rabbit has received it also turns out they did a pretty poor job with code and left a bunch of API Keys exposed meaning that anyone that has the rabbit R1 is at risk of having their data leaked a company called etched announced a new chip which is apparently the world's first specialized chip for Transformers the Transformer is pretty much what all large language models that you use are built on top of it's also what Sora Gemini stable diffusion and all of these tools use and these chips are more than 20 times faster than the Nvidia h100s which is what most of the AI companies these days use to train their AI models this looks to be pretty impressive when they become available and they also appear to be a competitor to grock the grro Q grock which I just so happened to be holding a pen for and the day they announced this grock announced that they've achieved 400 tokens per second with the Llama 370b model this message appears to be a direct response to the etched comment as pointed out here by AE meta started testing user created AI chat bots on Instagram we've seen some of these AI chat Bots with like Snoop Dog but under a different name and Mr Beast under a different name and I think Tom Brady under a different name well now they're starting to roll it out so individuals can create their own custom chat Bots it appears to be rolling out an early test in the US of the AI studio so you might start seeing AIS from your favorite creators and interest based AIS in the coming weeks on Instagram they will primarily show up in messaging for now and will be clearly labeled as Ai and finally with the Olympics coming up they are going all in on AI this year including an AI version of Al Michaels people can get customized feeds of what they're interested in from the Olympics and it will be commentated on by Al Michaels but it's AI Al Michaels here's a quick example of what that might look like hi Kelly welcome to your daily Olympic recap your personal rundown of yesterday's most thrilling Olympic moments since you're a swimming fan let's head right to the pool Team USA secured a stunning victory in the men's 4X 100 meter medley relay smashing the world record over at the diving venue Christ the Palmer showcased resilience and skill overcoming past knee surgeries to qualify for the women's springboard final I'll go ahead and stop it there you get the idea you're going to be able to tell it what sports you're interested in and it will create a custom little video of a recap with Al Michaels commentating on that recap even talking to you by name if you want to and that's what I got for you a lot of AI news came out this week it was weird cuz it kind of felt like a slower week to me but as I was rounding up all of this AI news I was like holy crap there's a lot going on it's gotten to the point where 30 new AI announcements in a single week feels like a slow week which is just wild but anyway if you enjoy videos like this and you want to stay looped in on the AI news give this video a thumbs up and make sure you subscribe to this channel I'll make sure that videos like this keep on showing up your YouTube feed and if you want to find more AI tools to play with and stay looped in on the AI news on a daily basis check out futur tools. where I keep the news up to date every day I'm adding cool AI tools as I find them and I've got a free newsletter you join the free newsletter I'll hook you up with the AI income database a database of cool ways to make money using these various AI tools and you'll be entered to win cool prizes in our monthly competitions which is exclusive to people who subscribe to both the YouTube channel and the newsletter so again head on over to futur tools. to find all that stuff thank you once again to LTX studio for sponsoring this video and thank you so much for tuning in nering out with me and uh keeping your finger on the pulse of the latest in the AI world thank you so much once again I really appreciate you I'll see you in the next video bye-bye