Transcript for:
AI News Highlights

I'm in the final days of my vacation out here in Colorado just about ready to head home but that does not mean I have not been keeping up with the AI news there has been a ton a ton of AI news this week and I'm excited to dive into it with you right now so let's get into it so of course the one week that I'm out on vacation happens to be probably the biggest week in AI that we've had so far this summer because this week is the week that meta finally released their newest version of llama called llama 3.1 now llama 3.1 is an upgraded model of the previously released llama 3 and it now comes in three sizes instead of just two we've got our 8 billion parameter model our 70 billion parameter model and our 405 billion parameter model this is essentially the amount of data that was trained into the model the bigger the model typically the better it is at things like reasoning and math and just generally has the most amount of use cases for it some of the things that these new versions of llama can do are things like use tools it's got multilangual agents so for an example translate the story of Hansel and gredle into Spanish it's better at complex reasoning and it's better at coding if we look at some of the benchmarks that meta shared we can see here that a lot of the benchmarks actually outperform all of the other state-of-the-art models in a lot of areas the current sort of leaders are gp4 Omni and clad 3.5 Sonet and we can see that this llama 3.1 45 billion parameter model actually outperformed these other models in a lot of the various Benchmark testing so we actually have an open-source model now that is pretty much on par with what you get out of things like gp4 Omni and Claude 3.5 Sonet and when looking at their 8 billion parameter model the smaller model that they just released it pretty much outperforms in basically every area compared to the other similar size smaller models but what really makes this one powerful is the fact that it is open source you can actually download these models they're publicly available and anybody can update them and fine-tune them something that can't really be said about clad 3.5 Sonic or GPT 40 we can see here it says developers can fully customize the models for their needs and applications train on new data sets and conduct additional fine-tuning so developers and Engineers now have a model to play around with and fine-tune and optimize for whatever they need it for that's as powerful as something like GPT 40 and Cloud 3.5 Sonic this is huge for the open- source AI Community there's one slight caveat to the open sourness of this and that's that if we look at the additional commercial terms here under the licensing agreement it says if you have more than 700 million monthly active users you must request a license from meta which meta May Grant to you in its sole discretion so if you use these llama models to build something that gets more than 700 million active users then you need to sort of work it out with meta and they're probably going to want a piece of that pie or something I don't know now most people aren't going to be able to download and run the 405 billion parameter model on their own local machines not only are the files hundreds of gigabytes but you're going to need more vram than most people have even if you're running a 4090 from Nvidia you're probably still not going to be able to run it on your computer however you can go and play with these models in a couple places you can use it at AI you can use it in WhatsApp you can use it in Instagram messenger Facebook Messenger all of the various messaging platforms that Facebook offers are now using the latest model and if you have the meta rayb bands as well they're getting the latest model built into them as well I don't think it's in there yet but they're going to be getting the latest versions of llama built into them as well but probably the easiest place to go and play around and test the models right now is just over at meta AI you can see we have a chatbot here if I ask the chatbot what llm are you currently using using it'll tell me I'm meta Ai and I'm based on llama 3 a state-of-the-art language model developed by meta I'll say which version of llama are you running I'm a large language model based on meta llama 3.1 it's not really giving me the answers outright but I'll say how many parameters and you can see this one's based on the 70 billion parameter model here but if you want the absolute fastest way to use llama 3.1 well it was just rolled out onto grock's platform as well we can see here they put out an article that says grock supercharges fast AI inference for meta llama 3.1 grock is proud to partner with meta on this key industry launch and run the latest llama 3.1 models including 405b 70b and 8B at grock speed now if we jump over to gro.com where we can actually check we can see up in the right corner we can change which model we're using we now have llama 3.18 billion instant and llama 3.1 70b versatile we don't have llama 3.14 45 here in grock because apparently they're only making that available for Enterprises at the moment but if we go ahead and select llama 3. 170b here and ask it a question or give it a prompt for example explain the trolley problem and tell me how to solve it watch how fast this responds this is all real time I didn't uh pause anything it generated at 250 tokens per second and the entire output was back to me in 3.22 seconds with this whole explanation here but if you do want to use the 405 billion parameter model it was also built into perplexity right after it was announced aravan says llama 3.1 405b the most capable open source model is up and running on perplexity for All Pro users so you do have to be a pro user but if we go over to perplexity doai click into our settings and then come down here to AI model I typically use it with Cloud 3.5 Sonet I now have the option to run it on llama 3.1 45 billion parameter so let's go ahead and select that go home and now any prompt I enter into perplexity here is going to be running this newest model from meta I use perplexity a lot these days I'm mostly switching back and forth between clae 3.5 Sonet and perplexity I use clad when I want to sort of work through problems and answer questions and get help summarizing things things like that and I use perplexity if I know it needs to search the web or if I need to do some deep research and now we can also use llama 3.1 405b inside of perplexity now if llama 3.1 was the only that happened this week that would have been a pretty big week in AI on its own but that's not all that happened this week mistol AI another company that sort of competing with meta for creating the best open-source models released mistol large 2 now mistol large 2 is a 123 billion parameter model which based on the headline from mistl is large enough we can see from their little charts here on their web page that mistol large 2 is actually outperforming the Llama 3.1 70b the newest model in math performance and here's some more Benchmark comparisons from this mistal model we can see from the human eval Benchmark mistol large 2 just slightly underperforms GPT 40 it outperforms Claude 3.5 Sonet and it outperforms llama 3.1 405b the newest model from meta and it pretty much is on par with the best state-of-the-art closed and open source models in all these other benchmarks here when tested on code generation this mistol large 2 outperformed all these other models in Python C+ plus bash Java typescript PHP and C but noticeably missing from this list here is cloud 3.5 Sonet which if I had to guess they left that out because Claude 3.5 Sonet probably beat out mistol in a lot of these but I'm just speculating why did they put CLA up here in these comparisons but didn't mention it in these comparisons down here I don't know the point being here is that we are getting some really great open- source models that are starting to compete with the bigger closed Source models from company like anthropic and open Ai and when we get open source models like this it's just the beginning they're going to get fine-tuned they're going to get optimized and they don't have the same rules and boundaries and biases and things like that that the closed models do have I mean maybe these initial versions might but because they're open source and other people are going to go and sort of iterate off of them a lot of that stuff's going to be yanked away and perfected and optimized over time so this is a really powerful moment in time for AI especially open source Ai and Apple's getting into the game as well Apple showed off a smaller model this week a 7 billion parameter and a 1.4 billion parameter model and they're making these open source as well their new model outperformed mistol 7B and they claim they're getting close to llama 3 and Gemma but this is obviously before llama 3.1 came out so who knows how close it is to the newest small model from meta Google's Gemini model got a big upgrade this week as well according to Google's article here they're upgrading the free tier of Gemini to Gemini 1.5 Flash and they claim you'll notice improvements in quality and latency with especially noticeable improvements in reasoning and image understanding and they've also expanded the amount of tokens on the free version of Gemini to 32,000 tokens and according to this to get the most out of the larger context window we'll soon add the ability to upload files via Google drive or directly from your device which is already available in Gemini advance but will soon be available in the free version of Gemini so you'll be able to throw documents inside of Google drive and use that for context for whatever you're prompting inside of Gemini they're now displaying links to related content for fact seeking prompts in Gemini so if you think it's hallucinating you can actually double check the source that it's getting that information from they're adding directly in Google messages on select Android devices and a handful of other features are all coming into Gemini so all of the companies building large language models seem to have really big news this week and of course you can't have a big week in AI from meta and Google and apple and mistol without open AI getting in the game and trying to overshadow a little bit with their latest announcements including the fact that now you can actually fine-tune GPT 40 and you can do it for free for up to2 million training tokens a day now this is a sort of limited time thing this is only through September 23rd but they're trying to get people to get in use GPT 40 and fine-tune it on their own custom data so if you're a company that wants to fine-tune it on your own internal documents or you want to fine tunit to just work with sort of Health applications or biology applications you can go and do that and experiment with it for free to up to 2 million training tokens a day right now this week open AI announced search GPT now there was speculation that this was going to come out months ago back when we got the GPT 40 announcement a lot of the speculation was that this was going to be the announcement this search GPT Sam Alman here says we think there's room to make search much better than it is today we're launching a new prototype called search GPT they put out a blog post about it with some more explanation of what we can expect they do say it's a temporary prototype of a new AI search feature if we look at their little sort of demo animation here it looks like you can ask it a question and it gives us a response that looks very similar to what you get out of perplexity or even Google search with AI or Bing search where it actually gives you the answer with sources maybe some images some weather data was in that screenshot we can see here they do a example prompt of music festivals in Boon North Carolina in August and we've got a response here where it shows the various music festivals with images and links to the actual source of the information there's a little link button that opens up actual search results like you might see in something like Google so it looks pretty promising but it also looks kind of like what we get from perplexity right now and of course they're rolling it out to a select number of people to test it but you can join the wait list and maybe if you're lucky you'll be one of the people that gets to test it before everybody else and we also got another little tease from Sam Alman Marco here says when are we getting voice and Sam mman actually replied to him and said Alpha roll out starts to Plus subscrib next week so if you're a chat GPT Pro user we might be getting to see the voice feature that they showed off and got in trouble for because of the Scarlet Johansson stuff we might get to see that like next week maybe we'll see there's also some news out of anthropic this week albeit not as favorable or good news it looks like anthropics scraping Bots are scraping websites at a insane rate even when the websites are telling them not to scrape my website we can see Kyle wans here says hey anthrop AI I get you're hungry for data CLA is really smart but do you really need to hit our servers a million times in 24 hours you're not only taking our content without paying you're tying up our devops resources not cool and according to them they specifically are asking in their terms of service for companies like Claude not to scrape them now just adding it to your terms of service I don't believe is going to prevent some of these Bots from scraping your site but theoretically you should be able to add it to your robots. text file which should tell a lot of these Bots hey don't scrape this site and if I scroll down this thread a little bit here it says our terms of service banned machine learning training before their crawl afterwards we added them to robots. text so it looks like they're doing the proper things to avoid scraping and anthropics been scraping it anyway and this is going to be a hot debate over the coming months probably the coming years and we're even going to talk about it a little bit more in this video a little bit later when I get into some of the news around AI video but I'm still in the large language model weeds with you right now and xai and grock is another pretty substantial large language model while Elon Musk was on an interview this week he said that grock 2.0 is coming very soon and it's going to be on par with GPT 40 or Claude 3.5 Sonet but he also went on to say in this interview that grock 3.0 will be the most powerful AI in the world and they're expecting to release it by December now Elon sort of has a tendency to say something is coming at a certain time and then sort of underd deliver on that time promise so we'll see if we actually do get it by December but he's claiming that this will be the most powerful AI model that we've ever seen and to sort of back that up a little bit here we can see this other tweet from Elon saying nice work by xai team Nvidia and supporting companies getting the Memphis super cluster training started at 4:20 a.m. local time knowing Elon he probably specifically picked that time on purpose but I won't go there he says with 100,000 liquid cooled h100s on a single RDMA fabric it's the most most powerful AI training cluster in the world he's got the most powerful AI training cluster in the world at his disposal so it's not hard to believe that grock 3.0 could be the most powerful model we've ever seen but we don't know what opening ey is going to do maybe GPT 5 is going to come out before then and he's going to have some more catching up to do we just got to wait and see and since we're on the Elon train right now talking about tweets and things that he's announcing and predicting lately he said that Tesla will have genuinely useful humanoid robots in low production for Tesla internal use next year and hopefully high production for other companies in 2026 so he's saying these Tesla Optimus robots that we've all been seeing videos of he's going to be using them at Tesla next year and other companies are going to be able to use them as part of their Workforce by 2026 that feels fairly optimistic to me but it also feels very plausible Bing's AI got a bit of a redesign this week if we take a peek at this little video here you can see that when they give it a question it actually answers the question on the left side and the normal search results that you're used to are actually pushed over to the right side we can see here's another screenshot how long can elephants live the anwers right here with some more details the source of the actual response that it gave and then the traditional search results that you'd expect are down this right sidebar now I did a search on my own Bing account and it's still showing the old way but I guess apparently it's rolling out to some or rolling out soon I don't know again I'm still seeing the old way when I do a search all right shifting over to AI video news Luma AI just released a cool new feature called Loop they said today we are releasing Loops in dream machine to keep your imagination going and going and going get started here now if you've used dream machine already what I will say about it is that it's not super great at text to video yet but if you generate an image put the image in and then try to turn that image into a video it's actually really really good at that and now they've added a new looping feature where it can infinitely animate something over and over again like this flaming Loop video or this spaceship perpetually flying through space or a capy Bara riding a bicycle or this top that perpetually spins or just the prompt Loop that just keeps on looping now if we head on over to Luma labs. a I can take an image drop it in here and we can see that I've got a loop option now I simply click the loop option here and give it my prompt I'm just going to put colorful swirling background make sure a loop is checked I'm going to uncheck enhance prompt here and let's go ahead and see what it generates for us I'm just throwing up one of my old thumbnails to see what it'll do with it so it may not be great but I wanted to show you a quick example and here's what it gave me which I'm actually pretty impressed with you could actually see AI news flash onto the video for a second cuz that was in the original thumbnail but it just Loops this animation forever and ever and ever now I'm actually pretty impressed with this looping feature I think it's pretty cool we'll definitely have to do some more playing around with it in future videos but I want to move on and share with you even more AI video news because this week cing which a lot of people claim is the best text to video generator I mentioned Luma is great for image to video cing is probably the best text to video out there that comes closest to what we've kind of seen with things like Sora it's also fairly uncensored where you can actually plug in celebrities names and use IP and stuff like that right now we can see an example here of some you know baby puppies or baby wolves here all I guess eating something here's another one of like a samurai with a flaming sword running through the city humanoid robot like walking through a futuristic city it does pretty good text to video now if we head on over to cing ai.com we can actually log in and we don't actually need to give it a phone number before people were using workarounds and getting Chinese phone numbers and plugging them in to get inside of it but now all you got to do is register with an email and you can actually use it and you do get some free credits every day that you can play around with if I click on AI video Let's do one of my old standbys a monkey on roll skates and we'll generate you can see it takes about 10 credits they gave us 66 credits every 24 hours so roughly six videos a day let's go ahead and generate and after a few minutes this is what we got it actually looks like a monkey on roller skates I mean it's not amazing yet I still haven't gotten the perfect monkey on roller skates but I figure a few more rolls of the dice and I probably would come up with something pretty decent they also have some other cool features here like the ability to mess with the camera movement negative prompts the the ability to change between high performance and high quality changing between 5 seconds and 10 seconds different aspect ratios things like that something will definitely need to play with again in a future video which I'm really really looking forward to but I want to move on with the news here and circle back around to that topic of where the training data is coming from on a lot of these tools because this article came out this week on 404 media about how Runway was trained on thousands of YouTube videos without permission in last week's news video I talked about how companies like anthropic and apple and a bunch of these companies building large language models were scraping the transcripts from YouTube videos without permission as training data well now it turns out that Runway was apparently training on YouTube videos however it's a little bit different because there's no real confirmation that they're doing this it was just sort of leaked information from an anonymous source who then shared a spreadsheet of YouTube channels that supposedly this was trained on it's all sort of confusing and vague but there's a spreadsheet here and if we click on this tab that says recommended channels supposedly these are the channels that the videos were pulled from to train Luma but again it's from an anonymous source and they were using vpns and things like that to hide the fact that they were scraping so it's really hard to 100% confirm that Runway actually did train on these channels but I mean there's some pretty big channels in here Marquez brownley Casey neistat Mr Beast a lot of these channels are in this list of supposed videos that were scraped and used for runways training now I have very mixed feelings about this personally I don't really see a huge problem with a lot of these videos being used to train if the outputs that we're getting are substantially different than anything that it was trained on however if people are using these tools and managing to get videos that look like some of those creators out of these videos then it starts to feel wrong but there's also even more gray area than that because there is a lot of copyrighted content from you know TV networks and things like that that end up on YouTube because of commentary channels so let's say a commentary Channel watches The Real Housewives of Orange County or something like that and does commentary over it well the video footage which was aired on TV and then ripped to a YouTube channel and overdubbed with commentary well the video footage is still video footage from these TV shows and it just gets very very muddy and I don't know how this is going to be solved I think the general consensus among the AI people is that if it's publicly available and we can train our own brains with this data then why should a computer or company not be able to train that's the sort of arguments that a lot of people are making and I just out of curiosity did a poll on my Twitter account here and I asked if someone puts their content on YouTube for the world to see should it be fair game for AI companies to scrape and useing training data 54.5% of people said yes 28.3% of people said no you shouldn't be able to use it and 17.2% said I'm conflicted personally I fall in that category this is based on 2,169 votes but I also realize that my Twitter following is probably fairly biased towards very Pro AI so this is sort of the results I expected but me personally I'm still a little conflicted but at the end of the day I feel like if the videos that are being outputed from these systems are so far different than all of the stuff it was trained on it probably is going to fall under fair use but then you have those complications of sort of commentary channels and videos that were never meant to be trained which it just complicates it and I don't have the answers I want to see cool stuff get made but I also don't want to see stuff being made that actually hurts the creators the data was initially trained on and right now I don't really feel like it's hurting those creators but I don't know I'm sure there's some arguments out there that could change my mind my opinions on this stuff are very very fluid stability AI also released a new model this week called stable video 4D now what's really cool about this stable video 4D is that you can input a video of a single object and then it sort of predicts what that object would look like in video form from different angles so it says here stable video 4D transforms a single object video into multiple novel view videos of eight different angle SL views we could see the examples here you know we've got this camel and I believe this was the initial video that we're seeing down in this Cube but then you see this little simulated camera showing what this video would look like if the camera angle was changed on that video here's another example of a flag here this is the original footage and then you can see the cameras are pointed at different perspectives on that flag as the camera moves around we get different views of that original video now it is available over on hugging face right now so you can download the files here it doesn't look like anybody's used it inside of spaces yet so there's not like a free version that we can use but I imagine within the next few days there'll be platforms online that have used this model that will allow us to use an online cloud-based user interface to play with this I just haven't found one yet and I can't install it locally on my own computer because I'm still on vacation in Colorado Adobe also rolled out some new AI features inside of Illustrator and Photoshop we can take a peek at this example here where somebody used illustrator to draw this butterfly and then they gave the prompt array of spring flowers and leaves and then the AI actually filled in the butterfly with the flowers and leaves here's another example of just a square drawn in Adobe Illustrator they gave the prompt Flora and Fauna of navy blue and you can see it actually generated this in a pattern if I fast forward here a little bit you can see that it actually extend this out into a duplicating pattern so that's a pretty cool feature that was just rolled out into illustrator and there's a handful of smaller releases inside of adobe as well I will obviously make sure that this is linked up in the description so you can check some of them out while we're on the topic of AI image generation Leonardo just rolled out teams inside of Leonardo where people can collaboratively work on images and Creations inside of Leonardo multiple people can get in there and iterate and build off of images that are created in there we can see here some of the features shared Team collections so you can collaborate on the whole collection of images consistent outputs so you can build fine-tune models that are used between users which will help you generate consistent assets so if you're Building images for games or something like that you can make sure if multiple people are working on those game image assets that they all look very consistent no matter who's using the platform at what computer they're using it there's a shared team feed and a handful of other features available inside of this new release a company called Sakana AI also dropped an image model to generate Japan's traditional ukio e artwork I'm not sure if I said that right this was a company that was founded by some ex Google members who went to create their own AI company it does appear that this new model that they created uses a sort of fine-tuned version of stable diffusion XL the model is available over on hugging face and you should be able to use it in tools like automatic 1111 or comfy UI or whatever you use to generate stable diffusion images sunno the AI music generation company also rolled out a new feature this week called stems Pro and premere users can now separate the vocals and instrumentals from songs which will give you more control over how you use sunno so you can break out just the drum tracks just the bass tracks just the vocal tracks things like that to get more creative with whatever music you're generating and while we're on the topic of AI audio generation this tweet went viral this week uh from Matthew Sabia there was a voice recording of Joe Biden going around talking about how he was dropping out of the presidential race here in the US and Matthew here claims that he took that audio pulled it into 11 labs and 11 Labs claimed that it was fake audio and that this was actually audio generated with 11 Labs however 11 Labs stepped up and said that's a lie he actually pulled in a completely different audio clip made it look like it was the Joe Biden clip but it really wasn't the Joe Biden clip was probably legitimate but he edited a video to make it look like he was dropping in the Joe Biden clip but it was really a different actually AI created clip that he was dropping in so it gave him an output saying No this was AI generated it was all just basically a hoax on Twitter so if you saw that recording going around basically claiming that the Joe Biden audio clip was made with 11 Labs 11 Labs claims no it wasn't that was faked and finally I thought this was really really cool I used to love the old college football games I loved them more than the Madden video games I always thought those games were so much fun but they got discontinued over sort of arguments about the players being added into the games but not getting compensated but because college players can actually get paid now they are actually starting to make college football games again and they used AI to quickly scan all of the players from the NCAA into this game so you can play as your favorite college football team and the players on those teams faces are actually in the game they actually used an AI model plugged in 11,000 different faces of these players and had AI actually generate those players in the game I just thought that was really cool this is a use case where AI rapidly sped up the process of game development for a very high-profile very bigname exciting game and I just think that's awesome cuz I'm a gamer and I'm an AI nerd and these are my two worlds colliding here so I'm happy to hear this anyway that's all I got for you I'm actually going to be home next week my Colorado vacation is coming to an end this weekend I typically would test out more of these tools and play around with them but since I've been on vacation I've sort of slowed the pace of my videos but I've seen so much stuff this week I can't wait to get home and start playing with more of these AI tools and making more videos upping the pace of the videos that I'm doing and it's going to be some fun weeks ahead with all of this cool new AI stuff rolling out and I can't wait to make more videos about it for you if you want to see more AI tutorials more AI news cool behind the scenes stuff from AI companies that are building all this stuff make sure you subscribe to this Channel and like this video it will make sure more videos like this and all this cool AI Tech show up in your YouTube feed and as always if you haven't already check out futur tools. where I curate all of the coolest AI tools that I come across I share all of the AI news is pretty much updated on a daily basis and we've got a free newsletter where a couple times a week I'll update you with just the coolest AI tools I've come across and just the most important AI news that I found all directly to your inbox totally free you can find it all over at futur tools. the site is actually in the middle over an overhaul right now I'm working with a developer there's going to be some Community elements some commenting and ratings and all sorts of really cool stuff to make this a much more interactive community- based site to help you find the exact tool for what you need so again check that out at futur tools. and thank you so much for nering out with me and I just just love talking about this stuff I'm so excited to make these videos every week and share all the cool stuff that I'm coming across in the AI world I hope you appreciate it I hope you like videos like this I hope you subscribe to channels like mine and thanks again for nering out I will see you in the next video bye-bye