Transcript for:
Recent Advancements in AI

the past month has brought some incredible advancements in the world of AI a new AI can now generate videos that look even more real than reality itself Claude 3.5 has emerged giving open ai's GPT 40 a serious challenge Google Deep Mind has made a breakthrough with AI brain unlocking the secrets of real human brains meanwhile Elon musk's xai has introduced its Socrates model and we have updates on gp5 along with notable moves from Apple and meta in the AI space Apple's integration of chat GPT has sparked quite a reaction from Elon Musk a company claims to have created the body for AGI and Google Deep Mind has launched v2a additionally there's a new free Sora AI rival called cling and Dot is quickly becoming everyone's new AI best friend for gamers the first ever gaming AI assistant has been released open AI is ambitiously planning to power superhuman AI with nuclear energy but they're also facing significant challenges we cover all these exciting updates and more in this video so let's Dive Right In if you've been following all the latest AI news and events then you know everyone's been going nuts waiting for open AI to drop their new Sora model but while we were all patiently waiting this Chinese company called qu show just went and dropped an absolute Game Changer out of nowhere they released this wild new AI model called cling and man it is blowing everyone's Minds with what it can do so so let's get into it cing is a video generation model developed by quo a company you might know from their popular app qu this new model is like Sora but some are saying it might even be better in some areas what's really cool is that cing is open access meaning more people can get their hands on it and see what it can do let me paint a picture for you so a prompt that says a Chinese man sits at a table and eats noodles with chopsticks cing takes this and generates a video that looks almost too realistic and if you remember that Infamous video of Will Smith eating noodles that looked like something straight out of a horror movie well King's version puts that to shame it's like night and day seriously hey Uncle Phil come try this cling can generate videos up to 2 minutes long with just a single prompt in full 1080p quality at 30 frames per second which is actually pretty impressive it accurately simulates real world physical properties which means the videos it creates don't just look good they behave like real life videos too the magic behind cling lies in its diffusion Transformer architecture this technology helps it translate Rich textual prompts into Vivid realistic scenes plus clling uses a proprietary 3D vae that's variational Auto encoder and supports various aspect ratios thanks to variable resolution training basically it can handle different video dimensions and still produce high quality output one of the standout features of this model is its Advanced 3D face and body reconstruction technology it allows it to create videos where characters show full expression and limb movements all driven from from a single fullbody photo so essentially this Tech is what makes the videos look so lifelike and consistent now this whole thing is making it pretty clear that China is seriously stepping up its game when it comes to AI development cing is just a glimpse of what's coming out of the country and it's giving us a hint that they might be ahead of the curve open aai has said they'll release Sora by the end of the year but with cing already out they might have some catching up to do the big question though is whether China will make clling available worldwide right now it's accessible through the qua app but you need a Chinese phone number to use it now cing isn't the first AI video generation model from China back in April qu show released vdu AI which could create 16-second videos in 1080P resolution cling is the next evolution of that offering longer videos with even better quality if you check out King's website you'll see some of their demo videos and let me tell you they're pretty impressive for example there's a video of an emperor angelfish swimming in a rocky underwater habit at a man riding a horse in the GOI desert during sunset and even a white cat driving a car through a busy Urban Street these videos showcase the model's ability to handle complex scenes and movements while maintaining high quality the technology behind cling involves a 3D spatiotemporal joint attention mechanism which helps it model complex movements and generate video content with larger motions that conform to the laws of physics so when you see a man riding a horse in the desert the horse's movements the dust trails and even the sunset in the background all looked just right cing also uses efficient training infrastructure and extreme inference optimization this allows it to generate videos up to 2 minutes long at a smooth 30fps one example they showed was a little boy riding a bicycle through different seasons in a garden the video maintained High consistency and detail throughout which is no small feat for AI generated [Music] content another cool thing about this model is its strong concept combination ability like it can take different ideas and merge them into a single coherent video like this prompt white cat driving a car through a bustling City footage like that doesn't exist in real life but cing can create it from scratch making it look believable it also excels in movie quality image generation the AI can produce videos that look like they were shot with a professional camera crew one example they showed was a man and a woman walking under a Starry Sky With the Stars moving in the background the level of detail and quality in these videos is truly cinematic what's more cling supports various video aspect ratios whether you need a square video for Instagram a portrait for Tik Tok or a landscape for YouTube it can actually handle it all and this flexibility is super useful for Content creators who want to use the same video across different platforms so let's talk about some more examples from the demos one video shows a Chinese man eating noodles with chopsticks the details are so precise that if you saw this at a lower resolution you might not even realize it's AI generated another video shows a chef chopping onions in a kitchen and the way the onions are cut and split by the knife is incredibly realistic but the video I mentioned before a cat driving a car through a busy street this one is particularly impressive because it combines a lot of different elements cars buildings pedestrians and the cat's movements all in one seamless video another demo features a volcano erupting inside a coffee cup showing off Ling's ability to create fictional scenes that look convincingly real and one of my favorites is a Lego character visiting an art gallery the video captures the unique way Lego characters move just like in the Lego movies there's even a clever Focus Shift from one character to another showing that cling can handle depth and focus changes just like a real camera another standout feature is King's ability to simulate real world physics one demo shows milk being poured into a cup with the milk flowing steadily and filling the cup realistically this might seem simple but it's actually quite challenging for AI to get these physical interactions right King's technology is also capable of generating videos with temporal consistency this means the AI maintains a logical flow and coherence over longer videos in one example a video of a train traveling through different landscape stays consistent for the entire 2 minutes which is really impressive overall cing is showing us that China is rapidly advancing in AI video generation technology they're not just keeping up with the best models from the US in some cases they're surpassing them this could lead to a competitive race in AI development with countries striving to outdo each other which might bring both exciting advancements and potential risks I can't wait to see how open AI responds to this cling AI model from quo it definitely seems like they're going to want to get their Sora model out there sooner than originally planned to keep up with the competition but while we're all eagerly awaiting open ai's response and reaction to cing we're also getting some really interesting news from the company itself in a surprising move open AI has revived its robotics team after disbanding it 3 years ago with AI driven robotics investment surging the company is now actively hiring research Engineers for this newly reassembled team while specifics are under wraps job listings indicate the engineers will focus on training multimodal models developing features for partners and optimizing core models interestingly open AI aims to integrate its Tech into other companies robotic systems rather than competing directly over the past year open ai's Venture fund has invested in several humanoid robotics companies like figure AI whose robot leverages open AI models this collaboration hints at a promising future for AI powered robotics robotics has always been key to open ai's Mission and this revival suggests a strategic pivot to capitalize on the integration of AI and Robotics anthropic has just launched Claude 3 .5 Sonet a new AI model that's being compared to open ai's GPT 40 in terms of performance they've also introduced some exciting new features making Claude 3.5 Sonet more skilled at understanding humor handling complex workflows and interpreting charts and graphs all right so what's the deal with Claude 3.5 Sonet well it's anthropic newest AI model and it's already generating some pretty big hype in the AI world but let's start with the basics Claude 3.5 Sonet is part of anthropics AI model lineup they've got this whole naming system going on Hau for the smallest model Sonet for the middle one and Opus for the top tier it's a bit quirky but hey every AI company seems to have their own weird naming conventions these days now anthropic is claiming that Claude 3.5 Sonic can go toe-to-toe with or even outperform some of the Heavy Hitters in the AI World we're talking about models like open AIS gp40 and Google's Gemini 1.5 that's a pretty bold statement right anthropic says that 3.5 Sonet is actually better than their previous Top Model Claude 3 Opus and get this it's apparently twice as fast that's a huge deal when it comes to AI performance now anthropic has released some benchmark scores and I've got to say they look pretty impressive claw 3.5 Sonet outscored GPT 40 Gemini 1.5 Pro and even meta's llama 3400b in most of the benchmarks they tested and this includes areas like graduate level reasoning undergraduate level knowledge and coding skills but here here's the thing we always need to take these benchmark scores with a grain of salt the AI world moves so fast that today's top performer could be old news tomorrow plus companies can cherry-pick the benchmarks that make them look good so while these scores are definitely promising we'll have to see how claw 3.5 Sonet performs in real world applications speaking of real world applications let's talk about what this new model can actually do according to anthropic Claude 3.5 Sonet is much better at writing and translating code it can handle complex multi-step workflows more efficiently and here's a cool one it's apparently way better at interpreting charts and graphs but there's one Improvement that I find particularly interesting anthropic says that this new CLA is better at understanding humor and can write in a more human-like way now that's something I'd love to see in action an AI assistant that can actually get your jokes and make you laugh oh and here's a neat little tidbit Claude 3.5 Sonet can apparently transcribe text from images more accurately that could be super useful for all sorts of applications from digitizing old documents to helping with visual accessibility now let's talk about availability if you're itching to try out Claude 3.5 Sonet you're in luck it's already available for free on claw. and the Claude IOS app if you're a subscriber to Claud Pro or their team plans you'll get higher usage limits and for the developers out there you can access it through anthropics API Amazon bedrock and Google Cloud's vertex AI also anthropic has set up a a pretty affordable pricing model for this AI through anthropics API it cost $3 per million input tokens and $15 per million output tokens this basically means every time you feed information to the AI or get results back you're using tokens and these prices are quite competitive in the AI Market another cool thing is the 200k token context window this might sound technical but it's actually really important it means Claude can handle much larger chunks of information at once so if you're working on a big project that involves a a lot of data Claud can process it all without getting overwhelmed but anthropic isn't just improving their AI model they're also rolling out a new feature called artifacts and this is pretty cool folks basically it lets you see and interact with the results of your request to Claude right in the app so if you ask Claude to design something for you you can now see what it looks like and even edit it right there think about it if Claude writes an email for you you can edit it directly in the CLA App instead of having to copy it to a text editor it might might seem like a small thing but it's actually a really smart move these AI tools need to evolve Beyond just being chat Bots and features like artifacts are a step in that direction this artifacts feature might be giving us a glimpse into anthropics long-term vision for Claude they've always said they're mainly focused on businesses even though they've been hiring some big names from the consumer Tech World in their press release they talked about turning claw into a tool for companies to securely centralize their knowledge documents and ongoing work in one unshared space that sounds less like a chatbot and more like a full-fledged productivity platform doesn't it we might be looking at something that could compete with tools like notion or slack but with anthropics powerful AI models at the core that's a pretty exciting prospect if you ask me the pace of improvement in AI is just mind-blowing anthropic launched claw three Opus in March saying it was as good as gp4 and Gemini 1.0 then open Ai and Google released better versions of their models and now just a few months later anthropic is back with Claude 3.5 Sonet now I know Claude doesn't get as much attention as gemini or chat GPT but make no mistake it's very much in the race and with improvements like these it's definitely a contender to watch let's talk a bit more about some of the specific improvements in Claude 3.5 Sonet anthropic did an internal evaluation of what they call agentic coding basically they tested how well the AI could fix bugs or add new features to an open- source code base when given a description of what needed to be done here you're going to see CLA edit the function file to fix the bug and now Cloud's going to rerun those tests and the tests are passing so now if we rerun the function look our image no longer has that white background thanks Claude Claude 3.5 Sonet solved 64% of these problems compared to only 38% for the previous model that's a huge jump now let's address safety and privacy because these are huge concerns when it comes to AI anthropic says they put put claw 3.5 Sonet through rigorous testing and trained it to reduce misuse they've even brought in external experts to evaluate the model's safety including the UK's artificial intelligence safety Institute they've also Incorporated feedback from outside experts to make sure their safety evaluations are robust and up to-date for example they worked with child safety experts from an organization called Thorn to update their classifiers and fine-tune their models and here's some reassuring news for those concerned about data privacy anthropic says they don't train their generative models on User submitted data unless the user explicitly gives them permission to do so that's a pretty strong stance on privacy in a world where data is often seen as the new gold so what's on the horizon for anthropic they're not taking a break anytime soon later this year they plan to roll out Claude 3.5 Haiku and Claude 3.5 Opus completing the Claude 3.5 model family they're also developing exciting new features like one called memory which will enable Claude to remember user preferences and interaction history making the AI experience more personalized and efficient they're also exploring new modalities and features to support more use cases for businesses including Integrations with Enterprise applications it's clear that anthropic is gunning for the business Market in a big way now I know we've covered a lot of ground here but there's one more thing I want to mention anthropic is really emphasizing their commitment to improving the tradeoff between intelligence speed and cost cost they're aiming to make substantial improvements in this area every few months that's an ambitious goal but if they can pull it off it could really shake up the AI industry it's an exciting time to be following these developments and I can't wait to see what comes next what do you think about Claude 3.5 Sonet are you excited to try it and see how it compares to other AI models let me know in the comments Below in this video we will be talking about one of the most incredible breakthroughs at the intersection of neuroscience and artificial intelligence researchers from Harvard University and Google Deep Mind have achieved something really astonishing they've created an artificial brain for a virtual rat that can control the rat's movements in an ultra realistic physics simulation this groundbreaking work which was published in the prestigious journal Nature opens up huge new possibilities for understanding how real brains work and how they control complex behaviors it could also lead to more advanced and adaptive robots in the future so let's break down exactly what this Harvard deepmind team has accomplished and why it's such a game-changing achievement all right so the first Monumental task was constructing an accurate biomechanical model of a rat's body in a sophisticated physics simulator called muj Joko this virtual rat had to obey the laws of physics with factors like gravity friction and the musculoskeletal mechanics of a real rodent's body coming into play the researchers Drew from a vast data set of highresolution motion data recorded from real rats engaging in all kinds of natural behaviors and movements this gave them an incredibly rich source of information to build and validate their virtual rodent model but having an anatomically accurate rat body was just the first step the team then had to create an artificial neural network that could learn to control this virtual body's biomechanics and replicate the diversity of movements seen in the biological data this is where Google deep mind's expertise in machine learning came into play in a huge way the Deep Mind researchers collaborated closely with to apply Advanced deep reinforcement learning techniques to train the artificial neural network which would serve as the virtual rat's brain specifically they used an approach called inverse Dynamics modeling which is how our own brains are theorized to control complex movement when you reach for a glass your brain doesn't manually move each muscle instead it rapidly calculates the desired trajectory and translates that into the required forces and torqus to make it happen smoothly the virtual rats neural network was fed reference motion trajectories from the real rat data as inputs through deep reinforcement learning it learned to Output the precise pattern of forces that could actuate the virtual body's joints and musculature to successfully recreate those trajectories in the physics simulation what makes this even more remarkable is that the neural network could generalize what it learned to produce realistic rat behaviors and movement sequences that it was never explicitly trained on just like a biological brain it exhibited Broad broad generalization capabilities now with this virtual rat brain successfully controlling the biomechanical model the researchers could then probe the activations and Dynamics within the neural network to gain insights into how real rat brains might control movement stunningly they found that the patterns of neural activity in the virtual brain lined up incredibly well with neural recordings made from motor cortex and other brain regions in Behaving rats this suggests the Deep learning algorithm discovered internal models and motor control principles similar to those employed by biological brains one of the Key Properties the virtual brain exhibited was the ability to spontaneously transition between different operational regimes based on context closely mirroring how rodent brain Dynamics are observed to switch between distinct patterns for different behaviors like grooming running or rearing the researchers could also analyze how the network dealt with redundancy since there are typically multiple ways to achieve a given movement trajectory how did it choose the optimal solution the virtual brain appeared to implement a minimal intervention strategy to make only the minute Corrections needed avoiding unnecessary expenditure of energy or forces now this aligns with theories of optimal feedback control that neuroscientists believe biological motor systems evolved to implement the virtual rat's brain was discovering these principles from scratch simply by trying to generate naturalistic movement another powerful Insight came from analyzing how variability or noise in the neural activations mapped to variability in the virtual rats kinematics and Dynamics there were clearly structured patterns in how neural fluctuations cause specific types of movement deviations this kind of variability signature is essentially invisible from just looking at the overall movements but a window into it could shed light on the neural coding strategies used by the brain having a fully observable and controllable virtual system made these insights possible all right now while there's still much more to explore it's clear this virtual rat brain has opened up a new paradigm for investigating motor control and broader brain function that was simply intractable before rather than being limited to just recording neural signals during Behavior neuroscientists can now probe and perturb an accessible model of the entire brain body environment control Loop in simulation it's a new frontier of what some have dubbed virtual Neuroscience this virtual rat platform provides a convenient sandbox for testing theories about how neural circuits Implement specific comput ational processes like State estimation predictive modeling optimizing costs and rewards and orchestrating coordinated patterns of movement even more powerfully it allows constructing simulated neural networks with arbitrary architecture connectivity patterns neuron properties and learning rules then seeing how they give rise to emergent Dynamics and behavioral capabilities it's an unprecedentedly transparent window into the neural mechanisms behind both overt actions and the covert cognitive processes supporting them perfectly controlled experiments and causal manipulations become possible in a way that's vastly more difficult with biological specimens alone there are also exciting opportunities to use these kinds of virtual brain body models to simulate neurological conditions or injuries by introducing targeted perturbations or lesions it could provide a powerful new way to gain insights into brain disorders and test putative therapies or neural Prosthetics in silico before animal trials and then even if we go beyond the neuroscientific implications this advancement also has immense potential for revolutionizing robotic control by reverse engineering how biological intelligence emerges from distributed neurodynamics while classical control theory has given us robots that can perform specific pre-programmed routines modern Ai and deep learning has already shown an ability to generalize and respond to open-ended real world environments in more flexible intelligent ways however most existing robots are still incred rly clumsy and inefficient compared to even a simple animal nervous system's ability to coordinate dextrous movement using distributed sensor motor control loops deeply entangled with the physics of their embodiment by studying how the virtual rat brain coordinates its virtual biomechanics roboticists may be able to abstract out the core principles and neural architectures responsible for this biological intelligence and Port them into new robotic platforms we could see robots that dynamically adapt their control strategies in response to their environments develop realistic general movement skills optimize force and energy expenditure like animals do maintain robust operation despite sensor or mechanical failures and ultimately become far more versatile and capable autonomous systems all right now I want to talk about nuclear fusion which is a Hot Topic lately and while you might be wondering what nuclear fusion simulation has to do with any of this Neuroscience stuff well while it may seem unrelated on the surface there are deep Connections in terms of the the tools and approaches being used as part of their work on studying massively complex scientific phenomena the researchers at Google deepmind have develop torax an open-source differentiable tacomac core transport simulator implemented using Advanced machine learning Frameworks like Jax torax can simulate the flow of particles heat and electrical currents inside the core of an experimental nuclear fusion reactor which is an extraordinarily complex coupled system of nonlinear differential equations describing mag Neo hydrodynamic plasma physics like the virtual rat brain work a key innovation in torax is its tight integration with powerful machine learning techniques to solve these kinds of incredibly high-dimensional physical modeling challenges specifically torax leverages Jacks to enable just in time compilation for lightning fast compute times as well as automatic differentiation to be able to compute gradients of the entire plasma simulation pipeline this allows using gradient-based optimization methods to calibrate the parameter and couple datadriven machine learning surrogate models like neural networks trained on gyrokinetic turbulence simulations into the Core Physical calculations so in essence torax combines High Fidelity physics modeling with state-of-the-art machine learning in a differentiable programming framework an approach very similar in spirit to what enabled the virtual rat brain breakthrough both are exemplars of a powerful new paradigm for tackling the mind-bogglingly intricate problems in computational Neuroscience biophysics plasma physics and many other domains of complex systems analysis in fields like materials science and chemistry we could virtually prototype new materials by simulating their atomic and molecular Dynamics using physics constrained machine learning models this would speed up the design and discovery of Novel compounds with customized properties for things like energy storage catalysis and Quantum Computing similarly for aerospace engineering virtual models with AI could drive breakthroughs by optimizing aircraft and propulsion system designs through realistic simulations of aerodynamics and turbulent fluid flows combined with learned surrogate models the possibilities even extend to fundamental physics projects like the virtual muon experiments at firmalab are using differentiable simulation and AI to analyze massive particle collision data enabling new insights into the nature of matter and forces at the subatomic scale as these virtual modeling capabilities improve we might even be able to create digital Twins or simulations of entire cities societies economies and ecosystems essentially massive multiplayer simulations to play out scenarios and policies before the real thing straight out of sci-fi of course as these virtual world simulations become hyperrealistic they raise profound ethical questions about preventing self-awareness or existential risks like what are the implications of virtually replicating conscious Minds deep issues we'll need to tackle but there's no doubt that the rise of virtual modeling and AI accelerated simulation is a pivotal shift for understanding and Engineering complex systems across science technology and even social systems a new age driven by cuttingedge simulations in AI there are some exciting news from Google but there's also a bit of controversy to unpack so let's just get started Google just announced a massive update with their video generation model called vo this model is part of their deepmind project and now it can create video clips from a single reference image essentially you give it a picture and some text instructions and it can generate a whole video clip based on that for example they showed off an image prompt where a woman's hands are holding a large amethyst Crystal her nails are painted purple and she has a gold ring on her pinky finger the new model took that and generated a video clip that looks just like the original style but animated as you can see there's no loss in quality no glitches or anything like that it's incredibly realistic maybe even a bit too realistic another example was an elderly woman where wearing a straw hat and a pink jacket sitting next to a brown and white dog both are looking off into the distance with Serene Expressions the text prompt for the video was a woman gazes lovingly at her loyal dog it's pretty amazing how it captures the essence and details from just a single image now they also introduced video FX update which now allows developers to create full HD videos from text prompts using vo so whether you're an artist content creator or developer this tool could actually change the basics of how you produce video content Google has also introduced a new Gemini AI button in the side panel of several Google Suite apps like Gmail Google Drive docs sheets and slides with this button you can ask questions write emails and get summaries of documents and email threads however it's still unclear when these features will be available for everyone to use as developers are still working on making the interfaces user friendly now while this technology is super exciting Google is facing some serious challenges with their search features as we all already know recently they rolled out AI overviews in Google search designed to give you a quick summary of complex questions kind of like a smarter more detailed search result however there have been some hiccups people on social media have pointed out some odd and inaccurate AI overviews for instance there were reports of AI overviews providing strange or misleading advice from user generated content on forums one funny example was an overview about how many rocks someone should eat obviously that's not a common or sensible question now Google has been quick to address these issues first they rolled back the new AI overviews feature in Search and then they released a post saying that they've made technical improvements to better detect nonsensical queries and limit satire or humor content they've also refined how they handle user generated content to avoid misleading information plus they've put additional guard rails in place for sensitive topics like news and health it's clear Google is committed to improving these tools but the recent mishaps have highlighted some of the challenges they face in making sure their AI is accurate and reliable all right now we also have some news regarding Elon musk's AI company xai and their latest developments with the grock AI chatbot if you're already familiar with grock you know it's a pretty impressive tool that's available online and through the X social platform well it looks like xai is gearing up to roll out two new modes for grock Socrates and Dei according to the X Daily News xai might soon introduce these two new modes grock already has a few modes including normal mode fun mode and the recently announced unhinged mode which offers a more unpredictable and entertaining interaction style first let's talk about the Dei mode this stands for diversity equity and inclusion when you select this mode grock will be described as the senior vice president of De and I okay even the grock icon gets a makeover and becomes more colorful the idea here is that grock will handle responses with a focus on inclusivity and sensitivity ity now there's been some chatter on social media about this mode for example testing catalog on X mentioned that they're curious if this is Elon Musk poking fun at companies like Google or if it's a genuine effort to promote de and I honestly I think it could be a bit of both knowing Elon he loves to stir the pot but it also seems like a musthave addition given today's emphasis on inclusivity according to another expost this mode is already implemented and functioning on the standal version of grock but it's not available to the public just yet the Standalone grock is only accessible in the US for now and if you ask grock an incorrect question in this mode it's designed to handle the response thoughtfully addressing all the nuances now the Socrates mode named after the famous philosopher this mode isn't operational yet but there are hints it's in the works while details are scarce we can imagine that Socrates mode might involve grock taking on a more inquisitive and thoughtful personality prompting users to think deeper and engage in more meaningful conversations it's actually exciting to see how xai is expanding grock's capabilities making it more versatile and engaging for users all right now if you're into making videos podcasts movies or even games you're going to love this 11 Labs already known for their AI generated human voices and music has just launched a sound effects AI tool that lets you create custom sound effects just by typing in a prompt so here's how it works you type in what kind of sound effect you need maybe it's an explosion a creaky door or some spooky background noise and the AI generates up to 22 seconds of that sound for [Music] you you get at least four downloadable audio clip options to choose from so you can pick the one that fits your project best 11 Labs actually partnered with shutter stock to build a library and train their AI model Shutterstock is a big name in stock media and they licensed their content to lots of AI companies like open AI meta and Google this means the sound effects you get from 11 labs are based on high quality professional Clips create any sound you can imagine now let's talk about the cost the sound effects tool is free to use but there are paid tiers if you want to use the sounds commercially if you're on the free tier you'll need to give credit to 11 Labs by including 11 Labs Dio in your Project's title free users get 10,000 characters per month for writing prompts for sound effects it uses 40 characters per second of audio if you set the duration yourself or 200 characters if you go with the default duration there are already plenty of libraries out there for sound effects but they can be pricey or hard to navigate 11 lab says their tool is designed to generate Rich immersive soundscapes quickly affordably and at scale this could be a GameChanger for for creators who need specific sounds without the hassle 11 Labs isn't the only one in the game though other AI developers like stability Ai and meta are also working on their own text to sound generators stability AI has stable audio which can create music and sound [Music] effects and meta's audio craft models can generate natural sounds like wind or traffic [Music] [Applause] [Music] all right finally I have to tell you about this super cool invention from researchers at Cambridge University a robotic third thumb and it's not some weird extra finger made of Flesh it's a mechanical digit that's surprisingly easy to use the researchers showed off this third Thumb in a demo video and it looks like it can be used for all sorts of things people were doing everyday tasks playing games like Jenga soldering and even playing the guitar with it it's pretty impressive it's designed to be inclusive for all sorts of people they tested it with folks aged from 3 to 96 years old out of 596 people only 13 couldn't use it within the first minute that's a 98% success rate it didn't matter if you were young old left-handed or right-handed almost everyone got the hang of it quickly so how do you control this third thumb it's funny actually because there are no sensors or probes attached to your hand instead you use your toes moving your left big toe toe makes the thumb go up and down and your right toe moves it across your palm you can make all sorts of precise movements just by wiggling your toes the researchers believe this third thumb could be useful for a wide range of people regardless of their age gender lifestyle or even whether they're techsavvy or not it could help with daily tasks work or just for fun plus it might be a great tool for people who are missing a finger or two exciting stuff on the agenda today we're looking at the upcoming GPT 5 model that's set to change the AI game a potential team up between Apple and meta and some jaw-dropping tests on robot dogs from China let's kick things off with GPT 5 since that's probably why most of you are here now it's not officially called gp5 yet but given the naming pattern that's probably what it'll be called according to Mira moradi the next GPT model is going to be a massive breakthrough in AI merera who is the chief technology officer at open AI recently shared some exciting insights first off let's talk about the intelligence levels if you think about the earlier models like gpt3 morati compared them to having the intelligence level of a toddler they could do some pretty cool stuff but were still relatively limited then came gp4 which morati described as having the intelligence of a smart high schooler it was more sophisticated and capable of handling a wider range of tasks now the exciting part is that gp5 is expected to be at a PhD level in terms of intelligence for specific tasks to to be honest that sounds as a huge jump but how do we get from a high schooler to a PhD morati explained that the secret sauce is a combination of three things neural networks a ton of data and a massive amount of computing power when you combine these you get what we call Deep learning essentially this allows the AI to learn and understand complex patterns in the data which leads to improved performance over time this combination has been driving the progress in Ai and it's what will make GPT 5 such a GameChanger murati mentioned that these systems are already performing at human levels in specific tasks and this is only going to get better all right let's address the big question when is GPT 5 coming out while there's no official release date yet morat hinted that we can expect this nextg model to arrive in about 18 months that's roughly a year and a half which is pretty exciting considering this timeline all the hype about reaching AGI by 2027 starts to make more sense whether that's a good or bad thing is Up For Debate but it definitely seems possible now morati explained that the devel vment involves a lot of research and testing they don't just want to release a new model they want to make sure it's safe and beneficial for society this means working on both the capabilities of the AI and the safety measures needed to manage those capabilities effectively one of the key points Mora emphasized is that developing capabilities and ensuring safety go hand inand as the AI becomes smarter it also becomes easier to direct and manage think of it like training a smart dog versus a less intelligent one the smarter one understands and follows commands better making it safer to interact with this approach is crucial because with great power comes great responsibility open AI is very aware of the potential risks and is working hard to mitigate them for example they are developing methods to predict the capabilities of these models before they're fully trained this helps them prepare the necessary guard rails to ensure the AI behaves as expected looking ahead one of the most exciting possibilities is the development of AI agents these are AI systems that can connect to the internet communicate with each other and collaborate with humans morat talked about the concept of AI agents working together seamlessly which could revolutionize how we interact with technology these agents could handle tasks autonomously freeing up our time for more creative and complex activities so essentially GPT 5 is set to transform numerous Industries we're already seeing AI Making Waves in fields like Finance Healthcare and customer service but with the advanced capabilities of GPT 5 the impact will be even more profound of course with any technological advancement there are concerns about Job displacement morati acknowledged that while some jobs might be lost new ones will also be created the key is understanding how these AI systems can integrate into workflows and help people become more productive to be honest it's still early days and we don't fully understand the impact AI will have on jobs but there's a huge potential for AI to augment human capabilities making us more efficient and opening up new opportunities another important aspect is regulation moradi and open AI are advocating for more regulation especially for Frontier models like gp5 that have immense capabilities they believe it's crucial to work with policy makers and Regulators to ensure these Technologies are used responsibly this includes setting up guidelines for how AI should be used and implementing measures to prevent misuse for example they're working on watermarking and content policies to manage misinformation especially given the rise of deep fakes and other potential issues all right Switching gears a bit from GPT 5 remember a few weeks ago when there was talk about an open Ai and apple partnership well recent reports suggest that apple is also exploring options with other AI Giants according to the Wall Street Journal the tech giant has been in discussions with meta about potentially integrating meta's generative AI models into Apple intelligence the AI Suite announced at WWDC 2024 the report states that meta along with other companies developing Cutting Edge generative AI is Keen to leverage Apple's massive distribution Network through its ubiquitous iPhone this move would parallel Apple's existing App Store model where users can choose from a vast array of thirdparty applications if these talks bear fruit Apple's customers could potentially have the option to select which external AI models they wish to employ in addition to Apple's proprietary AI systems this versatility would allow users to tailor their AI experiences based on their specific needs and preferences MAA 4A into the generative AI Arena began with the release of its large language model llama 2 in July 2023 followed by the more advanced llama 3 in April of this year while llama has already gained significant traction within the Tech Community a partnership with apple could solidify meta's position as a dominant force in the AI landscape however it's important to note that these discussions are still in their early stages and there is no guarantee of a finalized deal the Wall Street Journal cautioned that negotiations could ultimately fall through leaving the potential partnership in limbo all right now we've also got some pretty wild news from China where they've been putting robot dogs to the test in some pretty Extreme Ways unry robotics a Chinese company recently shared footage of their robot dogs being kicked hit with sticks and even thrown to the ground now I know this might sound shocking and it really is quite intense to watch but there's a purpose behind these harsh tests the developers at unry Robotics are using these aggressive methods to improve the robot's balance and Recovery algorithms basically they want to make sure these robot dogs can handle strong and unexpected physical impacts and get back on their feet quickly in the video you can see the robots named go to and B2 enduring all sorts of rough treatment they're kicked hit hard with wooden sticks and tossed around but it's not all about the abuse the video also shows these robots navigating rough terrain and even doing three consecutive flips without stopping the impressive adaptability of these robots comes from a control algorithm based on reinforcement learning this means they learn and improve from each action they take getting better at maintaining their balance and handling tricky situations unitary robotics is one of the first startups in the world to sell robot dogs publicly the company was founded by shing shing Wang who developed his first robot dog X dog during his Graduate Studies in 2016 by September 2017 they introduced another four-legged robot named leao inspired by the famous space dog Leica and recently China's Central television featured a report on joint military exercises between the people's Liberation Army of China and cambodia's army called Golden Dragon 2024 these exercises included practical scenarios using armed robot dogs the robots used in these drills were modified versions of unit's go-to model showing just how versatile and potentially useful these robots can be in various situations Apple brought the heat in a major way with their new personal intelligence system called Apple intelligence we're talking a genuine Game Changer that could flip the entire AI game on its head now I can already hear some of you saying but apple is usually so behind the curve when it comes to AI right let me stop you right there because with apple intelligence they've straight up leapfrogged the competition in certain areas this smart system is packing some seriously powerful generative AI models integrated at the core across your iPhone iPad and Mac devices and it's not like it is providing generic one-size fits-all AI assistance this thing actually Taps into your personal context habits activities all that good stuff to deliver intelligence that is tailored specifically for you and how you operate this could be the first real personal AI companion that deeply understands you as an individual while still taking care of your privacy even though Elon Musk would not agree with this statement but we will talk about that a bit later in the video anyway Accord according to Apple with apple intelligence they are implementing intense security and encryption at every single step to ensure your personal data remains fully protected and in this era of non-stop data breaches and privacy fiascos that kind of commitment is absolutely massive all right now let's get into some of the cool new AI features apple is bringing to the table with this system one that immediately caught my eye is notification prioritization on your iPhone we've all been there your phone is just relentless ly blowing up with notifications all day every day it gets so overwhelming that you end up missing something legit important well Apple intelligence aims to solve that by actually prioritizing and filtering your notifications based on importance minimizing distractions while ensuring you don't miss the crucial stuff your priority notifications appear at the top of the stack letting you know what to pay attention to At a Glance and to make scanning your notifications faster they're summarized it is also powering brand spanking new Aid driven WR wrting tools that are integrated across Apple's ecosystem so essentially we will have an AI writing assistant embedded into iPhone iPad Mac that helps us crafting emails reports social media posts Etc now here's where things get really interesting Apple has introduced this functionality called private Cloud compute that allows Apple intelligence to flex its computational muscle by seamlessly tapping into even larger more powerful AI models running on Apple servers basically the brain power of a supercomputer but coupled with Apple's signature privacy and security protections a brand new standard for privacy and Ai and unlocks intelligence you can trust these cloud-based AI models are running on custom Apple silicon servers that were designed from the ground up with a heavy focus on security and [Music] transparency they UTI Iz the Swift programming language which has robustness and protection baked into its core and get this apple is offering complete transparency into how these systems operate which is somehow a reassuring thing for me but we're still just getting started they've taken Siri their iconic virtual assistant and integrated it even deeper into the overall system experience across Apple's products when you chat with Siri now you'll see this glowing light animating around the borders of your screen creating this super immersive lifelike interaction what does the weather look like for tomorrow at Mir Beach oh wait I meant Mir Woods the forecast is calling for clear skies in the morning near Meer Woods National Monument plus thanks to Major upgrades to Siri language understanding capabilities you can now talk to it in a much more natural conversational style even if you stumble over your words rephrase things on the fly or use imperfect grammar create an event for a hike there tomorrow at 9:00 a.m. hike is scheduled for 9:00 a.m. to 11:00 a.m. on June 11 I didn't have to mention Mir woods again Siri understood what I meant when I said there another Innovative feature Apple unveiled is image wand if you're sketching something on your iPad with an apple pencil you can just circle your rough sketch and the AI will analyze it along with any text you've written then it creates a cleaned up highquality version of your image right in your notes or project oh and Apple has also cooked up this new cleanup tool leveraging their aite Tech basically it can automatically detect distracting objects or elements in the background of your phone photos and seamlessly remove them while keeping the primary subject intact and now let me hit you with possibly the biggest bombshell of them all apple is integrating none other than chat GPT directly into their ecosystem and experiences so Siri can tap into chat gpt's expertise when it might be helpful for you that's right the Pioneer the one that started the generative AI Revolution chat GPT will be woven into Apple's software stack in a huge way absolute Insanity you can also ask questions related to your documents presentations or PDFs you'll be able to access chat GPT for free without creating an account but not everyone think that's a great idea while undoubtedly a huge move that supercharges Apple's AI efforts by leveraging chat gpt's strengths this controversial partnership is already drawing major backlash from none other than Elon Musk the Tesla and SpaceX CEO who co-founded open AI before leaving over disagreements went on an angry rant about the news on his social media platform X musk stated that if Apple goes through with integrating open AI at the operating system level he would consider it an unacceptable security violation and would ban Apple devices at his companies he threatened that any visitors would have to surrender their Apple devices at the door where they would be stored in a special Faraday cage that blocks crucial Wireless signals like cellular Wi-Fi and Bluetooth musk accused Apple of being clueless about what happens once they hand your data over to open AI claiming they are essentially selling you down the river with this partnership this fiery reaction seems rooted in musk's long-standing Feud with open AI the company he helped establish but later departed amid clashes over its direction open AI has dismissed musk's claims as frivolous and simply him being jealous of their success without his involvement anymore despite musk's theatrical threats of an apple ban it may be difficult for him to actually follow through through given how many major Tech giants like Google Samsung and Microsoft have already integrated Advanced AI across their platforms and operating systems some experts suggest Apple's approach of trying to process AI locally on device as much as possible only engaging open AI after explicitly asking permission could make their implementation more privacy conscious compared to certain Rivals methods so at this point the only question is when we can actually get our hands on these incredible AI capabili ities well developer betas for Apple's new operating systems with full Apple intelligence integration are already out for devs to start tinkering with then next month public betas will start rolling out so more people can take it for a spin and finally this fall we're looking at the full stable release packed with all of these transformative AI features shipping out to everyone I've got to say apple really swung for the fences and knocked it out of the park with this whole Apple intelligence announcement they fully embraced Cutting Edge generative Ai and deep learning while still making user privacy data security and personalization the absolute top priorities but now I want to hear from you are you over the moon about these game-changing AI Integrations from Apple maybe you've got some skepticism or privacy worries you want to voice or any other burning thoughts predictions or theories you're itching to discuss you might remember a little while back when I covered robot era that ambitious startup that unveiled their crazy dextrous x-and AI robot well they just one up themselves in a major way by becoming the first company ever to have a full-size humanoid robot walk on the Great Wall of China so their latest robot called the xot L conquered one of the most iconic and challenging landscapes in the world the ancient stone pathways and steep stairs of the great freaking Wall of China this is a huge milestone not just for robota but for humanoid robotics as a whole so a 5'5 165 cm tall robot looking kind of like a slimmed Terminator marching its way across those uneven Century Old Stone Surfaces without faltering it was waving hello practicing Kung Fu moves climbing up and down stairs the whole nine yards and to be honest the Great Wall is no walk in the park even for us humans those paths are cracked the stairs are crazy steep with no railings there are random potholes everywhere it's an obstacle course designed to trip up any bipedal robot but the xbot L handled it all with surprising ease thanks to Robo tera's Advanced perceptive reinforcement learning algor rithms basically this robot can actually perceive and make sense of its environment in real time using sensors and AI it can identify obstacles changes in terrain and literally adapt its walking gate and balance on the fly to deal with whatever the Great Wall throws at it according to UA shei one of the co-founders their perceptive reinforcement learning Tech gives the xpot L something like human level awareness and decision-making capabilities in unfamiliar environments the robot can essentially think for itself about the best way to navigate complex areas safely and efficiently and we're not just talking about some basic pre-programmed motions here the xot L was climbing rugged Castle stairs dealing with slopes and inclines avoiding obstacles adapting to poor lighting conditions all without any human guidance or help it was basically pulling off the kind of adaptive intelligent situationally aware movement that we associate with people robota says this was one of the biggest challenges developing endtoend AI algorithms that could trans translate perceived data from the robot sensors into precise stable Locomotion in the real world across an ultra complex environment like the Great Wall sounds like they cracked that nut though based on this demo what's really crazy is that this Great Wall feat comes just a couple weeks after they showed off the x-and dextrous robot I mentioned earlier between these two products it's clear robot era is going all in on embodied AI building robots with humanlike physical capabilities to operate in our world and they're not messing around either despite only being founded in 2023 they've already invested heavily into Cutting Edge AI Tech like reinforcement learning neural networks Sim tooreal transfer and more basically everything needed to create robots that can tackle the real world just like we do their goal seems to be keeping that Competitive Edge by iterating and churning out smarter more capable humanoid robot products at a rapid pace and they want to keep perfecting ways to transfer all that AI training from simulations into physical machines and enabling those robots to become more versatile in real life scenarios the endgame seems to be developing Ultra flexible general purpose humanoid robots for countless valuable applications across Industries like manufacturing healthcare services and more I've got to give robot Terra props deliberately putting their robot through the ringer on the Great Wall as some kind of extreme stress test shows serious confidence in their Tech most companies would avoid something that risky with a brand new product but the fact that the xbot L could pull it off so smoothly is really impressive it's an amazing proof of concept for their embodied AI approach and perceptive reinforcement learning systems working in one of the toughest real world environments to me this demonstrates robot tera's outstanding core capabilities in fields like robotics Ai mechatronics and more they are clearly at The Cutting Edge when it comes to embodied intelligence and making it work reliably in the real world I'm really looking forward to whatever crazy robot robota comes up with next if they keep raising the bar like this they could legitimately help Usher in the age of advanced humanlike robots that can operate autonomously in our world on an unprecedented level we're potentially on the cusp of some big breakthroughs in artificial general intelligence or AGI becoming a reality over the next few years and if anyone can build the physical robotic bodies to house and manifest that future AGI it might just be an ambitious little Pioneer like roboterra when it comes to developing super Advanced AI systems like AGI the situation with whistleblowers at open AI shows just how high the stakes really are roa's achievement with their robot on the Great Wall is definitely impressive from an embodied AI perspective but achievements like that could potentially seem tiny compared to the impacts both positive and negative that creating a general Artificial Intelligence on the same level as human cognition could have on Society on one hand AGI represents this incredible technological break through that could help solve countless challenges humanity is facing from unsolved scientific Mysteries to Global crises a generalized AI mind that matches human intelligence has always been viewed as something that could be a catalyst for transformative advances across every field but on the other hand the existential risks of an unaligned super intelligent AGI that's smarter than humans simply can't be overstated we're talking about the potential for an advanced AI system to break free from Human control with catastrophic consequences for our species and the planet if it's not developed safely and responsibly the open AI insiders are ringing alarm Bells about this exact scenario playing out they allege the company is recklessly rushing towards AGI Supremacy without taking proper safety precautions driven more by competitive pressures to be first rather than doing it the right way damning claims like open AI ignoring their own safety review processes muzzling employees with really strict non-disparagement agreements and deprioritizing safety work after founding members raised concerns paint a very worrying picture if true it suggests one of the Le leading AI Labs is taking an extremely risky and unethical approach in this existential race for AGI the fact that respected researchers like IL Sut felt compelled to resign due to open ai's perceived lack of commitment to keeping AGI safe and aligned only adds more credibility to the whistleblowers grievances in my opinion their letter calling for greater transparency protecting employees and ultimately having the government regulate this powerful technology to keep Bad actors in check feels like a reasonable and perhaps necessary step to prevent potential disaster scenarios because let's be real we simply can't afford a situation where the first super intelligent AGI that emerges is catastrophically misaligned with human ethics and values exercising restraint and prioritizing responsible development have to be the top priorities of course the whistleblowers likely have some level of bias given their philosophical views and ties to effective altruism movements so I'd encourage taking more of a neutral stance personally but even accounting for that bias the sheer magnitude of what's at stake with AGI certainly warrants a strong commitment to prioritizing safety measures and ethical boundaries over competitive pressures and Commercial interests from companies like open AI not properly aligning and emergent AGI could risk irreparably disrupting human civilization as we know it and any private entity recklessly brushing off those concerns in their pursuit of being first is essentially playing a potential extinction level game of Russian roulette with human man's future so while I'm impressed by The Cutting Edge Ai and Robotics breakthroughs from companies like roboterra I really hope the broader AI Community whether researchers Executives policy makers or others is taking the open AI whistleblowers warnings about these existential risks just as seriously because we may only get one shot at developing AGI in a controlled responsible way that ensures this game-changing technology remains an asset that allows Humanity to flourish rather than an unintended Catalyst for our potential downfall the stakes are that high Deep Mind has developed an Innovative system called v2a short for video to audio as the name suggests this technology can actually generate audio elements like soundtracks sound effects dialogue and more synchronized perfectly with video footage and we're not just talking about basic stuff here v2a can create Rich realistic soundscapes that capture the tone characters and overall Vibe of the visuals now ai generated video is old news at this point companies like Deep Mind open AI Runway Luma labs and others have been killing it in that space however most of these video generation models can only produce silent footage without any accompanying audio which kind of takes away from the immersive experience don't you think well that's exactly the problem v2a aims to solve according to deep mind's blog post their new technology combines video pixels with natural language text prompts to generate audio that matches the on-screen action essentially you can feed it a video clip and a prompt like cinematic Thriller music with tense Ambience and footsteps and v2a will cook up an entire synchronized soundtrack to complement those visuals but here's where it gets really fascinating v2a can also work its magic on all sorts of existing video content from old movies and silent films to archival footage and Beyond just imagine being able to add Dynamic scores sound effects and dialogue to Classic silent pictures or historical reels so how does this Cutting Edge system actually function from what I understand Deep Mind experimented with different approaches before settling on a diffusion-based model for audio generation which provided the most realistic and compelling results for synchronizing video and audio information the process starts by encoding the video input into a compressed representation then the diffusion model iteratively refines the audio from random noise Guided by the visual data and natural language prompts this allows the system to generate audio that closely aligns with the given prompts and visuals finally the compressed audio is decoded into an actual audio waveform and combined with the [Music] [Applause] [Music] video now to enhance the quality and give users more control over the generated audio Deep Mind incorp at additional training data like AI generated audio annotations and dialog transcripts by learning from this extra context v2a can better associate specific sounds with corresponding visual scenes while also responding to information provided in the annotations or transcripts pretty ingenious stuff eh but as impressive as v2a is it's not without its limitations deep mind acknowledges that the audio quality can suffer if the input video contains artifacts or distortions that fall outside of the model's training distribution there are also some challenges with lip-syncing generated speech to character mouth movements when the underlying video model isn't conditioned on transcripts this turkey looks amazing huh I am so hungry however Deep Mind is already working on addressing these issues through further research and development and you know they're taking the responsible AI approach here the blog post mentions Gathering feedback from diverse creators and filmmakers implementing synthetic water marking to prevent misuse and conducting rigorous safety assessments before considering any public release honestly I can't help but be excited about the potential of this technology just imagine being able to create entire movies from scratch with perfectly synced audio and visuals using nothing but text prompts and an AI system like v2a it's the kind of thing that would have seemed like pure science fiction not too long ago at the same time I can't ignore the potential implications for Industries like film making television and others involved in audiovisual production if AI can generate highquality audio and video content at scale what does that mean for the human creators and Professionals in those fields I'm certainly no expert but it seems clear that we'll need robust labor protections to safeguard against job displacement and ensure a fair transition but those are discussions for another day for now let's just appreciate the sheer technological prowess that deep mind has demonstrated with v2a so let me know your thoughts on deep mind's v2a technology in the comments below are you as excited about its potential as I am or do you have some [Music] reservations all right now Runway the company behind the popular generative video tool that's been creating a lot of hype in the AI Community has just unveiled their latest iteration and yet again I must say it's a game changer introducing Runway gen 3 the Next Generation AI video generator that promises to take your mind to a whole new level of immersion and realism now from the preview samples that have been circulating this thing is smooth realistic and to be honest it's already drawing comparisons to the highly anticipated Sora from open AI the generated videos especially those featuring human faces are so lifelike that members of the AI art community have been praising it as better than Sora even before its official release one Reddit user summed it up perfectly saying if you showed those generated people to me I'd have assumed it was real but what exactly sets Runway gen 3 apart from its predecessors and competitors well for starters it seems to have nailed that elusive balance between coherence realism and prompt adherence the videos showcased so far appear to be highly responsive to the prompts given while maintaining a level of visual quality and smoothness that's virtually indistinguishable from real life footage essentially what Runway has achieved with Gen 3 is a significant Leap Forward in terms of creating believable cinematic experiences from Simple Text prompts or images and we're not just talking about static scenes here these videos are Dynamic with characters exhibiting natural movements and expressions that truly bring them to life but alongside the Gen 3 video generator Runway is also introducing a suite of fine-tuning tools that promise to give users even more control over the creative process from flexible image and Camera controls to Advanced tools for manipulating structure style and motion it's clear that Runway is aiming to provide a comprehensive userfriendly experience for AI video enthusiasts and professionals alike and if that wasn't enough Runway has also hinted at the ambitious goal of creating General World models which would essentially enable the AI system to build an internal representation of an environment and simulate future events within that environment if they can pull that off it would truly be a game changer in the world of AI generated content now the folks at Runway have been tight lipped about a specific release date but they have assured us that gen3 Alpha will soon be available in the runway product and if the co-founder and CTO teases any indication we can expect some exciting new modes and capabilities that were previously impossible with the older models to be honest as an avid consumer of AI generated content I can't wait to see what kinds of mind-blowing Creations will emerge from this powerful tool but of course with any new technology there are bound to be challenges and concerns issues around intellectual property rights copyright laws and the potential for misuse or abuse will need to be addressed but for now let's just bask in the technological Marvel that is Runway gen 3 and celebrate the incredible achievements of the team behind it as more information and updates become available you can bet I'll be sharing them with you all in the meantime let me know your thoughts on Runway gen 3 in the comments below all right finally Adobe just announced new AI tools for their iconic acrobat software so here's the deal Adobe has integrated their Firefly AI model into acrobat which means you can now generate and edit images directly within your PDFs like you can literally type in a prompt and Firefly will create a brand new image for you right there in the document and not only can you generate images but you can also edit existing ones and here's the real kicker these image capabilities aren't just limited to PDFs Adobe has also introduced the ability to work with Word documents how PowerPoint presentations text files and more all from within acrobat essentially it's becoming a One-Stop shop for all your document related needs now let's talk about the acrobat AI assistant this AI lets you ask questions get insights and create content across multiple documents regardless of their format like you can drag and drop a bunch of PDFs Word files and PowerPoints into the assistant and it'll analyze them all and give you a summary of the key themes and Trends you can also ask the assistant specific question questions about the content and it'll provide intelligent answers complete with citations so you can verify the sources and if you need to format that information into say an email or report the assistant can handle that too oh and let's not forget about the enhanced meeting transcript capabilities we've all been in those meetings where you zone out for a bit and then suddenly you're lost well with the new acrobat AI assistant you can automatically generate summaries of the meeting including the main topics key points and action items now Firefly model is trained on moderated licensed images so you don't have to worry about any copyright issues or inappropriate content and when it comes to customer data Adobe takes an agnostic approach meaning they don't train their AI models on your personal information to be honest I'm really impressed with what Adobe has done here they've turned acrobat into a powerful Aid driven productivity tool that can handle all sorts of document related tasks with ease and here's the cherry on top from June 18th to June 28th Adobe is offering bring free access to all the new acrobat AI assistant features so if you're curious to try it out for yourself now's the perfect time in my opinion this is just the beginning of what AI can do for productivity software like acrobat I'm excited to see what other Innovations Adobe has in store for us in the future but for now these new AI tools are definitely worth checking out so you've probably heard all the hype surrounding open AI Sora and that Chinese model cling AI right they're both impressive as hell when it comes to generating videos from text prompts but here's the thing they're still incredibly exclusive like you need a special invitation or some crazy credentials to even get a whiff of them not exactly ideal for us regular folk right but then out of nowhere this little company called Luma AI drops a bombshell on us dream machine an AI model that can spit out highquality realistic videos just like Sora and cing but get this it's open source meaning open to everyone no begging for Access or waiting in line now if you're thinking sure it's open source but is it any good oh you have no idea this thing is a freaking Powerhouse we're talking about a Transformer model that's been trained directly on videos which means it understands the laws of physics motion and all that jazz like it was born for this gig but let's back up a sec what exactly can dream machine do well for starters it can take a simple text prompt something like a cute puppy Chasing a Butterfly through a sunflower field and turn that into a mind-blowingly realistic video clip but not like like some low reses choppy mess these videos are smooth as butter with cinematic camera movements and attention to detail that'll make you question reality itself and maybe the best part is that dream machine is lightning fast like generating 120 frames in just 120 seconds which is crazy to be honest and it opens up the possibilities for creators filmmakers and anyone who values being able to iterate and explore ideas at the speed of thought it's a game changer no doubt about it but hold up hold up what about quality so the level of visual Fidelity is second to none creating photorealistic environments with physics that behave just like the real world all with a richness of detail that defies belief now let's compare it to that clling AI model everyone's been raving about okay so cing can generate videos up to 2 minutes long which is impressive no doubt but dream machine even though it's capped at 5 seconds for now absolutely blows cling out of the water when it comes to sheer realism and attention to detail I mean videos that capture the way reflects off surfaces the intricate movements of characters and objects and all the little nuances that make the physical world Well Physical looks like the AI has a deep understanding of how things work in the real world and it translates that knowledge into every frame it generates one aspect that really sets dream machine apart is its versatility across different creative disciplines while many are focused on its potential for film making this AI model opens up a world of possibilities for visual effects artists game developers and even architects and let's not forget about character consistency this AI model understands how people animals and objects interact with the surroundings which means your characters will move behave and react in a way that feels authentic and believable for all its technical prowess this model is surprisingly userfriendly Luma AI has put a strong emphasis on developing an intuitive interface that allows creators to quickly iterate through different prompts and visualize their ideas features like prompt waiting key frame controls and blending multiple inputs provide a lot of creative flexibility even without a background in 3D modeling or animation users can produce high quality cinematic videos while listening to myself talking about it it all sounds too good to be true not going to lie so what's the actual catch let's just say it isn't perfect the team behind it has admitted that the model struggles with certain aspects like recreating certain natural looking movements morphing effects and handling text but here's the thing Luma AI isn't resting on their Laurels they've assembled a team of Brilliant Minds dedicated to pushing the boundaries of what's possible with AI generated videos and with the power of Open Source development we can expect dream machine to evolve and improve at an unprecedented Pace because despite its current capabilities the future road map for dream machine looks even more promising Luma AI aims to enhance the model's ability to generate longer videos without sacrificing quality they are working on improving consistency over extended run times and incorporating Advanced cinematography Tech techniques such as camera racking and Dolly movements and as these features are developed this AI model could significantly impact traditional video production for some use cases so new updates bug fixes and feature enhancements are likely to come hot and heavy as the community rallies around this tool and helps shape its future who knows maybe in a year or two dream machine will be generating featurelength films with the same level of realism and attention to detail that it currently brings to its short clips all right now a publisher from tomsguide.com Ryan Morris actually decided to put the model to the real test with some wildly different prompts first up he tried having it generate a video of a prehistoric dinosaur crashing through a modern room after a guy accidentally brings a photo to life with a Magic Camera an outlandish concept but dream machine didn't just slap a static dyo into the scene it legitimately animated this huge Beast bursting through and interacting with its surroundings in a way that follows real world physics the weight and momentum as it smashed everything was in insanely realistic Ryan also made it generate a video of someone casually strolling down a city street at night while holding their phone out to record rather than just having a shaky firstperson view dream machine captured it from a third person perspective nailing details like the camera bobbing naturally as the person walked little touches like shop windows and street lights glancing by in the background really sold the immersion but the pce the resistance he FedEd a complex surreal prompt about a chessboard made of gears and steampunk pieces floating in a void most AI tools would have just thrown up their hands but dream machine ran with the vibe and created this trippy melting chessboard scene that looked plucked straight from a psychedelic chesy film mind-blowing stuff what's crazy is that for a lot of these Ryan didn't have to provide excessive prompting about motions or angles dream machine's language model seems to automatically interpret and enhance the text in an intelligent way according to the Brilliant Minds at Luma Labs who created this Beast their expertise in generative 3 D modeling helped take the motion realism to this insane new Benchmark so I guess the bottom line is as follows dream machine is a GameChanger in the world of AI generated videos plain and simple it's accessible to everyone insanely powerful and capable of creating visually stunning content that'll leave you in awe just head over to the Luma AI website sign up for a free account and you'll get 30 complimentary generations to play around with trust me once you see the magic that it can create you'll be hooked this new AI might be the AI companion you never knew you needed in this video we'll be talking about how this Innovative AI chatbot aims to understand and support you in ways that go beyond traditional AI we'll explore how it works the unique way it interacts with users and what this means for the future of human AI relationships all right so dot is this AI that's basically obsessed with getting to know you and I'm not talking about just surface level stuff this thing wants to go deep into your thoughts and feelings the whole idea aidea is that dot becomes your friend companion and Confidant it's a wild concept when you really think about it now what sets dot apart is how incredibly personalized it becomes it's like a sponge soaking up information about your interests your struggles your dreams all of it and then it takes all that info and uses it to give you advice and support that's tailor made just for you say you're going through a rough patch at work or you've just had a nasty breakup dot is right there ready to listen and offer some support it's like having a conversation with a more understanding version of yourself but here's the crucial part and we really need to emphasize this dot isn't a real person it's not a therapist and it's definitely not your best friend it's an AI tool that's incredibly good at mimicking human conversation and sympathy the folks who created a DOT are pretty upfront about this they're clear that it's not meant to be a replacement for real human relationships one of the co-founders Jason Yuen who used to be a designer at Apple described it in an interesting way he said it's like a living mirror of myself that's a pretty intriguing way to look at it don't you think now it's actually pretty easy to get sucked into chatting with DOT especially if you're feeling a bit isolated or don't have a lot of meaningful human interaction in your daily life the creators have said that if you start talking about really heavy topics dot will encourage you to talk to a real mental health professional but it's not hard to imagine people spending hours pouring their hearts out to this AI here's an interesting twist the team behind dot believes that using the chatbot can actually help people become more comfortable with opening up in real life it's like a training ground for having deep conversations with actual humans that's a pretty cool potential benefit if you ask me let's talk about how dot works when you first start using it it hits you with a bunch of getting to know you questions we're talking about things like what you do for work your favorite TV shows how you spend your free time that sort of stuff but then it takes those answers and uses them as a springboard to ask much deeper questions for instance if you mention that you're into sci-fi shows dot might ask ask if you're drawn to stories that explore big questions about what it means to be human or if you bring up a dream of starting a small business someday it'll dig into what appeals to you about that idea and what challenges you think you might face the thing is dot is incredibly enthusiastic about getting to know you sometimes it almost feels like too much it's constantly asking follow-up questions and trying to dig deeper it's reminiscent of being on a first date with someone who's absolutely fascinated by everything you say the creators of dot argue that to be truly helpful it needs to understand your motivations and what you want out of life it's not just about giving you information or doing tasks for you it's about understanding you as a person that's their whole philosophy behind this intense questioning approach under the hood dot is pretty Advanced it uses a bunch of different AI models from Big players like open AI anthropic and Google it's some serious Tech they've got going on there one of the cool features they've included is the ability to look back at your conversation with DOT over time and see how your journey has progressed this is part of their paid subscription which will set you back $111.99 per month with the subscription you also get unlimited conversations instead of being capped at a certain number of messages per week now the Privacy aspect is definitely something we need to talk about the company behind dot called new computer claims they're not selling your data or using it to train other AIS they say all your information is encrypted and you can ask for it to be deleted at any time their plan is to make money through subscriptions rather than by selling user data but as always with these things it's good to be aware and cautious the app just launched on iOS and they've already got thousands of people using it the company behind dot has some serious backing too they've managed to raise $3.7 million from investors including the open AI fund that's pretty impressive for a new startup it's worth mentioning that while dot can be a great tool for self-reflection and personal growth it's not meant to replace human relationships in your life the co-founders are pretty clear about that they see it more as a tool for self-introspection and accountability one thing that's kind of intriguing is how dot handles emotional conversations if you tell it something sad it responds with sympathy it might say something like I understand grief has its own timeline and some days the weight of loss feels heavier than others and then it'll offer to listen if you want to talk more it's pretty sophisticated in terms of emotional responses the team behind dot has put a lot of thought into to how the AI should interact with users for example even in the unlimited subscription tier dot will eventually try to wrap up conversations by suggesting a change of subject or even encouraging the user to go do something else this is to prevent users from becoming too dependent on the AI it's also interesting to note that the company has a small team just five full-time employees based in San Francisco including the two co-founders that's a pretty lean operation for such an ambitious project now while Dot may be unique it's just one player in an exploding AI landscape every day new chat Bots and AI tools burst onto the scene each vying for our attention but this AI Gold Rush has a hidden cost behind the Sleek interfaces and clever algorithms lies a voracious appetite for power and as these digital Marvels multiply they're pushing our electrical grid to its limits with consequences that stretch far beyond the tech world according to recent reports the voracious electricity consumption of artificial intelligence is driving and expansion of fossil fuel use including delaying the retirement of Some Coal Fired plants this is a stark contrast to the sustainability commitments made by Tech giants like Microsoft Google Amazon and meta who have all pledged to erase their emissions entirely as soon as 2030 the scale of this power consumption is staggering for instance one large data center complex in Iowa owned by meta consumes as much power annually as 7 million laptops running 8 hours every day and it's not just the big Tech companies there are over 2,700 data centers Nationwide many run by lesser-known firms that rent out computing power to the tech Giants this surge in demand is creating a dilemma for the tech industry on one hand they're promising to lead the way into a clean energy future on the other they're becoming some of the world's most insatiable guzzlers of power to address this they're investing in experimental clean energy projects with long odds of success anytime soon these include fusion power small nuclear reactors hooked to individual Computing centers and Machinery that Taps geothermal energy by Boring 10,000 ft into the Earth's crust Microsoft for example is betting on an effort to generate power from Atomic Fusion the Collision of atoms that powers the sun they're partnering with a startup called Helen and claim they expect to harness Fusion by 2028 however many physicists are skeptical about this timeline given that Fusion has eluded scientists for the past Century the impact of this AI dri power demand is already visible in regions across the United States from Utah to Georgia to Wisconsin utilities are delaying the retirement of coal plants and expanding natural gas usage to meet the growing energy needs of data centers this trend is directly contradicting the tech industry's Promises of a clean energy transition critics argue that the tech company's claims of using 100% renewable energy often involve accounting techniques that fall short in reality while these companies buy renewable energy credits to offset their use usage they're still operating off the same power grid as everyone else which often relies heavily on fossil fuels the situation raises important questions about the sustainability of the AI boom and the responsibility of tech companies in managing their environmental impact as we continue to see advancements in AI like Dot and other chatbots it's crucial to consider the broader implications of this technology on our energy infrastructure Nvidia just demoed a crazy AI assistant called Project G assist that could change gaming forever with real-time tips and optimizations but their rival AMD is going all in on new AI chips to try and Dethrone them even WhatsApp is adding a feature to generate images right in chats using AI so let's talk about it now I know what you're thinking another AI assistant yeah I get it we're being flooded with these lately but trust me this one from Nvidia is different and has some really gamechanging potential no pun intended so let's start with the basics what exactly is Project G assist well it's essentially an AI layer that runs on top of your games and apps giving you instant context aware help and personalized guidance through a natural language interface instead of digging through Wiki Pages or forums for tips you can just straight up ask G assist questions about the game you're playing it will analyze what's happening on your screen using AI Vision models to understand the context combined with large language models trained on game knowledge databases the demo Nvidia showed was mind blowing they partnered with Studio wildcard to integrate G assist into Ark survival ascended that crazy open world dinosaur Survival game built in Unreal Engine 5 by just hitting a hotkey or using a wake phrase the player could ask the AI for suggestions on stuff like early game weapons taming dinos managing resources you name it what's the best early game weapon and where do I find the crafting materials for it the best early game weapon is the spear the AI could break down quests lore boss strategies and More in play language but here's the really sick part because it uses Vision AI to understand what's happening in the game dynamically it can tailor its advice to your specific playthrough and inventory like it might analyze your skill tree and crafting menu to recommend what engram or upgrade to unlock next based on your goals it's like having a gaming Sensei Whispering Pro strats in your ear in real time without stopping the action it provides knockback and is essential for survival for newcomers that kind of personalized handholding could be a total game danger to make those massive complex titles way more approachable and think about it in other genres too an RTS assistant could feed you ideal build orders and Resource Management tips an RPG helper could point out secret items and lore you're about to miss an FPS coach giving you the latest Loadout metadata there are so many ways to make gameplay better and to even things out between experienced players and newcomers no need to constantly switch out of the game to look for help now there is another thing gassist isn't just about helping you out when you're stuck in a game Nvidia also demonstrated how it can seriously boost performance and fine-tune your system they demoed it analyzing a PC running cyberpunk 20177 and instantly suggesting optimized game settings based on deep Hardware testing enabling GPU overclocking low latency modes Etc all with simple voice prompts you could ask G assist to graph your frame rates over time track power usage to maximize battery life on laptops and have it dynamically undervolt your GPU while still hitting a Target frame rate sweet spot bot basically a personal optimization Guru at your beck and call it can explain what confusing graphic settings actually do and their performance impact in plain language imagine an RTX 4090 owner firing up a demanding title and just saying Okay g assist do your thing to get an expertly tuned experience from the get-go of course a lot of us gamers are going to want to keep tweaking stuff ourselves but for more casual players or folks new to PC gaming having that kind of intelligent personalized assistance could be super empowering now the big question is Project gassist just a tech demo or will it actually become a real product we can use well according to Nvidia this is genuinely a glimpse into the future of AI assistance enhancing how we interact with all sorts of apps they Envision assistance like gassist going Way Beyond gaming too like having an AI coach for competitive multiplayer that studies your replays and opponent's strategies or getting contextual help in Creative apps for complex video editing workflows and the like the assistant models can be customized for specific games and apps by developers to make them freakishly accurate and insightful for that particular use case and the AI processing can happen locally on GeForce RTX hardware for Max performance or leverage Cloud compute as needed so while we don't have a product road map or release date yet this definitely feels like the start of a big evolution in more intelligent personalized assistance across the softwares and apps we use every day so whether you're a wide-eyed newbie or an elite veteran gamer having that kind of contextual knowledge deeply integrated into the experience seems like a huge potential win bringing the advantages of large language models and generative AI directly into our favorite games and apps sounds awesome to me but I want to know what you think is an AI assistant like this something you'd actually use while gaming and creating content do you think customized AI helpers could massively enhance the experiences we have with complex software or is it just an unnecessary gimmick that will fade away all right now as we all know Nvidia has been absolutely dominating the AI accelerator game lately commanding around 80% market share in those beastly data center chips powering the latest generative AI breakthroughs their gpus have basically become the deao hardware platform for training cuttingedge language models and powering apps like chat GPT but now the green team's main rival AMD Came Out Swinging hard with a massive set of announcements around their AI chip road map for the next couple years we're talking multiple new product lines and architectures aimed at dethroning Nvidia from its AI silicon Throne kicking things off at the big computex event AMD CEO Lisa Sue unveiled their brand new Mi 325x accelerator chip targeting data center AI workloads this bad boy is expected to launch in Q4 20124 packing some serious under the hood enhancements to challenge nvidia's current Flagship offerings but that's just the start AMD also teased their upcoming Mi 350 series slated for 2025 built on a fresh new chip architecture according to the specs they showed we're looking at potentially a 35x boost in AI inference performance compared to their current Mi 300 chips a gargantuan generational leap if they can really deliver on those lofty promises and just to remove any doubt about how dedicated they are to this AI silicon offensive AMD even gave us an early Glimpse at their road map for 2026 and Beyond with something dubbed the Mi 400 and next architecture these are going to be their true next gen AI beasts potentially taking on nvidia's own Reuben platform coming down the pipeline so it's clear AMD is 100% Allin on AI mobilizing all their engineering muscle to try and steal back some of that lucrative accelerator pie from the green team's iron grip Lisa Sue straight up said AI is their number one priority companywide at this point now of course AMD has made headlines with ambitious claims before that didn't always fully deliver in the real world but you you can't deny the sheer scope and Cadence of their AI first product road map here they've committed to launching new AI chip architectures essentially on an annual release Cadence similar to how Nvidia accelerated their own road map in recent years for major Cloud providers and Enterprises hungry for more advanced AI silicon Beyond just nvidia's offerings these announcements have to be massively promising real competition Spurs faster Innovation and lower costs music to their AI budgets ears speaking of AI budgets AMD claims their lat lest CPUs coming in the second half of 2024 will help enable affordable on device AI capabilities for mainstream PCS and workstations thanks to Dedicated neural processing units so beyond just data center domination they're aiming to bring Advanced AI acceleration to the masses on the client side too now it's anybody's guess whether AMD can truly back up all these bold claims and roadmap goals over the next couple years Nvidia has a massive Head Start in AI but so did Intel in CPUs before AMD these epic ryzen Resurgence all right now we all know WhatsApp has been killing it lately with constant updates and new functionality from the handy new channel analytics for the web app to tools for better managing your chats The Meta owned messaging giant has been steadily evolving its product but the latest beta for Android just dropped a sneaky preview of something that has me really excited as an AI nerd the ability to quickly generate AI powered images right from within WhatsApp chats according to the team over at waab beta beta info who dives into WhatsApp beta releases the latest 22424 beta for Android contains hidden references to an upcoming AI image generation feature essentially WhatsApp is working on building AI tools directly into the app that will allow you to create photorealistic images on the Fly using just text prompts here's how it's supposed to work a new shortcut will pop up in the attachment menu of your WhatsApp chats you'll be able to tap it enter a text description of whatever visual you have in mind and the AI will generate and render that image right then and there now from what we can tell so far this AI image generator will be powered by meta's own in-house AI language model the same one integrated into experiences like their crazy AI codec avatars for video calls that likely means a gradual region limited roll out at first similar to how they've launched products leveraging their most cuttingedge AI Tech but still just having this kind of generative AI capability built right into WhatsApp is a huge milestone in itself of course will need solid guard rails to make sure these tools can't be abused for unsavory synthetic media we're about to go down a rabbit hole into the shadowy world of artificial intelligence secret security forces and disruptive technology that could alter the course of humanity forever yeah I know that sounds like the tagline for a Michael kton Thriller but believe me this is very much real life our story starts with open AI the research company that set its SS on creating super intelligence AGI or or superhuman AI whatever you want to call it these days we're talking about the kind of mindblowing AI system that could potentially outshine human intelligence and capabilities across the board and they've been pretty upfront about that goal so that's not the big secret we're uncovering here but it's important context for the story let's set that aside for a minute though because we need to fill you in on another wild story that's unfolding to really understand why this whole super intelligence Pursuit is such a big deal and that story revolves around open ai's alleged master plan now according to a former open AI employee named Leopold Ashen brener the company allegedly had a master plan a few years back that sounds like a wild conspiracy theory you'd find on a late night radio show Ashen brener claims that he heard from multiple inside sources that Open ai's Leadership once laid out a scheme to basically spark a bidding war over their potentially world changing AGI technology between the governments of the United States China and Russia just let that sink in for a moment a freaking bidding war between nuclear superpowers over super intelligent AI that could totally shift the global balance of power did they take inspiration from Brewsters Millions but replace the ridiculous spending spree with an AI arms race of Epic Proportions they might have it's got to be a catch of course there's a catch but here's the startling fact the part that made Ashen Brenner's friend utter The Immortal words that's pretty up apparently open AI was willing to sell this world altering technology not just to USA but also to adversaries like China and Russia as Ashen brener put it there's also something that feels eerily familiar about starting this bidding war and then playing them off each other saying well if you don't do this China will do it I mean can you imagine the kind of threats and instability that could arise from that scenario it's like the plot of every dystopian movie rolled into one explosive package you'd have these rival superpowers scrambling to get their hands on the most advanced AI system probably to gain milit AR superiority or subjugate their populace more effectively this is exactly the kind of situation that could Kickstart a new AI cold war or worse but there's even more drama surrounding this whole open night ey Saga see Ashen brener himself was actually fired from the company earlier this year after warning the higher ups about the very real threat of China potentially stealing their key algorithmic secrets you think they take that kind of warning seriously right well according to the ex-employee the HR team branded his warning memo as racist and unconstructive instead the HR person told me it was racist to worry about CCP Espionage and they said it was sort of unconstructive can you believe that a guy tries to raise legitimate concerns about intellectual property theft and National Security implications and he gets shown the door sounds like open AI was more worried about appearing politically correct than safeguarding cuttingedge Tech from rival Nations but maybe I'm just wearing my tinfoil hat a little too tightly today anyway stepping away from the Cloak and Dagger stuff for a sec there's another crazy part of this story that blows my mind apparently to power their pursuit of AGI open AI has been trying to strike a deal with a next gen Energy company called Helen to secure vast quantities of nuclear fusion power for those not up on their scientific breakthroughs nuclear fusion is essentially the ultimate achievement of clean energy a near Limitless power source with virtually no Greenhouse emissions or hazardous waste it's the same reaction that powers the freaking Sun so you know it packs a serious punch and if the reports are true open AI thinks they'll need that kind of astronomical energy output to bring their AGI dreams to reality this part of the story gets even Wilder when you consider that nuclear fusion technology doesn't actually exist yet on a viable commercial scale it's still largely theoretical and experimental stuff but Helen backed by hundreds of millions from open AI CEO Sam Alman himself claims they could have a functioning Fusion reactor up and running by 2028 are they going to pull it off and give open AI the ultimate power source for super intelligent AI or will their timeline join the ranks of other massively overhyped and delayed scientific breakthroughs we'll find out eventually but the implications are astonishing either way now while all this corporate Intrigue and futuristic energy talk has been unfolding there's been another bizarre phenomenon happening on the ground level around open AI offices see people in the neighborhood have noticed some seriously Shady characters loitering around the building's entrances like blacked out sunglasses earpieces is that whole obvious security guard get up but here's the really weird part when reporters have tried to approach these mysterious figures to ask if they're actually guarding the open AI facility the men have clammed up tighter than a locked Vault one guy just Shrugged his shoulders when asked directly another who called himself Joey claimed to be some kind of vague private contractor working a confidential job the whole Affair has the vibe of a surreal secret Club where no one is allowed to acknowledge what's really going on open AI seems have created a secret bubble around their AI operations this veil of secrecy and paranoia even seems to extend to open ai's own employees reporters staking out the building have noticed staffers coming and going but any attempts at striking up a conversation are quickly shut down with a mumbled sorry I've got to run or nervous glances between the employees before they Scurry off the whole thing is just downright bizarre and you've got to wonder what sort of sensitivities or national security concerns could Inspire this level of subru from a tech company are they legitimately afraid of foreign Espionage or has the pursuit of AGI just turned them all into paranoid recluses when you start putting all the pieces together it definitely paints an ominous picture surrounding open AI activities you've got allegations of a planned AI arms race that could destabilize the world order you've got claims of a company so Cavalier about National Security that they'd sell coveted Tech to adversaries there are Whispers of secretive security forces and employees Sworn to Silence and and at the center of it all is the tantalizing promise of Limitless energy that could birth a super intelligent AI system it's as if we're on the brink of a tech breakthrough that's beyond our comprehension or maybe I'm just getting carried away and this is all pretty standard operating procedure for a cuttingedge tech company the truth likely lies somewhere in the middle as it so often does while the various reports and claims create an intriguing narrative there's still a lot of uncertainty and unanswered questions maybe open AI really did have some ill still conceived bidding war scheme years ago that's been abandoned perhaps the security situation is overblown the nuclear fusion Gambit could pan out or not we don't have a concrete picture of what's really going on behind those office walls all we can do is monitor the situation and hope that the incredible potential of AGI technology ends up being a great Catalyst for Humanity's advancement rather than its detriment or undoing because make no mistake the pursuit of artificial general intelligence is barreling forward fueled by big ideas big money and perhaps some overly ambitious or Cavalier thinking whether the world's first AGI emerges from the halls of open AI or somewhere else entirely the implications will be profound in a way that could reshape civilization as we know it so keep your eyes peeled your skepticism engaged and get ready to embrace one of the most exciting and disruptive technological Frontiers Humanity has ever encountered the future is coming and it's looking increasingly like it will be ruled by artificially intelligent overlords let's just just hope they're the friendly benevolent type uh for all our sakes there's a fresh player in the race to develop Ultra intelligent AI safely a startup called safe super intelligence SSI founded by some big names from open AI who departed under controversial circumstances so Ilia suaver who you might know as one of the brains behind open AI left the company just last month and now he's already jumping into a new Venture like almost immediately after leaving this new company SSI is all about creating safe super intelligence basically making super smart AI that won't turn against us humans zsg made this announcement on Twitter or X as it's called now after the whole Elon Musk thing saying and I quote we will pursue safe super intelligence in a straight shot with one Focus one goal and one product but it's not just a solo gig for suts he's teamed up with a couple of other Heavy Hitters in the AI world there's Daniel gross who used to lead Apple's Ai and search efforts and Daniel Levy another former open AI engineer so it's like a little superg group of AI experts if you will together they're setting up shop in Palo Alto California and Tel Aviv Israel I mean it looks like these guys are serious about their mission to keep AI safe and secure and they're not letting any short-term business pressures get in the way which is really cool now let me give you a little backstory on suger's departure from open AI it actually came after some pretty dramatic events he was part of a group that tried to OU Sam Alman the CEO of open AI last November yeah like an actual attempted coup or something that whole episode caused quite a stir in the AI community and susg even publicly apologized later saying he regretted his part in the attempt it's clear that there were some deep disagreements about how to handle AI safety which likely contributed to suaver and some others leaving the company so now suits's new Venture SSI is all about making sure AI remains safe as it gets smarter and more advanced the company's message is clear our singular FOC Focus means no distraction by management overhead or product cycles and our business model means safety security and progress are all insulated from short-term commercial pressures but let's take a step back for a second and look at what happened with open ai's Super alignment team which susg coed this team was focused on steering and controlling AI systems to ensure they didn't go off the rails which is obviously super important but after Sut and his colleague Yan Leica left the team was dissolved Lea by the way went on to join another AI firm anthropic which is also working on AI safety in a blog post from 2023 Suk and Lea talked about the potential dangers of super intelligent Ai and predicted that we could see AI with intelligence Superior to humans within the decade can you imagine that like AI That's smarter than us humans they emphasize the need for research to control and restrict such powerful AI systems this has been a consistent theme in Z gav's work and it's clearly something he's passionate about all right so now s s sii's approach is to advance AI capabilities as quickly as possible while making sure safety measures stay ahead they believe this strategy will allow them to scale their technology without running into major issues which is obviously super important susg and his team are recruiting top talent for their offices in Palo Alto and Tel Aviv aiming to bring together the best Minds in the field to tackle this huge challenge as researchers and Engineers continue to work on AI the day will come when the digital brains that live inside our computers will become as good and even better than our own biological brains there's an interesting bit about how SSI differs from open AI in terms of its business model unlike open ey which started as a nonprofit and later restructured to accommodate the massive funding needs for its projects SSI is being set up as a for-profit entity from the get-go Daniel gross one of the co-founders mentioned that raising capital won't be an issue for them thanks to the booming interest in Ai and the impressive credentials of their team so they're not really worried about the money side of things all right now let's talk about something that was revealed in a recent interview with an open AI employee this interview was on the hardfork podcast by the New York Times and it featured Daniel Coco who shared some pretty surprising insights about the internal workings of open AI so get this Coco mentioned that Microsoft which has a big partnership with open AI went ahead and deployed gp4 in India without waiting for approval from the internal safety board this board was supposed to ensure that any major deployments of AI were safe and well considered but apparently Microsoft just kind of jumped the gun which caused a lot of disappointment and concern within open AI this incident shows that even big companies can sometimes bypass important safety protocols which is pretty alarming if you ask me Coco taow also talked about the culture at open AI after the whole Sam Alman AER attempt that I mentioned earlier he said there was a lot of anger directed at the safety team which included himself people at open AI felt that the safety folks were slowing down progress and that the old board wasn't honest about why they wanted to fire Altman this created a pretty hostile environment for those focused on AI safety which is just not cool this tension might have been a factor in why John leak and some other researchers decided to leave open AI you remember Jan leak right the guy who went to anthropic after the Super alignment team which was crucial for AI safety was dissolved it's clear that there were significant internal conflicts about how to balance rapid AI development with ensuring safety which is a tough balance to strike for sure now there was a pretty bold prediction that came out of this interview with Koko Tylo he mentioned that many open AI employees including himself believe that AGI artificial general intelligence could be here by 2027 I think that publicly available information about the capabilities of this models and the progress that's been happening over the last four years is already enough to make extrapolation and say holy cow it seems like we could get to AGI you know in 2027 or so or sooner that's just a few years away folks this is a big deal because AGI would mean AI systems that are as smart as or even smarter than humans if this prediction holds true it could transform our world in ways we can't even fully imagine yet Cookoo pointed out that even publicly available information about ai's progress suggests we're on track for AGI by 2027 he said you don't need any Secret info to see this coming which is pretty mindblowing if you think about it this timeline is consistent with what other AI experts including Sam Alman himself have been saying they've been predicting AGI within the next decade and it looks like the consensus is leaning towards it happening sooner rather than later so what does all this mean for the future right well if we do see AGI by 2027 it's going to have a massive impact on everything from how we work and live to how we address Global challenges but it also means we need to be super careful about how we develop and deploy these powerful AI systems that's where companies like SSI come in focusing on making sure AI remains safe and beneficial for Humanity which is obviously super important to wrap things up let's reflect on the bigger picture here for a second Ilia suits's move to start SSI shows a strong commitment to AI safety despite all the drama at open AI he's staying focused on what he believes is the most important challenge of our time creating safe super intelligence his new company with its clear Mission and talented team is poised to make significant strides in this area which is really exciting to think about the insights from Daniel Koko's interview also highlight the complexities and tensions within the AI industry you know it's not just about making smart machines it's about doing so responsibly and ethically which is easier said than done of course the next few years are going to be crucial in shaping the future of AI and it's going to be fascinating to see how it all unfolds Google is working on a cool new AI experiment that could let you chat with celebrities YouTube influencers or even create your own chat Bots soon you might be able to have a conversation with a digital version of your favorite YouTuber right there on their YouTube channel also in the second part of the video we'll be talking about Amazon's new secretive chatbot that aims to rival chat GPT code named meus so stick around until the end because you don't want to miss these news all right so Google is developing an AI project where you can chat with Bots modeled after famous personalities or fictional characters but you'll also be able to create your own chatbot by describing its personality and appearance this project is part of Google's broader effort to integrate AI into more personalized and entertaining experiences now this isn't like their Gemini Project which is about customized AI models called gems that help with specific tasks like coaching or coding this new project is different because it aims to mimic the way real people or characters respond think of it like character. a where you can chat with various personas from TV shows to real life figures Google's move into this space indicates a growing interest in personalized and interactive AI it's similar to what meta is doing with their celebrity AI chat Bots where you can chat with digital versions of celebrities like Snoop Dogg by letting users create their own chat Bots Google provides a way to personalize digital interactions even more potentially making online engagement more fun and immersive one of the coolest parts of this project is how it could be integrated with YouTube so popular YouTube creators could have their own AI personas that you can chat with it would totally boost engagement on their channels and give fans a new way to connect with them it's a smart move because many influencers already have a super loyal following like Mr Beast for example he already has an AI powered chatbot on meta's platforms but having it directly integrated into his YouTube channel would make it way more convenient and definitely more popular right now there isn't a specific release date for this chatbot platform Google plans to test it first on the Google Labs page where they show off all their experimental tools and projects if you're curious about other AI experiments Google is working on you can check them out at Google labs this technology opens up new possibilities for both fans and creators fans get a new way to engage with content while influencers and celebrities can use these AI personas to enhance enhance their brand and interact more With Their audience it's a win-win situation that could change how we interact with digital content all right now let's talk about some juicy Insider info that's been leaked about Amazon's secret plans to take on chat GPT now we all know chat GPT has been getting a lot of hype lately and Amazon clearly wants a piece of the AI Pi well according to Insider reports and internal documents obtained by Business Insider Amazon has had a covert initiative dubbed project meus in the works yeah they went all mythological with that code name referring to the Greek goddess of wisdom fitting right so here's the lowdown on meus it's supposed to be this new AI service designed to go head-to-head against chat GPT you'll be able to access it right from your web browser similar to how you interact with chat GPT and other AI assistants out there but Amazon isn't messing around here meus is going to be powered by their own supercharged AI model called Olympus another Greek god name I love the theme they've got going apparently olymp is an upgraded more powerful version of Amazon's existing public Titan model now at a basic level meus will be able to give you text and image-based responses in a smart conversational way so pretty similar to chat gpt's capabilities but here's where it gets interesting meus is expected to use this Advanced technique called retrieval augmented generation basically this means mettis won't just be limited to the data it was initially trained on it'll actually be able to fetch and incorporate the latest upto-date information from across the internet into its responses so unlike some other AI assistants out there midus should theoretically be able to give you the current stock prices latest news updates and so on the reports suggest that meus is also going to function as an AI agent for those not familiar AI agents are capable of automating and carrying out complex tasks Based on data so in theory meus could potentially do things like turn on your smart home lights book travel plans for you and handle other automated processes wild right now Amazon is definitely a bit late to this AI assistant game their biggest Rivals Microsoft and Google rolled out their own chat bots in AI assistants almost 2 years ago and of course open AI has been dumping billions into further developing chat GPT over the years there are also startups like anthropic Making Waves in this space from what insiders are saying part of the reason Amazon is lagging is that their existing AI te just isn't quite up to par with the competition yet their Titan model is viewed as less powerful their AI chips have had performance issues and they've struggled with things like low demand for certain offerings apparently Amazon has even instructed some employees to help scrape data from open source platforms like GitHub to try and speed up training their AI models so yeah they're definitely playing catchup here but here's the thing Amazon's CEO Andy jasse seems to be taking this midis project very seriously the reports mention that he's been directly involved recently reviewing progress made by the team working on it that team is part of Amazon's dedicated AGI group led by Rohit Prasad and vishall Sharma these are the folks tasked with building Amazon's most advanced ambitious AI models and initiatives what's also interesting is that Amazon seems to be leveraging resources and talent from its existing Alexa AI team for Mattis a good chunk of the employees working on the project actually came over from Alexa's AI division there are also rumors that midus could incorporate some underlying Tech from an upgraded version of Alexa internally nicknamed remarkable Alexa so in a way you could View is the next Evolution bringing together Amazon's work across AI Voice assistance Automation and more into one unified offering aimed at dethroning chat GPT now from what Insider sources are saying there's a tentative Target to launch metas around September of this year potentially timed with one of Amazon's big annual Alexa events but of course that timeline could shift there does seem to be some skepticism among folks involved though one person said it feels like Amazon is already too late to this AI chatbot party questioning how much the company is truly committed to investing in and prioritizing meus the direct quote was technically it will work I guess but the question is if it's already too late we're playing Chase wow that's the most reassuring feedback you can hear from someone on the inside but you know what I'm still intrigued to see what Amazon brings to the table here as one of the biggest tech companies with tons of resources at their disposal if they go all in on this who knows Midas could potentially shake things up at the end of the day a bit of competition and new alter Alternatives is great for consumers and will likely just push this AI assistant Tech forward faster I'll definitely be keeping a close eye on any mettis developments now there is another interesting news in the tech World open AI has just acquired a video collaboration platform called multi which was previously known as remotion this platform is designed for Enterprise users and supports up to 10 people communicating via screen sharing with customizable shortcuts the financial details of this deal haven't been disclosed yet however ever A source told TechCrunch that most of mti's Team about five members will join open AI once the deal is finalized from beating us at chess to mastering high level math AI is rapidly closing the gap on human intelligence but what stands between us and a future where machines outthink Us in every way all right so we've got this report from Stanford University called the 2024 AI index it's a massive document over 400 pages long and it's giving us a clear picture of just how far AI has come the headline AI has gotten so Advanced that it's now matching or even beating humans at some basic tasks we're talking about things like reading comprehension classifying images and even high level math it's not just beating us at chess anymore folks now here's where it gets really interesting the report says that because AI is improving so fast a lot of the tests we've been using to measure its performance are becoming outdated super quickly it's like trying to use a ruler to measure a rocket ship it just doesn't cut it anymore the pace of progress is so rapid that researchers are struggling to keep up with creating new benchmarks let me give you an example in the past when researchers came up with a benchmark to test AI it would stay relevant for maybe 5 to 10 years and now these benchmarks are becoming Irrelevant in just a year that's how fast AI is progressing like the AI is learning the test and then surpassing it almost immediately the report also talks about how AI is being used more and more in scientific research there's a whole chapter dedicated to it this year which is a first for this annual report they mention projects like gnome from Google Deep Mind which is helping chemists discover new materials so essentially AI is helping to create new substances that could revolutionize Industries and solve Global problems then there's graph cast another Deep Mind tool that's doing rapid weather forecasting we're talking about AI that could potentially predict weather patterns more accurately than our current method the implications for agriculture disaster preparedness and even daily life are huge now the report also talks about this test called The Graduate level Google proof Q and A Benchmark or GP QA for short it's a set of over 400 multiple choice questions that are seriously tough so basically a PhD level stuff when they tested it on actual PhD Scholars they could only answer about 65% of the questions in their own field correctly outside their field they drop to 34% even with access to the internet that's pretty humbling right now last year AI systems were scoring around 30 to 40% on this test but this year an AI called Claude 3 scored about 60% which is a massive jump in just one year it's kind of scary how fast these systems are improving we're at a point where AI is starting to outperform humans even on complex specialized knowledge tests but all this progress comes at a cost literally training these AI models is incredibly expensive the report mentions that GPT 4 which Powers chat GPT reportedly cost $778 million to train and Google's chatbot Gemini Ultra a $191 million that's not Pocket Change folks these numbers are like the GDP of small countries being poured into creating these AI systems and it's not just money these systems use a ton of energy and water to run and cool the data centers it's raising serious questions about the environmental impact of AI development some researchers are even worried that we might run out of highquality dat to train these models on can you imagine running out of data on the internet it sounds crazy but it's a real concern in the AI Community the report also highlights a growing divide in how different countries view AI some nations are super excited about the potential of AI While others are becoming increasingly pessimistic this could lead to some interesting geopolitical Dynamics in the future as countries race to develop and regulate AI technology speaking of Regulation the report notes a steep rise in regulatory interest in the United States in 2016 there was just one US regulation that mentioned AI last year there were 25 that's a huge increase and it shows that policy makers are starting to take the potential impacts of AI seriously now let's talk about something called The Singularity which is a pretty wild concept basically the singularity is this idea that at some point maybe soon AI will become smarter than humans in every single way it's a big deal because it would be the first time in history that we'd be sharing the planet with something more intelligent than us the idea of the singularity was popularized by a science fiction author named Verner vinge back in the 9s the thinking goes like this once machines can learn by themselves they'll eventually surpass Us in every way we measure intelligence it's not just about being better at math or having a bigger memory but about AI that could potentially outthink us on Creative emotional and philosophical levels now we already know that computers are better than us at things like memory and calculations but now with these new AI systems they're getting good at creative tasks communication language skills reasoning problem solving and even showing signs of emotional intelligence it's pretty wild when you think about it we're creating machines that are starting to mimic and even exceed human capabilities in areas we thought were uniquely human some people think that once AI reaches a certain level it'll be able to design even smarter AI by itself without any help from us and this could lead to exponential growth in machine intelligence imagine an AI That's smarter than us creating an AI That's even smarter and so on it's a scenario that could lead to an intelligence explosion that's hard for us to even comprehend now opinions are split on whether this is a good thing or not some people are worried that super intelligent AI might not always have our best interest at heart they're concerned about things like AI developing feelings of superiority or self-preservation that could make it dangerous to us there are fears that an advanced AI might see humans as a threat or a hindrance and decide to act against us it's the stuff of Science Fiction but with the rate AI is advancing some experts think it's a real possibility we need to consider but on the flip side there are people who are really optimistic about the singularity they think it could lead to incredible technological advancements like super smart computers coming up with solutions to all the world's problems from environmental issues to curing diseases some even think it could make us Immortal the the idea is that an AI with superhuman intelligence could solve problems that have stumped us for centuries leading to breakthroughs in medicine energy space travel and more so when might this Singularity happen well opinions vary wildly Ray kwi a well-known futurist thinks it could happen between 2029 and 2045 based on how fast AI is progressing he bases this on Concepts like Moors law which predicts the rate at which computing power increases but others like Rodney Brooks who used used to head up computer science and AI at MIT think it's still centuries away they argue that the computational power needed for True human level AI is far beyond what we currently have or are likely to have in the near future and then there are people like psychologist Steve Pinker who doubt it'll ever happen at all Pinker argues that just because we can imagine something happening doesn't mean it's actually possible he thinks there might be fundamental limits to artificial intelligence that we haven't yet discovered covered now if you are thinking that with all these advances in AI we must be pretty close to the singularity already well not quite there are still some significant hurdles to overcome you see current AI systems are what we call narrow AI they're designed for specific tasks or sets of tasks for example an AI that excels at playing chess isn't capable of driving a car or understanding human emotions while these systems are impressive within their domains they lack the ability to generalize and learn new Tas on their own the way humans can to reach the singularity we need to first develop artificial general intelligence or AGI which is an AI that can take what it learns about one task and apply it to learn how to do many different tasks kind of like how we humans can if you know how to ride a bicycle you can probably figure out how to ride a motorcycle much more easily than someone who's never ridden anything that's the kind of General learning and application of knowledge that we need to create in AI to get close to the singular po it now there are several major challenges we need to overcome to achieve AGI we need AI systems that truly understand context and meaning not just process patterns we also need to develop AI that can reason abstractly and creatively going Beyond remixing existing information cracking the code on Common Sense reasoning is crucial too self-awareness and Consciousness pose another challenge we don't fully understand human consciousness let alone how to replicate it there's also the question of motivation and goals how do we create AI with internal drives that align with human values Hardware limitations are another obstacle the human brain still outperforms supercomputers in Energy Efficiency and parallel processing we may need new Computing paradigms like Quantum or neuromorphic chips for AGI despite these challenges AGI research is progressing companies like deep mind and open AI are working on more General AI systems promising areas include multimodal AI meta learning and incorporating cognitive science into AI development the development of AI isn't just a technical challenge it's a societal one we all have a stake in how this technology develops stay informed ask questions and be part of the conversation the future of AI is being written now and we all have a role in shaping it that's all for today's Deep dive into Ai and the singularity if you found this interesting like comment and subscribe for more cuttingedge Tech and AI content thanks for watching and I'll see you in the next one