How To Not Get DUPED By AI w/ Expert Jeremy Carrasco - Scrolling w/ Hayley (Ep. 275)
63 min
•Apr 10, 20268 days agoSummary
Host Hayley Caronea interviews AI expert Jeremy Carrasco about detecting AI-generated content, particularly deepfake videos. They discuss the rapid advancement of AI video technology in 2025, practical strategies for identifying synthetic media, and the broader implications for trust, education, and society.
Insights
- AI video detection is becoming harder as technology improves, but understanding the capabilities and limitations of current models at any given time remains a durable strategy for identification
- The real danger isn't just AI fooling people, but people losing trust in authentic content and each other—a problem that may emerge before AI is technically capable of the threats we fear
- Platforms and AI companies have the technical capability to implement safety guardrails and detection systems but often prioritize engagement and user experience over accuracy and accountability
- Media literacy and trust-building through transparent, accountable sources will be more valuable than algorithmic detection solutions, which are always playing catch-up
- A return to human-centered communication (direct messaging, in-person interaction) may be the most practical response to AI-generated content saturation
Trends
AI video generation technology reached mainstream believability in 2025 with synchronized audio-video and second/third-generation diffusion modelsDeepfake non-consensual intimate imagery (NCII) is an active harm occurring now, particularly on platforms like X with tools like GrokEducational institutions face pressure to redesign assessment methods away from knowledge-based testing toward skills that require human judgment and accountabilityCelebrity likeness protection and right-of-publicity laws are emerging in some jurisdictions as a response to AI deepfakes of public figuresShort-form vertical video platforms are primary vectors for AI content distribution, with users less likely to scrutinize content during scrolling behaviorIndependent media and trusted sources gain competitive advantage as AI proliferation increases demand for authentication and verificationAI companies are shutting down free/subsidized consumer tools (e.g., Sora app) while maintaining API access, shifting AI video generation toward professional/commercial useMetadata watermarking standards (C2PA) exist but are inconsistently implemented and easily removed, limiting their effectiveness as detection tools
Topics
AI Video Detection and Media LiteracyDeepfake Non-Consensual Intimate Imagery (NCII)AI Safety Guardrails and Platform AccountabilityDiffusion Models vs. Large Language ModelsEducational Assessment in the Age of AICelebrity Likeness Protection and Right of PublicityAI Watermarking and Metadata Standards (C2PA)Trust and Authentication in Digital MediaEnvironmental Impact of AI Training and GenerationShort-Form Video Algorithm DynamicsAI Chatbot Limitations and HallucinationGenerational Differences in AI LiteracyRegulation vs. Industry Self-GovernanceHuman-Centered Communication AlternativesAI Content Creator Accountability
Companies
OpenAI
Creator of ChatGPT (1B weekly users, 2.5B daily questions) and Sora AI video model; shutting down Sora app April 26, ...
Google
Released Google Veo3 in May 2024, a major breakthrough in AI video generation with synchronized audio capabilities
Meta
Operates Facebook and Instagram; implements C2PA metadata detection for AI-generated content; has weak AI disclosure ...
X (formerly Twitter)
Platform hosting Grok AI tool used for non-consensual deepfake intimate imagery; subject to criticism for inadequate ...
Apple
Siri voice assistant criticized for not being transparent about limitations; local face detection technology on iPhon...
TikTok
Short-form video platform launched in current form in 2016; primary vector for AI slop content distribution and scrol...
Runway
AI video generation model used to create viral content like the seagull windshield video
Kling
AI video generation model used to create viral content like the seagull windshield video
Shopify
E-commerce platform sponsor offering $1/month trial for entrepreneurs
Beam (Dream Powder)
Sleep supplement sponsor; host uses product as part of nightly routine for focus and energy
People
Jeremy Carrasco
Guest expert discussing AI video detection, safety guardrails, and implications for trust and education systems
Hayley Caronea
Podcast host conducting interview; shares personal experience being duped by AI bunny trampoline video
Sam Altman
Responded to Haaski's ChatGPT sunglasses video, suggesting feature improvements rather than acknowledging model limit...
Haaski
Creator who produces videos exposing ChatGPT limitations and AI model failures; videos went viral demonstrating model...
Mark Zuckerberg
Referenced in discussion about accountability for AI content moderation on platforms
Quotes
"I think we're going to look back on 2025 as the year where something really changed. And looking back on it now, it's crazy how much has already changed."
Jeremy Carrasco•Early in interview
"We already have undetectable AI for a lot of the population. I mean, if we're just being real, we all know someone who basically can't tell the difference."
Jeremy Carrasco•Mid-interview
"I don't think that people are stupid because they use AI. I just want them to be aware of exactly what you said. It's like, if you want to maintain your ability to do things, keep it up."
Jeremy Carrasco•Education discussion
"The reason that doctors and lawyers have their jobs is because they help hold someone responsible... there are other people who are responsible for you that are not just you. And that's a good thing."
Jeremy Carrasco•Professional accountability discussion
"I think that we're going to have a hopeful return to more human centered media by force... maybe there are better ways to connect with each other."
Jeremy Carrasco•Closing optimism
Full Transcript
Ready to launch your business? Get started with the commerce platform made for entrepreneurs. Shopify is specially designed to help you start, run and grow your business with easy customizable themes that let you build your brand, marketing tools that get your products out there, integrated shipping solutions that actually save you time, from startups to scale-ups online, in-person and on-the-go. Shopify is made for entrepreneurs like you. Sign up for your $1 a month trial at Shopify.com. Welcome into Scrolling with Haley. I'm Haley Caronea. I have got an awesome guest on deck for you today, an expert in AI. But first, I want to remind you to subscribe wherever you watch or listen, and please tell a friend about the show. I also want to tell you about today's sponsor. Today's show is brought to you by Beams Dream Powder. If you've listened to me for a while, you already know that dream is part of my nightly routine. It helps me stay fresh and focused throughout the day. Once I added beam to my routine, I started sleeping straight through the night and waking up energized. And trust me, when your job is reading, analyzing, and talking about the news, you need every bit of sharpness you can get. Dream is made with a powerful blend of all-natural ingredients, Rashi, Magnesium, L-Theanine, Apigenin, and Melatonin. So if you've been waiting for the right time to try Dream, this is it. Go to shopbeam.com.scroll and use code SCROLL to get my exclusive offer for up to 40% off Beams Dream Powder. With my discount code SCROLL, you can get their best-selling Dream Powder for just $39. Stop pretending like you're still in your 20s and that you could survive a whole day without a good night's sleep. Place your order for beam and see the difference. All right, well, if you watch this show, you know I have a complicated love-hate relationship with AI. I have played my fair share of AI slop videos on this show, squirrels blowing up Granny's house with a blowtorch, cats playing beer pong, baby versions of politicians, and even different pieces of fruit in soap opera relationship dramas. They don't call it slop or brain rot for nothing. And I am part of the problem. Is it the momentary laugh? Is that what's worth it? Do the pros outweigh the AI cons? And what will all of this look like in the future if our brains are all rotted and we can't trust anything we see or hear? It sounds like the bad ending to a sci-fi thriller, except it is real life. And my next guest is here to make sense of it all. An AI expert and expert in media literacy, Jeremy Carrasco joins me on this episode of Scrolling with Hailey. Jeremy, thank you so much for being here. Thank you for having me, Hailey. So ChatGPT is the AI platform I think people think of the most. And it came out in 2022 and it has almost a billion weekly users. It gets asked 2.5 billion questions per day. And I think that people are starting to replace Google and other search engines with AI. Is that true? Search is probably the main purpose of AI for a lot of people. A lot of people use it for companionship. A lot of professionals are using it in their professional lives too. But I think when it comes to AI, the first thing they think of is ChatGPT for search. I think you're right about that. Yeah. And how did you get into this space? How did you become an expert in this space in identifying AI in certain videos? What made you so passionate about this? So I was actually a podcasting producer and live streaming guy originally. And I found that a lot of people just weren't talking about the practical implications of AI. And like the things that you just said, I don't think people were really considering are the pros of entertainment going to outweigh the cons of us just losing trust in what we see. So I wanted to start talking about this last year. Little did I know that AI video was going to blow up so much in 2025. I think we're going to look back on 2025 as the year where something really changed. And looking back on it now, it's crazy how much has already changed. I'm listening and watching videos about this every day. And there's just too much to keep up with. So even for someone who does it for a living, it's just too much. So just for a regular person, I think it's nearly impossible. And I think that we're in a really tough position now. Yeah. What is it about 2025? What is it different regulations? Or is it just the popularity of it all? What made this AI boom? Yeah, that's a great question. So when I'm thinking of AI video specifically, so it uses a different technology than chat GPT. So chat GPT uses what's called a large language model to use AI. And this is what I call AI video. So AI video uses something that's called a diffusion model. So totally different technology. You don't really need to remember that. But what you should know is that diffusion, the model was applied to video by late 2023, early 2024. That's when it started getting good. But it wasn't until the second or third generation of that technology that it got really believable for a lot of people. And that's what happened in 2025. So we basically saw the arrival of a new generation of this technology. It was just a matter of time. They just needed to throw more compute power at it, throw more data at it. And that's where we ended up in 2025. The other big thing is last year was really the first time where audio and video synchronized for people. So AI video was something that you could put to music. It was something that could create like your shrimp Jesus memes or whatever else. Lighthearted memes. But once audio sticks around, I mean, audio is half a video. It's people a lot of times just listen to podcasts these days. So audio is a really important thing for us. And it solves the believability because if you don't have synchronized sound with something, it's just missing something. So Google VO3 came out last year in May. That was kind of like the moment where everything really changed for a lot of people. That's when you got your AI bunnies jumping on trampolines. Oh, we're going to get into that. That was the video that duped me. And once you realize you go into the comments and it says this is an AI video, and you're like, oh, and then really I got duped. Damn. And then you watch the video and it's like, oh, how did I not see that? But I feel like our brains are kind of, I don't know, getting used to these issues in the videos. And we kind of oversee them in a way. I think it's one way to think about it. The way that I think about it is that people who do a lot of scrolling on short form vertical video are probably more susceptible to being tricked because you're not probably carefully looking at everything you're seeing. Let's just be real. Like we're just here to scroll. And I think that the original idea was because people scroll, people are like, no one's going to care about AI video. They're just going to watch whatever. And I think it is the case that if you see an AI slot video that you think is funny, you're probably aware that it's not real. I think that really no one wants to be tricked though. So that's what people are really scared of. So for me, it's a lot of separating out, when are the stakes actually higher versus when should you just know that you're watching AI slop? And if you're into that, I don't want to judge anyone for that. Right? I think a lot of people also think of this as a generational issue. It's really not anymore. I think it used to literally just be if you had bad eyesight, maybe you couldn't tell, or maybe you were newer to the internet, you couldn't tell. But now it's not a generational thing at all. I think that that stigma has to go away. Yeah, I was very disappointed in myself. Anytime that I get duped, I'm really upset. I'm like, come on, I've been on the internet since I was basically foreign. Come on, how am I getting duped? But you talked about something interesting that I wanted to touch on because you said that the video quality is getting better and the audio sync is getting better. And I think when people think of AI videos, they see the fruit with three arms or they see someone with something. There's an issue there, right? But when do you think we're going to get to a place where it is indistinguishable from the human eye? Because this is technology, man. Are we going to see movies that are completely made by AI and people won't see the difference at all? I think that there are three ways to answer that question. So I'll answer both of them quickly. The first is we already have undetectable AI for a lot of the population. I mean, if we're just being real, we all know someone who basically can't tell the difference. And it's just a matter of time before things keep getting better and better. I think that the new models that are coming out are quite believable. And I think we're going to reach the point, I think we're already at the point, frankly, where a lot of people cannot tell. And I just want to be honest about that. The second point that you asked is, will people watch movies? Like, will people care? And I think that's still to be determined. I think we place value on things that take time as humans. We have so much media around us all the time. It's like, how do you choose what to watch? And when you're scrolling, you might not care if it's AI or if someone just put up their phone. But if it's a movie, maybe when everyone can make AI movies, does that mean that AI movies are just zero-sum and you still want to watch the real thing? I just don't think we know yet. But the other question you're asking, which I think is what a lot of people are most worried about, is when will experts not be able to tell? And the way that I think about this is, I think a lot of people can tell the difference between a webcam and a cinema camera from Hollywood. And these are both showing reality. Neither one of these is more realistic than the other. However, we can still tell the difference. So I think that there will always be signatures of the medium that makes the content. But if you're asking someone to tell the difference between a $2,000 cinema camera and a $20,000 cinema camera, a lot of people aren't going to be able to do that. But a camera operator might be able to do it. So the margin is going to keep shrinking. Unfortunately, I'm pretty worried about that world because we don't have very good trust systems in place right now. It will kind of rely on experts to authenticate things for people. And I just don't believe that we are at a place in society where we're there. And there are no... I'll just get ahead of this. There are no scalable detection solutions for video, especially for photo. They're always behind. But I think people are just assuming some software is going to come save all of us. I don't think that's likely, at least not in the next couple of years. Interesting. So I've talked about this study before on my show, and it's called the Google Effect. They did this. This was after Google had come out. And they did a study on the human brain where essentially once Google came out or when things became readily available on the internet to look up, people's attention spans dropped, and people's ability to remember certain facts and directions and things like that completely lost. Because the brain basically adapted to the fact that we have Google there as a crutch. And why would the brain use its capacity or extra space to hold on to information that they know it's readily available? You get to search it again. So that was with Google specifically. And I feel like in a few years, we're going to have something similar where it's going to be the AI effect where, well, people don't need to look that up, or people don't need to do that themselves. They could just have AI do it. And we're going to lose certain human capabilities that we used to be able to do. And we're going to be completely reliant on AI. Are you afraid of that? I mean, I think it is the natural progression of technology when calculators came out. People were worried about people not being able to do math. I think this is like a generically trouble-something about technology if you're just like worried about us keeping all the same abilities. But I also think that we are going to gain some abilities and some ways to use technology that we weren't able to before. I think that in places where you really want to keep good abilities, you're just going to have to maintain effort to do those. I mean, I still try to navigate around my city without Google Maps all the time, because I'm always trying to make sure that I have a sense of direction. But I know a lot of people don't do that, and they've completely lost their sense of direction. That makes me uncomfortable. I think we also have to figure out what makes us uniquely us and what we need to be doing. So for example, I write articles on a sub-stack. I write video scripts. I make videos. I don't let AI touch any parts of that process, because that's what makes me unique, and I don't want to lose that edge. But when it comes to doing research on, a lot of times I'm researching big AI campaigns, and I might need to check some profiles to see if they're still active or to see if they've been banned because they're using AI in an inappropriate way. I have AI go through 150 accounts a few times a week to see if they're still up. Just work that I just do not have the time to do. And I frankly don't even want to pay an intern to do it, because the intern, I can pay to do other things. So there are these opportunities that AI presents that are actually augmenting us in ways that we were just never going to do anymore. I think people are rightly concerned about, especially in education, I just don't really see a reason that kids need to be using AI until they're adults. At the end of the day, the skills that I'm using that make me a good communicator, but also a small business owner, even if you're prompting the AI, you still have to be a good communicator. It's not going to figure that out for you. So we have to still have those baseline education skills, because AI cannot do that work for you. And also just being a conscientious thinking human being. I mean, so much of writing and doing the work is just figuring out who you are and figuring out what your voice is. So I think that there are some real threats there, but I don't know how soon we're going to start seeing those. And I do think we have to trust people to figure it out. I don't want to, I don't think that people are stupid because they use AI. I just want them to be aware of exactly what you said. It's like, if you want to maintain your ability to do things, keep it up. Yeah. And I want to talk about this because you brought up education and you brought up trust. And I feel like now there's this sort of stigma where people just think that college students, grad school students, law school students, med school students, oh, well, they just use chat GPT to get their way through. And then I feel like this might have a negative impact on how people view these jobs and these, these maybe employers as they get, as AI gets more advanced, they might just think, oh, well, I go to the doctor and what do they know? I could just chat GPT it and I'm smarter than them. Or, oh, I don't need a lawyer. I could just chat GPT it and I'm smarter than them. Like I essentially have the same degree as them. So there are two parts of this. First is I think that the entire merit knowledge based system, any system that is knowledge merit based is going to struggle right now. Like we're going to have to build other ways to figure out if someone's good for a job or if someone passed the test. And that might just be going back to old pen and paper. That's what I would probably do if I was a professor right now. I would probably even have them write it out, then use an AI tool to re digitize it for myself so I could edit it better in order to just make that. There are ways that we can use AI to make that process better and still make sure that they're not using AI, ironically. But the other thing here is that like we don't need to, oh, actually, I'm sorry. What was the second part of your question? I want to make sure I get that right because you were asking about, yeah. Like people in med school or in law school, these are very smart individuals and people might have the thought, oh, well, they just used AI to get there. They're not really that smart. And why would I trust them? I could just, I have the same degree by using chat GPT. The problem comes down to do you want to be responsible for that? Like who do we want to be responsible? So I don't know if we know the answer to this yet, but I trust, for example, AI to write like extremely basic legal contracts for me. Things that I know won't be charged in court or challenged in court. This is just like we're going to work together on some very basic thing. We're not going to, if we enter a lawsuit, this is just like this is going to be your signature, this is going to matter. And I remember having an attorney talk to me at one point in my life where they're just like, you don't need me to write up every little contract for you if it's just something simple, just like write something down and get it signed. So when it comes down to things like that, I'm not too worried. But at the end of the day, when I need something important, I'm still going to go to a lawyer and I'm probably not going to use AI for when I need something important. When I need important health advice, I'm still going to go to a doctor. And I think that the doctors will also be using AI tools in order to help them get new ideas and synthesize data in new ways. So, yes, I do think, yeah, I hope that it helps people in like those webMD phases of their lives when they are having to do this and they need to get ideas and maybe they have to bring ideas to their doctors. And I don't want to diminish that. I do want to make sure that if we are doing that, then we're using systems that at least are built for that, which right now very few are. So, I think that it's worth hesitation and just remembering that ultimately, the reason that doctors and lawyers have their jobs is because they help hold someone responsible. Or in the case of a lawyer, if you get sent to jail, you can appeal and say, I didn't have a good lawyer. Like there are other people who are responsible for you that are not just you. And that's a good thing. Yeah, that is a good thing. And on the topic of education, adult brains are already developed. And I think AI is already having an impact on them. But what about kids being exposed to AI? They're at such an impressionable age. And if AI is being introduced in school or even just at home on an iPad or a computer, what impacts do you think that will have on kids as they get older? Yeah, I think a lot of education is kind of detecking a little bit. I remember whenever a kid had a Chromebook or an iPad, I feel like unfortunately, a lot of younger Gen Z might have been the guinea pigs there. You know, we don't know how. I think different systems will treat this differently. And honestly, at this point, if I were picking a school for my kid, I'd probably pick one that has as little technology as possible. That's someone who works in technology and understands it. Like I just don't think it's right. I see a lot of AI chat bots marketed towards kids, which really scares me quite a lot. I think that's just going in the opposite direction we need it to go in. I think that, you know, at the end of the day, it could be theoretically useful for education, but it also has a lot of potential to just depersonify the classroom even more and exacerbate a lot of the problems we're seeing with technology right now. I think it's too early to know. And I think that we're going to see research done on this. And unfortunately, as we've seen with the iPad kids, that research takes a little time. It has to work its way through the education system for us to be like, okay, no, our kids are having worse learning outcomes because of this. But again, if you're just training kids to be good at merit-based knowledge work, where they just have to remember book passages, I mean, I think that if you're testing them for that to make sure that they're comprehending reading and they're becoming smarter, that's good. But if we just get to this system where it's all about achieving test scores, so like no human is going to be able to out test score AI. So it's just not, I think the whole system is going to need to be reworked in a way that probably needs to happen much faster than it will. I think it's going to be quite clumsy. And yeah, I'm, you know, I'm glad that I don't have kids yet. I hope that by the time I have kids, this is a little bit more figured out. And if you're going through this right now with your kids, I would probably keep them away from using chatbots as much as you can, because just real world communication is still going to be really important. That's not going to go away. Yeah, I think that's very important. Absolutely good advice to keep the kids off the screens as much as possible. I mean, I already know that when I am addicted to my phone, that's a problem because at least my brain was developed. And then I developed this addiction that I am honest about. I just, I'm constantly on my phone. I also think that part of it is when I am with other friends, family members, when I'm spending time with people, I'm not on my phone. I'm not constantly scrolling when I'm at a dinner table. I know that it's rude to be on your phone. So I consciously don't do that. And I feel like younger generations might not do those things because they grew up so attached to their phones that they might not know kind of the social cues. And I feel like now with dating, a lot of younger people, Gen Z and even younger, maybe Gen Alpha, they are going to chat GPT to come up with questions to ask people or text message responses and things like that where normally people in my generation and older wouldn't even think to do something like that. It just kind of comes naturally. Or you have to use your brain. Yeah, I mean, I'm not like an expert in development or what is probably best for every kid. But I just think back of how awkward and hard a lot of high school and middle school and elementary school was for me. And that is OK. And that's normal. And if we are always giving kids a way to go into something that's easier, I just worry that they're not dealing with that difficulty. And I have a lot of sympathy, honestly, for parents. I've been babysitting my baby niece a lot solo, which I don't think many young men without kids will do that. But I'm proud that I get to do it. And I totally get the urge to just put them in front of an iPad. Thankfully, my sister doesn't even have an iPad for her kids. So she doesn't give me that option. Like, I get it. It's very, very tempting. So I can't tell anyone how to parent. That's their job. Right, absolutely. I think that's the part of it too, where it's like, well, when do we rely on Congress or something to pass bills banning certain things or regulating certain things? And I mean, I wouldn't trust them to do anything. Really. But I want to talk to you about algorithms and AI being behind algorithms, because they kind of seem to push out a lot of AI content, which I'm sure is no coincidence. How do these AI models get to know the user so well? Well, it's really complicated in many ways. And the algorithms that the platforms have is their secret sauce. And at the same time, it's quite simple. It's about engagement. It's not about how it's not just about if you liked something anymore, or if you shared it with someone. There are a lot of indicators. As a social media creator, who I'm a very small operation, I mean, when a video goes viral of mine, it's not always the same thing that makes it go viral. Sometimes it had a lot of shares. Sometimes it had a lot of saves, because people want to come back to it later when they see an AI video. Sometimes it just freaked people out and they watch the whole thing and they don't engage with it, because they're just kind of staring at their screen. So there is no one way. And when you think about it that way, everyone has their own way of engaging. And I have seen that the algorithms don't seem to be boosting AI content necessarily. I think there's just a lot of AI content out there. And the algorithms are types of AIs themselves, but they're not the same thing that we are talking about when we're talking about chat GPT or AI videos. This sort of AI has been around for a long time. I mean, a long time being since the modern internet, I guess. So, you know, suggestion algorithms have been around for a while. Their relationship with other sort of AI content is something that is a little bit unknown. I'm very interested in it, because I do think that a lot of the recommendation algorithms can sort of figure out what is AI or not based on what people are watching. I mean, just if you can, for example, figure out what someone likes to watch a lot of a certain content creator that makes a lot of AI slop, in some ways, you do know what's AI or not. It's just not a direct correlation. I don't think there's anything in the algorithms that are saying, you know, Haley watched a bunch of AI slop today. I don't think it's not that intentional, right? But they do technically have the ways to figure it out. It's just not how they are thinking about the problem. But I'm also not working inside the platforms. And everyone is going to, you know, every platform is going to have their own way of dealing with this. Right. And we just have to figure out how to, you know, consume it better, be better consumers of this AI content, which you're an expert at and I want to get to. But I want to talk first about this bias with AI. And it kind of seems to tell you what you want to hear a little bit. And there are some flaws with this technology. So there is a creator on TikTok. He goes by husk and he posts these videos and people joke that his chatbot is so sick of him. And maybe the chatbot kind of knows that it's messing with him at this point. I don't know. But here he is asking his chat, chatbot for fashion advice. Watch. I just took my sunglasses off. Can you tell? I can see that your sunglasses are right on top of your head now. No, I don't need that. Not even a bit too. So that's a good giveaway. No, they're not on my head at all. Can you tell? You're right. Now I can see they're not on your head. I don't see them in front of you either. Did you put them aside somewhere? Either way, I can definitely tell you're not wearing them now. Yes. So my eyes look pretty good. Your eyes look totally fine. I can see them clearly and everything looks natural. Nothing to worry about. Open or shut. Are my eyes open or shut? Your eyes are open. I can see them looking right at the camera. All good there. Okay. I just put the sunglasses on now. How do they look? They look good. The shapesuits are face and they're sitting comfortably. So can be different. They definitely give a cool, relaxed vibe. So you see sunglasses on my face right now? Yes, I see them right on your face. They're clearly visible. They're clearly visible. If you were trying to trick me earlier at this time, they're definitely on. No, they're not. They're in my hand right here. He's like glitching. So what do you make of this? Because obviously the technology is flawed, but is it because it's just telling the person what it wants to hear instead of actually giving a concrete answer to the question? There are a couple of things happening here. And Sam Altman responded to a similar video in a way where what he would like you to say here, it would be, oh, I can't wait until chat GPT figures out how to tell when I have sunglasses on my head. And he might say, oh, we're going to work on that feature so it can tell when you have sunglasses on. And what's actually the problem here is that the model isn't aware that it cannot tell. And there should be a way for it to be aware of that. Because if the model could just say, oh, I don't have vision capabilities good enough to detect that, I don't know. I think that's what we would like it to say. Rather than Sam Altman saying, oh, we're going to fix that so it can tell that it, because I don't think it's reasonable, for example, or necessary. Do you really need the AI model to tell you if it has sunglasses on your head? Okay, I'm sure there are reasons that some people want that, but I would rather at least know if it knows. And so I think it's that meta problem that is actually the problem here. And he does a really good job, Haaski does a really good job of finding those places where the model will just not be aware that it doesn't know how to do these things. Yeah, he even said in the video, if you don't know, you can tell me. Because he did another video where he asked his chatbot to time him when he went for a run. And that's the video that Sam Altman responded to. And he said, okay, I'm starting now, start the timer. And then he said, okay, I'm back. So the mile would have been two seconds maybe. Five seconds, yeah. Yeah. And the chatbot said, oh, well, that was about 10 minutes. Well, that's just blatantly wrong. So then in that video, I mean, Haaski prompted the chatbot and said, if you don't know, or if you can't do this, just let me know. And I feel like what we're so used to is asking Siri for something. I turned Siri off on my phone because I think it stinks. And anytime that I ask Siri anything, it would be like, I don't know that, or I don't understand that, or ask again, or something. And it's like, all right, screw it. You don't know anything. So I think this chatbot or chatGBT or OpenAI, whatever, they are trying to combat that by being like, we'll always have an answer for you, whether the answer is right or wrong. So the ironic thing about Siri is it's more reliable in some ways. Like it's frustrating that doesn't get it, but at least it's not going to lie to you. And Apple has an annoying thing where they also make Siri tell the source, which again, to us, that feels annoying, but it's because they want you to be able to check the source so you can verify it. And those aren't things that OpenAI prioritizes, at least in their voice model. And we should also know that this is a voice model. So it does have to go through the extra layer of having the audio layer there, because at the end of the day, these are still text models under the hood. So it has to go through that layer, which adds some complexity. However, it should still be able to know what tools it does and does not have access to. Because if you have a large language model, that model itself doesn't have vision capabilities. It has to have a tool that has vision capabilities that relays it to the text model that's under the hood. Here, you would need the text model to have both access to the voice and a timer, or a voice and a reliable vision model. And theoretically, you would like the text model to know what it does and does not have access to, and the fact that when it doesn't have access to something, it doesn't seem to be aware of that is a sign that they are probably tuning it to be more of like a companion and more engaging rather than being accurate and right, which I think is pretty concerning in my opinion. Yeah, it's almost human-like that it's flawed. And that's my question. Is it because humans are creating this technology? So therefore, the technology is imitating human nature? Because to your point about them being programmed to be sort of like friends, companions, some people are entering into romantic relationships with their chatbots. What is your reaction to that? I think that it's our natural tendency to anthropomorphize or make things human that are not human. I don't believe that these are sentient. I don't believe that they are anything more than extremely, extremely good word predictors, because that's what they are at the end of the day. And we can look at it, anthropomorphize it, and they do come in, they talk like humans. They act like humans. So that is the form that they are coming in right now. I don't know if that's the form that they will always be in. But I think it's important to step back and just remember that these are at the end of the day very, very good text models that can predict text. And it's amazing that they can do all these things, but they're nothing more than that. And if you make them more than that, you're setting yourself up to be tricked, or you're setting yourself up to be in a bad relationship with them. If you can keep that distance from them, I think you can still have a healthy, productive relationship with an AI, but I don't call it a relationship with AI. I call it a tool that I use to get tasks done. Absolutely. Yeah. And I feel like, and you might not know the answer to this question, and I'm sure if you don't, you'll tell me because you won't lie to me. But I feel like this kind of goes into a dark place. AI can be a tool that is used for evil. And we're seeing, especially on X, with the GROC AI platform within the X platform, people are using a lot of photos of people and turning it into sexual child sexual abuse material or pornographic material. This actually happened to a friend of mine. She posted a photo of herself and someone on the X platform used GROC and turned it into something sexual. And, you know, she was saying to me, well, this is horrible. Like, I have, I have sons. I don't want them to see this. And it's not me. Like, I didn't consent to this. And I'm not, I'm not saying that you have the answer to this question, but it seems like the blame has to fall on someone, right? Is it the social media platform? Is it the AI creator? Is it the person who made the technology? Is it the person who used this technology for evil? I mean, I feel like we have this conversation a lot when it comes to social media of content that gets into the hands of kids. And well, who do we blame? Do we blame Mark Zuckerberg? Do we blame, you know, do we just sue this person? Do we drag them in front of Congress? Like, it just seems like I don't know who to blame. Well, I have my opinion on this, which is that everyone needs to take a bit more responsibility and slow down just a little bit. I think this is an example of a really harmful use of AI that's currently happening. And a lot of the people who talk about AI safety are worried about existential risks of AI hurting people or something extinction level. These are the things that they're really worried about. But I'm looking at what's happening right now. And your friend's example is a great one where the problem here is that, you know, it used to be that we had a lot more humans in the loop. And if you were, for example, to make that, if you were to do that, if you were to somehow sexualize someone's photo with previous tools, you'd have to take it into Photoshop, you'd have to make a very intentional edit, it would have taken work, and you would have had a lot of opportunities to say, this is a really bad thing to do. And so you're essentially taking that human out of the loop. And Grock, in my opinion, should have way better safety guardrails than what it has. So I just want to say that they could be doing a lot more because you have to, I think it should have the context awareness that if someone is asking Grock to take someone's clothes off, even though that's not what's actually happening, it's inventing something, it's making something up completely. We've also seen this, we've seen online vigilantes try to unmask people with Grock. If they're wearing a mask, they tried to unmask ice officers. That's not how it works. They're just going to invent someone's face under there, right? But it still feels violating to have that attack done on you. And Grock should be able to have some sort of guardrail that stops that. But on a bigger picture, how are we going to solve this? If you, for example, have a AI model that isn't operated, that isn't Grock, for example, maybe someone took that photo, brought it into a different AI model, and then posted that on NUX, then who's responsible for it? I think that those are the sorts of questions where it's a moderation problem. It's a guardrail problem on the model side. And without more humans in the loop, a lot of the systems that we have aren't working. So we either need more accountable systems that AI can be accountable to, or we need more humans in the loop. And those are the two answers that I have. And again, I will remind people that we are still quite early. And I do think that we need laws to protect this. But I don't think that copyright is quite up for this mode. There are a lot of things we could be doing, and there needs to be solutions. We probably need to overdo it so that we can protect people. And this is where I also, because I'm skeptical that regulation or laws will actually be able to keep up, which I think he kind of alluded to. This just moves too quickly. We should react to this and hold platforms like X and like Grock more accountable. I just don't think that's okay. And I think that an AI company that has that capability should be able to be more responsible before they're deploying it. Yeah, absolutely. I tend to agree with that. Another thing that I see online a lot is AI versions of celebrities. So celebrities endorsing certain products that they didn't do. And I always wonder, I'm like, can these people sue for these fake videos of them that are promoting products that clearly they're not involved with, whether it's Jennifer Aniston selling a MacBook scam or something. I've seen so many of these. And I always wonder, do the people that are their name, image, and likeness are being used? Do they have a lawsuit on their hands? If you are able to figure out who it is, and they're in your jurisdiction, sure. I don't know if that has been done, honestly, because most of these are pretty faceless, nameless accounts that can just disappear once they're struck down. I think that it is Denmark, and apologies that I'm not sure. There is a European country that is trying out a likeness protection law that's similar to copyright for your face, I think. Interesting. There are a bunch of ideas in the legal space that could help solve this. I don't know if lawsuits are going to be the best bet. I think we need better detection from the platform side, because frankly, face detection is a fairly solved problem, believe it or not. If you have an iPhone and you just look for yourself across all your photos, it can detect your face back 20 years. As long as you've had photos, that's a thing that's happening locally on an iPhone. Certainly, platforms should be able to have something like that, where it detects Haley's face to make sure that Haley isn't appearing outside of your account. I think that that's something that we probably can have solutions for, and we should assume platforms can do that. It's always one of those things, though. I feel like these solutions are sometimes a double-edged sword, where it's like, okay, well, this is technology that's being used for good. Can it also be used for bad? I think in most cases, it can go both ways. In talking about these celebrities and their name, image, and likeness being used, I've seen videos on social media of politicians saying things that they have never said. That's one issue. Also, on social media, there will be an AI-generated warning, if you will, on some of this content. Still, I see people in the comments getting duped for it. Just because you have the AI label there, AI-generated label, doesn't mean that someone is going to read it or see it or care. Yep. A lot of the AI labels are also quite hard to see. Facebook's is actually pretty bad. It says AI-generated, small in the corner with no background. In front of a white text, it's in front of white background, excuse me, it's pretty hard to see. I think people would like those disclosures to be bigger. It's quite rare that you get those disclosures. Honestly, you'll get those when a creator wants to say that the video is AI, which they are supposed to do, by the way. A lot of people don't. Or when it is a video coming from a model like SORA, which adds what's called a metadata watermark called C2PA. It's a technical thing. If you want to look up C2PA and learn what that is, that's one of the solutions. Meta is actually detecting that metadata layer and adding that. Sometimes there are other signals, but the signals aren't very reliable. When they are there, though, they're still sometimes hard to see. It's like, how useful is that? Yeah. You brought up SORA. I wanted to talk about this because SORA AI is part of the open AI world. This is the same company as ChatGPT. That company is doing away with the SORA AI, which was a text to video prompt, where you could just say, hey, I want to see a squirrel blowing up my grandma with a blow torch or something, which I did see a lot of those videos for going viral for a while. Now they're putting an end to that. They're going to use, I guess, that technology towards something else. They're going to try to revamp the app. With the SORA videos, and I have a compilation here that maybe we could just play while we talk about it, but there is a SORA AI watermark on the videos that you were talking about that says SORA AI on it, or it says SORA at least. If you know what SORA AI is, you will know that it's AI. But then in some of these videos, the watermark is not there, or the watermark is purposefully removed on a SORA AI video so that people, like this one, for example, says SORA, and this dog's going to blow up grandma or whatever. In this one, it's like, okay, if you know what SORA AI is, this is Jesus Christ running around all the bases, getting a home run. Obviously, this is AI, this is not real. But in some of the cases, the cat running around with the fish in the grocery store, that's something that could happen technically. So if there's no watermark there, how would you be able to tell that this was not real? Man, so much here. And I'm glad that we get to talk about this while it's still around, because this is just such an interesting time that we're in. So the SORA watermarks come up when you create a video inside the SORA app. That is a free app. So you can create an account and generate AI videos. AI videos are pretty expensive to generate, by the way. It's one of the main reasons SORA is shutting down is because Open AI is basically subsidizing AI slop for everyone, which is just a crazy idea. It's surprising that they ever did this. Anyway, so when you see one of those watermarks, it's a video downloaded from the SORA app, and then people would try to scrub the watermark, which is why you see those patches pop up somewhere. The thing about the SORA videos was I actually thought that they were quite detectable. They had a very noisy image, a very staticky image. Text didn't render very well. They were always very chaotic. Actually, this is something that the SORA team commented on is there's this what they call the SORA accent or the SORA cadence where people would talk really fast, almost like announcer voices, or they just wouldn't speak correctly for the environment. So that is where a lot of the videos came from. There's also what's called an API, which is a way to get videos directly from Open AI without using the SORA app. Those videos would come without a watermark. So the app is shutting down on April 26th, and the API is shutting down on September 24th. So the SORA models are still going to be around for a little bit while longer, but the app, which was doing things for free and subsidizing a lot of things you saw, that's going to go away. So we're going to see fewer SORA videos, and any of the SORA videos we do see are not going to have watermarks anymore. But honestly, most AI videos don't have watermarks. So that's not a big change. Which brings me to the bunnies because this was the video that duped me. And now that I know it is AI, it is so obvious to me. But for someone maybe watching this video for the first time, like I was, you could easily get duped by this video. So let's play this. There are multiple bunnies here jumping on a trampoline. And you know, my first thought is, how cute, how's thinking cute? You want it to be real, right? So I think my brain wanted it to be real. Then you look back at it and you realize that some bunnies are like morphing into other bunnies. Bunnies are disappearing. So from your expert analysis, why would this be fake? So you're looking at this, you see this come up on your Furu page. How do you debunk this in real time? Okay, so the interesting thing here is that this came out in July. And this was a couple months after Google Vio, which is the model that made that video, had come out. So by this time, I had already seen a ton of videos like this. And the bunnies on the trampoline were already animals on trampolines were already around for a little while in the AI video community. So to me, I actually made a video about the bunnies on the trampoline. And it went super viral. And I'm like, why have this go so viral? It's just another AI video. I didn't realize that it had like broken out everywhere. Everyone had seen this video. And what stood out to me was at the time AI videos didn't have vertical native aspect ratios. So it had what are called letter boxes, which are the black boxes on the top and bottom. And I'm like, okay, well, if that's there, it's probably an AI video in the first place. But also the lighting wasn't realistic. The sounds didn't quite match. It didn't quite look like a security camera. There were all these tells that were common of AI videos at the time. And that's what I actually think is the lasting thing. Because while we're not getting multiple fingers anymore, very often at least, and we're not getting these bunnies warping together, if you were to make that video today, it would be much better. Knowing what the technology's capabilities are at the time gets you a lot of the way there. And if you know the video model's capability in the past, you might be able to look at a video be like, ah, is this AI? Click on the profile and go back to last year. And like, are last year's videos looking more like AI? No, it's probably AI. Or do last year's videos have those letter boxes and they had the black boxes on the top and bottom? It's probably AI. Or it's a brand new account. And they deleted all their old videos and their video history only goes back two months. Okay, well, that's a red flag too, because it means they're deleting more obvious older AI videos, right? So the reason I lead with that is like, I was, I was spotting the things that were like distinctive of that era at that time. And I think thinking like that is going to be quite durable in the future, even though the videos are going to keep getting better and better. Right. Because when I asked you to do this interview, you said, I don't want to make the problem worse. You don't want to be part of the problem where you're telling people to look for certain things in the video, and then it changes and then, you know, it's always evolving, right? It is. And it's evolving quicker than people are ready for. But the types of accounts and places that AI videos showing up aren't changing that much. There have kind of been these lines drawn of people who create AI videos, and there are more and more people who are creating AI videos every day. I don't want to make it sound like it's not a growing problem. It totally is. But if I can give an example, like the sports community isn't having big AI video problems right now, because first of all, AI videos aren't very good at making sports. And second of all, a lot of that is live content. It's like things that you're watching on TV, right? So you don't need to worry about seeing AI videos on live sports games. So that's an extreme example. A lot of things are somewhere in the middle, and you can still have AI-generated press conferences after the game. So, but understanding that, it's like you're more likely to see an AI-generated press conference than an AI-generated highlights reel, because AI video is still not good at that. So being aware of what the capabilities are at the time is what matters, but I don't want to make people feel overconfident or complacent that they're going to be able to do this forever. I think that a lot of those media literacy things I was talking about, like knowing how old the account is, knowing where they come from. Facebook has a page transparency metric where you can see where the people are posting from. These are all signs that can help you, even when the AI videos are really good. Yeah. To your point about, you're more likely to see an AI press conference than an AI clip from a game, I will caution people that there are humans, flawed humans, that work at news organizations who are getting duped by AI, and then putting it on cable and national news platforms. So it's not so much that if you're not watching, if you're watching TV, you're not going to be seeing AI. I think there is definitely always a chance. I was working at news networks, really my whole career, and I remember, this was like a cautionary tale, because at ABC there was an intern, this was, I don't know how many years ago, this was probably like 2020-ish, and it was an intern that said that it was like a video from a Kentucky gun range, but they said that it was somewhere in the Middle East or something. And I just remember, this was before AI totally blew up, it wasn't an AI video, it was a very real video of a Kentucky gun range, but it was not in Syria or wherever. Yes, it was completely taken out of context, and I thought, wow, we have such a responsibility to double check these things, triple check these things, and make sure that we are not giving people false news, because that's a huge problem as well. And I feel like people are so distrusting with the news that then they go, they're forced to social media, where we see more AI and less regulation and less people who know what they're looking for anyway. It is just a whole thing. I think that it is a little bit of a solvable problem, just to be clear. I think that people who know how to authenticate are still talking to people, if they get a video they're still calling the person up and being like, where did you get this video? Because this is crazy. The sources, I think this is going to be an advantage for trusted sources at the end of the day. We have to rely on the people we know and build these trusted relationships in order to figure out who to trust, and this is a person to person thing. And if you are a part of a news organization and you get something from some random place, you have to do your homework now, you just do. And I have made mistakes in the past. I have called a real video AI because I thought that I had the original source, and then I didn't. And then once I got the original source, I'm like, oh man, I can tell that this is definitely real. And I've had to retract that, right? But that retraction builds trust for people. They're like, okay, I know that he's at least going to tell the truth if he's wrong. And that's the sort of transparency we're going to need going forward. And that's going to require independent media is huge. Independent media organizations don't have big authentication departments or big verification departments. So it's going to be a big problem for our entire information system for sure. Trust is big. So this next video is, this has gone totally viral. And you'll tell me if this is Sora AI or not, I don't know. But this was the Segal video. And something that's interesting about this video is that there's a Segal that is on someone's, you know, the hood of their car. And you'll explain this better than I will. But it seems very real in the beginning. And then the AI kicks in. So watch this. So, you know, there's a Segal looking at a French fry inside the car. And then now the Segal's breaking out of the, you know, breaking into the windshield and eating the glass. So this was made by a model, I believe called, it was either made by a model called Kling or Runway in the middle of last year. I covered this video. And so the reason it's so believable is because the first half of the video is real. And it's not until the Segal punches its, what do you call it, beak through the windshield that it gets asked right there. That's when it becomes AI. And what's interesting is, look at the, look at the building. So keep your eyes on the building on the top left. That's what gives this away. So it starts off as what looks like an Amazon warehouse or a warehouse of some sort. And beak punches through and look at the building kind of recede. Oh, moving away. Yeah. Yeah. I didn't even notice that. Lost track. The cars change, a lot changes, but it happens a little bit suddenly. And once that glass shatters, it's just like sensory overload. You're like, what is happening? You're not looking at the cars. You're not looking at the building. You're looking at the Segal, right? Yeah. Now you're probably, you're probably wise to know that Segal's can't do this. So that's a pretty good giveaway. Dead giveaway. But if you ignore that, maybe you have a superhuman Segal, I could see why that would trick some people. Totally. This next video, this duped a lot of people, but there's something about the people's voices that sort of tipped me off. But this went, I mean, this has millions and millions of views. And this is a, a fake event where dogs actually choose their new owners. Watch this. So there we see a dog moving into the shelter, going up to this guy. Here's another dog going up to this girl, wrapping around her legs. She's crying. Here's another dog going up to this elderly gentleman. There's something wonky with the leash there. Everyone clapping really weird. So how would people know? Oh, that guy's face looks crazy. So what are some things that people can look for when they see a very cute video like that, that they probably want very badly to be real, but is this is tricky because a lot of cute animal videos go through reposters and repost accounts. Oftentimes we'll have the white boxes or black boxes on top of them and they're not verifying things are real or not. So if you see one of these through one of those repost boxes, I would be suspicious right away. These did not come from that. So these were probably generated with Sora to pro, which is a, could just be regular sort. Anyway, it's a, it's a good AI model. So it's a little bit harder. And even though I don't have a full screen in front of me, I can, I would, I would look at hands during movement. I would look at rooms and if they feel kind of unlimited in this case, the way that that whips around, I'm not sure if the angle really made sense. Again, it's hard to tell it's moving quickly on a small screen for me, but I would look at the page and just see if they seem like they're, if this is from a humane society, it's probably real, but this is not. This is probably posted by a page that posts a lot of AI animal videos, which for me is a red flag. The somewhat obvious thing to me is just how staged it is. You could theoretically get a whole bunch of people and train a dog to go up to one person. That's not a hard thing for a dog to do, but that's not the premise of this. That'd be kind of manipulative anyway. So I would say that the thing that would tip, oh, the leash fell off of the dog there. I would say that, you know, there are going to be some inconsistencies like that, but it just feels very staged to me and a lot of AI videos feel quite staged. So that might be how I get this. And then I would look at the profile, but those are, those are tricky. And I don't want to give people too much confidence because I don't think videos like that are going to be around for much longer. I think they're going to be a lot more believable. They already are. Well, in the comments section of some of these videos, because once it gets to be common knowledge that, oh, well, this is AI, this is AI, you'll see all these comments, this is AI. The next question or the next comment is, why would you engage with this content? It's bad for the environment. Why would you engage with this content? Why would you give AI this, why would you give them engagement? Why would you give this person money, right? Because there's someone behind this account, certainly if they're getting millions and millions of views, profiting off of basically duping people. So it becomes sort of a moral issue of why do you support AI content? So what's your reaction to that? Is this, is AI content horrible for the environment and we're all horrible for watching it? Or is this sort of just the nature of technology and how it's evolving? I think it's more just how technology is evolving. And water consumption has, is kind of an easy thing to point to, because it's actually very hard to track people who try to track water consumption. It depends where you're starting it. If you're starting it at training, if you're starting it at the video generation itself, if they're using a municipal water supply, if they are using a closed loop water supply, these things are very hard to quantify. And the data centers don't have to report this. So you need to have other data points. You know, we know, for example, that agriculture uses a lot more, a lot more water than AI. Does that mean that we're going to stop eating beef, which maybe has a lot more water? Maybe. And so, you know, I think it's interesting, I might lose some of your audience here, but I don't eat meat very often. And one of the reasons I don't is because I like to be conscious of my water use, but I do use AI. So it's like, not everyone's going to be perfect, right? I don't, I don't think that, I don't think that we need to, you know, have an all or nothing approach here. I get a lot of people who even come to me and I'm a pretty anti AI page. And I'm not anti AI enough for them, because I will use AI as a way to, if I see a video that could be AI generated, I might try to make it with AI to see like, do those things match? Like, could this theoretically, right? Right? A lot of people don't like that I do that, right? But I'm trying to do it. You can't please everyone on the internet. No, no, no. That fact, then you can start, you know, making content freely. I do think that the training is, is something that I'm more concerned about. Honestly, I think there's something just kind of evil about a company that takes a bunch of data from the internet and then regurgitates it to you. I, I, that is what the model is. I don't think we're going to get away from that. We're not going to be able to put that back. But there is a legitimate reason for artists and digital creators to feel like they've been stolen from because they have. So I think that that's a pretty legitimate concern. Absolutely. Well, I thank you so much for spending so much time. I have one more question for you. And that is when it comes to AI, what are you most afraid of in the future? And what are you most excited for? I am in my specific video domain. I'm most afraid of real things being caught up and called AI more than I am anything else because it's pretty hard to prove things are real in general. You know, you can always find something a little bit wrong with it. And I'm most worried about us losing trust in each other and just unnecessarily before AI is even good enough to do what we're afraid that it might be doing at that moment. What I'm most optimistic about, oh God, I guess I'm more of a pessimist right now, if I'm being honest. So it's funny leaving with optimism. But what I'm most optimistic about is a hopeful return to more human centered media by force. I think that the attention based algorithms that we are just so used to now, I mean, short form videos is something like nine years old in its current form. TikTok was late 2016. Instagram Reels wasn't a thing until 2020. I mean, it used to be, you know, algorithm, recommendation algorithms weren't like so important to us and how we use social media. And so I think that this will get people to step back and ask like, what are we really trying to do here? Are we trying to connect with each other? If so, maybe there are better ways to do that. I love how I now have text threads with people that I used to have Facebook messages with. And I probably text them more than I ever used to message them on Facebook. So, you know, use those other forms of social media, your iMessages, your WhatsApps, use those other forms. FaceTime is not social media, but it's like a positive way to use technology, right? Like connect with people. I think if it keeps going the way we're going right now, we're going to be using those forms of communication more and hopefully meeting in real life more often. So that's what I'm optimistic about. But it does rely on AI doing a lot more damage. So I don't know how optimistic that is. Well, that's optimistic enough for me because it's nice to have this feeling of, okay, we're going to move forward with AI. And maybe just maybe it'll make us more appreciative of human to human IRL communication, which I think is always wonderful to get back to. So Jeremy, thank you so much for being here. I appreciated this conversation and your expertise so, so much. So thank you so much. I appreciate you for being here. Thank you. Wow, what an incredible conversation. I hope you enjoyed it as much as I did, especially because I spend so much time on my phone getting through all of these videos that I play for the show. And I know that I get into the AI slot videos sometimes. There's good and bad out there, right? And you really can't do a show like this without covering AI because it's everywhere. So the better stewards we can be and the better consumers we can be of this content and the more informed we can be the better. So I hope you enjoyed this as much as I did. And I will see you right back here next time. This is a show with an obsessive focus on what's good for America. You are going to love Vince.