Trending Now
We have updated our Privacy Policy and Terms of Use for Eurasia Group and its affiliates, including GZERO Media, to clarify the types of data we collect, how we collect it, how we use data and with whom we share data. By using our website you consent to our Terms and Conditions and Privacy Policy, including the transfer of your personal data to the United States from your country of residence, and our use of cookies described in our Cookie Policy.
{{ subpage.title }}
Social media's AI wave: Are we in for a “deepfakification” of the entire internet?
In this episode of GZERO AI, Taylor Owen, professor at the Max Bell School of Public Policy at McGill University and director of its Centre for Media, Technology & Democracy, looks into the phenomenon he terms the "deepfakification" of social media. He points out the evolution of our social feeds, which began as platforms primarily for sharing updates with friends, and are now inundated with content generated by artificial intelligence.
So 2024 might just end up being the year of the deepfake. Not some fake Joe Biden video or deepfake pornography of Taylor Swift. Definitely problems, definitely going to be a big thing this year. But what I would see is a bigger problem is what might be called the “deepfakification” of the entire internet and definitely of our social feeds.
Cory Doctorow has called this more broadly the “enshittification” of the internet. And I think the way AI is playing out in our social media is a very good example of this. So what we saw in our social media feeds has been an evolution. It began with information from our friends that they shared. It then merged the content that an algorithm thought we might want to see. It then became clickbait and content designed to target our emotions via these same algorithmic systems. But now, when many people open their Facebook or their Instagram or their talk feeds, what they're seeing is content that's been created by AI. AI Content is flooding Facebook and Instagram.
So what's going on here? Well, in part, these companies are doing what they've always been designed to do, to give us content optimized to keep our attention.
If this content happens to be created by an AI, it might even do that better. It might be designed in a way by the AI to keep our attention. And AI is proving a very useful tool for doing for this. But this has had some crazy consequences. It's led to the rise, for example, of AI influencers rather than real people selling us ideas or products. These are AIs. Companies like Prada and Calvin Klein have hired an AI influencer named Lil Miquela, who has over 2.5 million followers on TikTok. A model agency in Barcelona, created an AI model after having trouble dealing with the schedules and demands of primadonna human models. They say they didn't want to deal with people with egos, so they had their AI model do it for them.
And that AI model brings in as much as €10,000 a month for the agency. But I think this gets at a far bigger issue, and that's that it's increasingly difficult to tell if the things we're seeing are real or if they're fake. If you scroll from the comments of one of these AI influencers like Lil Miquela’s page, it's clear that a good chunk of her followers don't know she's an AI.
Now platforms are starting to deal with this a bit. TikTok requires users themselves to label AI content, and Meta is saying they'll flag AI-generated content, but for this to work, they need a way of signaling this effectively and reliably to us and users. And they just haven't done this. But here's the thing, we can make them do it. The Canadian government in their new Online Harms Act, for example, demands that platforms clearly identify AI or bot generated content. We can do this, but we have to make the platforms do it. And I don't think that can come a moment too soon.
- Why human beings are so easily fooled by AI, psychologist Steven Pinker explains ›
- The geopolitics of AI ›
- AI and Canada's proposed Online Harms Act ›
- AI at the tipping point: danger to information, promise for creativity ›
- Will Taylor Swift's AI deepfake problems prompt Congress to act? ›
- Deepfake porn targets high schoolers ›
Beware AI's negative impact on our world, warns former Google CEO Eric Schmidt
Does Big Tech really understand AI? Ian Bremmer talks to former Google CEO Eric Schmidt & co-author of “The Age of AI: And Our Human Future,” who believes we need to control AI before it controls us.
What's troubling about AI, he says, is that it’s like nothing we’ve seen before, it's still very new. Instead of being precise, AI learns by doing– exactly like humans.
The coronavirus pandemic drove people’s lives even more online– we are now more connected than ever before. But we don't always know who runs our digital world.
The problem is that instead of governments, tech companies are writing the rules through computer algorithms powered by artificial intelligence.
The US and China competition in AI is intensifying. China is already doing pretty scary stuff with it, like surveillance of Uyghurs in Xinjiang (and also some fun stuff, like publicly shaming jaywalkers). Schmidt explains that it's because the Chinese ensures their internet reflects the priorities of the Communist Party --- he’s not a big fan of those values shaping the AI on apps his children use. Yet, he blames algorithms, not China, for the polarization on social media. Schmidt is all for free speech, but not for robots.
- Podcast: We have to control AI before it controls us, warns former ... ›
- Ian Bremmer explains: Should we worry about AI? - GZERO Media ›
- The AI addiction cycle - GZERO Media ›
- Eric Schmidt: We're not ready for what AI may do to us - GZERO Media ›
- Podcast: Why Scott Galloway is “cautiously optimistic” about AI - but not TikTok or Meta - GZERO Media ›