Trending Now
We have updated our Privacy Policy and Terms of Use for Eurasia Group and its affiliates, including GZERO Media, to clarify the types of data we collect, how we collect it, how we use data and with whom we share data. By using our website you consent to our Terms and Conditions and Privacy Policy, including the transfer of your personal data to the United States from your country of residence, and our use of cookies described in our Cookie Policy.
{{ subpage.title }}
Hard Numbers: Could Microsoft buy TikTok?, Get me the Operator, Meta and ByteDance spend on AI, ElevenLabs’ billions, Ready for “Humanity’s Last Exam”?
2020: Microsoft is in talks to acquire TikTok, according to President Donald Trump. If that rings a bell it’s because Microsoft sought to buy the social media app in 2020, the last time Trump tried to ban the app. The deal fell through, and Microsoft CEO Satya Nadella later called the attempted TikTok takeover the “strangest thing I've ever worked on.” This time around, all the company has said on the matter is that it “has nothing to share at this time.” Meanwhile, Trump has also nodded to there being “great interest in TikTok” from several companies.
200: OpenAI announced Operator, its AI “agent,” in an experimental “research preview,” on Thursday. The point is that this model can not only chat with you but can actually perform tasks for you, like booking a restaurant reservation or ordering food for delivery. It’s currently available to subscribers of ChatGPT Pro, a $200-a-month subscription.
65 billion: Meta said Friday it expects to spend up to $65 billion in 2025, up from $40 billion in 2024, to fuel its growing AI ambitions. Meanwhile, TikTok’s Chinese parent company ByteDance has reportedly earmarked $21 billion, including $12 billion on AI infrastructure.
3 billion: The AI voice-cloning company ElevenLabs has raised a new $250 million funding round announced Friday that values it at around $3 billion. We tried out ElevenLabs’ software last year to clone our author’s voice and translate it into different languages.
3,000: Researchers at the Center for AI Safety and Scale AI released “Humanity’s Last Exam” on Thursday, a 3,000-question multiple-choice and short-answer test designed to evaluate AI models’ capabilities. With AI models succeeding at most existing tests, the researchers strived to create one that will be able to stump most — or at least show when they’ve become truly superintelligent. For now, they’re struggling: All of the current top models fail the exam with OpenAI’s o1 model scoring the highest at 8.3%.Hello? Yes. This is Scott’s AI voice speaking.
There’s a voice on the internet that sounds just like me. In a way, it is me.
ElevenLabs, a startup software company, uses artificial intelligence to simulate natural-sounding language. If you need a realistic voiceover for a television commercial, short film, or an audiobook, ElevenLabs lets you use one of the human-ish voices in its library. While this technology isn’t ready to put voice actors out of business just yet — there are some definite hiccups — the technology is surprisingly effective.
In January, New Hampshire voters received a robocall from Joe Biden. Except it wasn’t Joe Biden. It was an AI-generated fake made by a political ally of then-Democratic challenger Dean Phillips. It was made with ElevenLabs.
In response, ElevenLabs banned the account responsible and in February changed its policy around impersonating public figures. While the company already banned impersonation of private figures without their consent, or anything intended to harm others, ElevenLabs added a list of “no-go” voices, specifically outlawing impersonation of presidential or prime minister candidates in the US and UK with the intention of adding additional “languages and election cycles.”
With elections in 64 countries this year, and OpenAI planning its own ElevenLabs-like tool too, the threat of mass confusion around elections feels palpable. AI has helped former Pakistani Prime Minister Imran Khandeliver a campaign address from prison, but it also caused problems for Indonesian political candidate Anies Baswedan, the victim of audio pretending to be a political backer chastising him. In addition to spreading disinformation, these tools could also be used to break voice authentication measures used by banks and other financial institutions.
Naturally, I had to give it a try.
Scott's AI-generated voice, courtesy of ElevenLabs.roar-assets-auto.rbl.ms
I paid $22 for the mid-tier version of ElevenLabs, which got me “professional voice cloning,” about two hours of downloadable AI-generated text-to-speech per month, and high-quality audio. To generate my voice clone, I had to upload 10-plus minutes of myself speaking, though it recommended 30 for the best results. After 12 minutes of gabbing into my USB microphone about the results of the NFL Draft, I hit upload.
The software took about two hours to do its magic before I was alerted that my cloned voice was ready to chat. Even just typing in, “Hey, this is Scott Nover,” I was amazed how much it sounded like me. I typed some more. The more text I included, the more the software struggled to sound human. There were unnatural pauses, odd diction, and my detached voice seemed upset about something.
Scott's AI-generated voice, speaking Mandarin, courtesy of ElevenLabs.roar-assets-auto.rbl.ms
ElevenLabs lets you choose from a few different models, and it offers sliders to adjust the voice’s tone, similarity to the uploaded recording, and sassiness. I tinkered with the settings until I found the best ratios and walked away pretty impressed with what I created. It even let me change the language — my polyglot editor Matthew Kendrick says the Spanish sounds good, with a mild Mexican accent and a few odd word choices. The Mandarin is a little less convincing — mostly because white dudes usually can’t hit their tones so accurately.
Scott's AI-generated voice, speaking Spanish, courtesy of ElevenLabs.roar-assets-auto.rbl.ms
If you’ve been following this newsletter, you’ll know we’ve been covering the rise of generative AI and have warned, in particular, about the rise of AI-generated audio.
Soon, AI voices will be everywhere. Some, like my homemade ScottBot, won't matter much. But some—say, a homemade clone of the Ayatollah Khamenei issuing a fatwa or Vladimir Putin inciting violence in Ukraine—could cause major problems on the world stage. And every government should be prepared.