We polled 2,000 people about how they’re using AI, what they want it to do, and what scares them about it the most.
AI is about to change the world — the problem is, no one’s quite sure how. Some look at the past year’s rapid progress and see opportunities to remove creative constraints, automate rote work, and discover new ways to learn and teach. Others see how this tech can disrupt our lives in more damaging ways: how it can generate misinformation, destroy or diminish jobs, and, if left unchecked, pose a serious threat to our safety.
Tech leaders, lawmakers, and researchers have all been weighing in on how we should handle this emerging tech. Some industry figures, like OpenAI CEO Sam Altman, want AI giants to steer regulation, shifting the focus to perceived future threats, including the “risk of extinction.” Others, like EU politicians, are more concerned with current dangers and banning dangerous use cases (while holding back positive applications, say skeptics). Meanwhile, many small artists would just like a guarantee that they won’t be replaced by machines.
To find out what people really think about AI and what they want from it, The Verge teamed up with Vox Media’s Insights and Research team and the research consultancy firm The Circus to poll more than 2,000 US adults on their thoughts, feelings, and fears about AI. The results tell the story of an emerging, uncertain, and exciting technology — where many have yet to use it, many are fearful of its potential, and many still have great hopes for what it could someday do for them.
Who’s using AI?
Who’s using AI?
AI is suddenly everywhere. Image generators and large language models are at the core of new startups, powering features inside our favorite apps, and — perhaps more importantly — driving conversation not just in the tech world but also society at large. Concerns abound about cheating in schools with ChatGPT, being fooled by AI-generated pictures, and artists being ripped off or even outright replaced.
But despite widespread news coverage, use of these new tools is still fairly limited, at least when it comes to dedicated AI products. And experience with these tools skews decidedly toward younger users.
Most people have heard of ChatGPT. Bing and Bard? Not quite.
Tool | I’ve used it | I’ve heard of it | I’ve never heard of it |
---|---|---|---|
ChatGPT | 20% | 37% | 43% |
Bing w/ ChatGPT | 12% | 34% | 54% |
My AI (Snapchat) | 12% | 33% | 55% |
Bard (by Google) | 10% | 28% | 62% |
Midjourney | 7% | 18% | 75% |
Stable Diffusion | 6% | 17% | 77% |
Only 1 in 3 people have tried one of these AI-powered tools, and most aren’t familiar with the companies and startups that make them. Despite the many insurgents in the world of AI, like Stability AI and Midjourney, it’s still the work of Big Tech that substantially steers the conversation. OpenAI is the major exception — but arguably, thanks to its market cap and deals with Microsoft, it is itself now a member of the corpo-club.
AI use is dominated by Millennials and Gen Z
Boomers: 4.8 million | Gen X: 15.8 million | Millenial: 36 million | Gen Z: 34.9 million |
One complicating factor, though, is that the definition of an AI tool is extremely fuzzy. We asked respondents about dedicated AI services like ChatGPT or Midjourney. But many companies are adding AI features to established software, whether that’s image generation in Photoshop or text suggestion in Gmail and Google Docs. And as the joke goes, AI is whatever computers haven’t done yet, meaning yesterday’s AI is, simply, today’s expected features.
Despite the limited usage of these tools so far, people have high expectations for AI’s impact on the world — beyond those of other emergent (and sometimes controversial) technologies. Nearly three-quarters of people said AI will have a large or moderate impact on society. That’s compared to 69 percent for electric vehicles and a paltry 34 percent for NFTs. They’re so 2021.
Will these technologies have a big impact on society?
AI | Electric vehicles | Virtual Reality | AR | NFTs | |
---|---|---|---|---|---|
Large/Moderate Impact | 74% | 69% | 60% | 52% | 34% |
How is AI being used?
How is AI being used?
The main fuel for the recent boom is generative AI: systems that can generate text, help brainstorm ideas, edit your writing, and create pictures, audio, and video. These tools are being quickly integrated into professional systems — Photoshop can reimagine parts of images, WordPress can write blog posts — but for most users, they generally require quite a bit of oversight to get it right.
Search, brainstorming, and art dominate current AI use
Category | Usage |
---|---|
Photos | 37% |
Music | 31% |
Videos | 31% |
Designs | 29% |
Stories | 29% |
Artwork | 27% |
Emails | 25% |
Answer a question | 68% |
Blog posts | 19% |
Coding | 18% |
Brainstorming | 54% |
Data analysis | 26% |
Lyrics | 18% |
For those who are using AI tools, creative experiments were most common. People are generating music and videos, creating stories, and tinkering with photos. More professional applications like coding were less common. And above all, people have simply been using AI systems to answer questions — suggesting chatbots like ChatGPT, Bing, and Bard may replace search engines, for better or worse.
One finding is particularly clear: AI is expanding what people can create. In every category we polled, people who used AI said they used these systems to make something they couldn’t otherwise, with artwork being the most popular category within these creative fields. This makes sense given that AI image generators are much more advanced than tools that create audio or video.
Most people thought AI did a better job than they could have
It was better than what I could have produced myself | It wasn’t as good as what I could have produced myself |
---|---|
61% | 11% |
60% | 5% |
59% | 8% |
57% | 8% |
56% | 10% |
55% | 4% |
52% | 9% |
45% | 11% |
The concerns around AI art
The concerns around AI art
AI image generators like Midjourney and Stable Diffusion provide a good case study for broader issues involving generative AI. These systems are trained on huge amounts of data scraped from the web, usually done without the consent of the original creators. While there’s heated debate about the ethics, the legality of this practice is currently being questioned in numerous lawsuits. These arguments are quickly spreading to other generative mediums, like AI song generation.
Our survey shows that people have mixed feelings about how to respond to these ethical quandaries. Most people think artists should get compensated when an AI tool clones their style, for example, but a majority also don’t want these capabilities to be limited. Indeed, almost half of respondents said they’d tested the system by generating exactly this sort of output.
A lot of people have copied an artist — and a lot of people want to ban it, too
Question | Agree |
---|---|
I think artists should get compensated when the AI creates content that is explicitly designed to resemble that artist’s work | 70% |
Companies should prevent the AI from creating work that is derivative of other work (e.g. artists, writers, musicians) | 43% |
I have instructed the AI to create something in either the style or voice of a writer, artist, or well known figure | 44% |
People want better standards for AI
People want better standards for AI
It’s not just tech leaders who are looking for AI tools to be controlled. More than three-quarters of respondents agreed with the statement, “Regulations and laws need to be developed regarding the development of AI.” These laws are currently in the works, with the EU AI Act working its way through final negotiations and the US recently holding hearings to develop its own legal framework.
Clearly, there’s strong demand for higher standards in AI systems and disclosure of their use. Strong majorities are in favor of labeling AI-generated deepfakes, for example. But many principles with wide support would be difficult to enforce, including training AI language models on fact-checked data and banning deepfakes of people made without their consent.
There’s broad support for regulations on AI
Question | Agree |
---|---|
Regulations and laws need to be developed regarding the development of AI | 76% |
AI models should be required to be trained on datasets that have been fact checked | 76% |
I think digital content that is created with AI should be required to clearly state that it was created with AI | 78% |
I think it should be illegal to create video & audio deep fakes that imitate a real person without that person giving their explicit permission | 76% |
AI futures: excited, worried, and both at once
AI futures: excited, worried, and both at once
The animal spirits surrounding AI are decidedly ambivalent but lean, just slightly, toward pessimism.
When trying to predict the effect of AI on society, people forecast all sorts of dangers, from job losses (63 percent) to privacy threats (68 percent) and government and corporate misuse (67 percent). These dangers are weighted more heavily than potential positive applications, like new medical treatments (51 percent) and economic empowerment (51 percent). And when asked how they feel about the potential impact on their personal and professional life and on society more generally, people are pretty evenly split between worried and excited. Most often, they’re both.
How do people feel about AI’s impact on society?
When you think about AI’s potential impact on society, does it make you feel more anxious or excited? | Total |
---|---|
I’m more excited than I am anxious | 21% |
I’m equally excited as I am anxious | 32% |
I’m more anxious than I am excited | 29% |
I’m neither anxious nor excited | 18% |
Quite a surprising number of people also consider the more adventurous predictions for AI to be reasonable outcomes. Fifty-six percent of respondents think “people will develop emotional relationships with AI,” and 35 percent of people said they’d be open to doing so if they were lonely.
Many people in the AI world are currently warning about the “existential risk” posed by AI systems — the hotly contested idea that superintelligent AI could doom humanity. If they want to speak to the population, they’ll find more than a few in agreement, with 38 percent agreeing with the statement that AI will wipe out human civilization. Perhaps that’s why more people are worried than not.
…and what about AI’s impact on personal life and jobs?
When you think about AI’s potential impact on your personal and professional life, are you more worried or excited? | Total |
---|---|
I’m more excited than I am anxious | 20% |
I’m equally excited as I am anxious | 31% |
I’m more anxious than I am excited | 31% |
I’m neither anxious nor excited | 18% |
Appropriate to the uncertainty around AI, there’s an expansive openness to the idea of what could happen next. Roughly half of US adults expect that a sentient AI will emerge at some point in the future — and close to two-thirds don’t have an issue with companies trying to make one. If you already feel like there are too many concerns to grapple with when it comes to AI, you’d better hope we don’t get there.
If companies want to make AI sentient, most people won’t stand in the way
Question | Agree |
---|---|
Companies should not be allowed to produce an AI that is ‘conscious’ or ‘sentient’ | 36% |
AI will become sentient at some point | 51% |
Methodology: The Vox Media Insights and Research team surveyed over 2,000 consumers at the beginning of 2023 with our partner, The Circus: an insight and data storytelling consultancy that specializes in original trend research, thought leadership, and strategic brand positioning rooted in a human-data centric approach.
Credits
- Study author: Edwin Wong
- Creative director: Will Joel
- Engineer: Graham MacAree
- Editorial operations manager: Kara Verlaney
- Copy editor: Liz Hickson
- Special thanks: Sebastian Fernandez
- Editor-in-chief: Nilay Patel
- VP and GM: Andrew Melnizek
- Publisher: Helen Havlak