What is generative AI, and why is it suddenly everywhere?


Jakub Porzycki/NurPhoto via Getty Images

Between ChatGPT and Stable Diffusion, AI suddenly feels mainstream.

Artificial intelligence is suddenly everywhere — or at least, that’s what it seems like to me: A few weeks ago, a friend mentioned in passing that his law professor had warned students not to cheat with AI on an upcoming exam. At the same time, I couldn’t escape the uncanny portraits people were generating with the image-editing app Lensa AI’s new Magic Avatar feature and then sharing on social media. A guy on Twitter even used OpenAI’s new machine learning-powered chatbot, ChatGPT, to imitate what I said on a recent podcast (which, coincidentally, was also about ChatGPT) and posted it online.

Welcome to the age of generative AI, when it’s now possible for anyone to create new, original illustrations and text by simply sending a few instructions to a computer program. Several generative AI models, including ChatGPT and an image generator called Stable Diffusion, can now be accessed online for free or for a low-cost subscription, which means people across the world can do everything from assemble a children’s book to produce computer code in just a few clicks. This tech is impressive, and it can get pretty close to writing and illustrating how a human might. Don’t believe me? Here’s a Magic School Bus short story ChatGPT wrote about Ms. Frizzle’s class trip to the Fyre Festival. And below is an illustration I asked Stable Diffusion to create about a family celebrating Hanukkah on the moon.


Stable Diffusion
Stable Diffusion’s take on a lunar Hanukkah includes a menorah with five candles and plenty of oversized Christmas ornaments.

Generative AI’s results aren’t always perfect, and we’re certainly not dealing with an all-powerful, super AI — at least for now. Sometimes its creations are flawed, inappropriate, or don’t totally make sense. If you were going to celebrate Hanukkah on the moon, after all, you probably wouldn’t depict giant Christmas ornaments strewn across the lunar surface. And you might find the original Magic School Bus stories more entertaining than my AI-generated one.

Still, even in its current form and with its current limitations, generative AI could automate some tasks humans do daily — like writing form emails or drafting simple legal contracts — and possibly make some kinds of jobs obsolete. This technology presents plenty of opportunities, but plenty of complex new challenges, too. Writing emails may suddenly have gotten a lot easier, for example, but catching cheating students has definitely gotten a lot harder.

It’s only the beginning of this tech, so it can be hard to make sense of what exactly it is capable of or how it could impact our lives. So we tried to answer a few of the biggest questions surrounding generative AI right now.

Wait, how does this AI work?

Very simply, a generative AI system is designed to produce something new based on its previous experience. Usually, this technology is developed with a technique called machine learning, which involves teaching an artificial intelligence to perform tasks by exposing it to lots and lots of data, which it “trains” on and eventually learns to mimic. ChatGPT, for example, was trained on an enormous quantity of text available on the internet, along with scripts of dialogue, so that it could imitate human conversations. Stable Diffusion is an image generator created by the startup Stability.AI that will produce an image for you based on text instructions, and was designed by feeding the AI images and their associated captions collected from the web, which allowed the AI to learn what it should “illustrate” based on the verbal commands it received.

While the particular approaches used to build generative AI models can differ, this technology is ultimately trying to reproduce human behavior, creating new content based on the content that humans have already created. In some ways, it’s like the smart compose features you see on your iPhone when you’re texting or your Gmail account when you’re typing out an email. “It learns to detect patterns in this content, which in turn allows it to generate similar but distinct content,” explains Vincent Conitzer, a computer science professor at Carnegie Mellon.

This method of building AI can be extremely powerful, but it also has real flaws. In one test, for example, an AI model called Galactica that Meta built to help write scientific papers suggested that the Soviet Union was the first country to put a bear in space, among several other errors and falsehoods. (The company pulled the system offline in November, after just a few days.) Lensa AI’s Magic Avatar feature, the AI portrait generator, sometimes illustrates people with additional limbs. It also has the concerning tendency to depict women without any clothing.

It’s easy to find other biases and stereotypes built into this technology, too. When the Intercept asked ChatGPT to come up with an airline passenger screening system, the AI suggested higher risk scores for people from — or who had visited — Syria and Afghanistan, among other countries. Stable Diffusion also reproduces racial and gender stereotypes, like only depicting firefighters as white men. These are not particularly new problems with this kind of AI, as Abeba Birhane and Deborah Raji recently wrote in Wired. “People get hurt from the very practical ways such models fall short in deployment, and these failures are the result of their builders’ choices — decisions we must hold them accountable for,” they wrote.

Who is creating this AI, and why?

Generative AI isn’t free out of the goodness of tech companies’ hearts. These systems are free because the companies building them want to improve their models and technology, and people playing around with trial versions of the software give these companies, in turn, even more training data. Operating the computing systems to build artificial intelligence models can be extremely expensive, and while companies aren’t always upfront about their own expenses, costs can stretch into the tens of millions of dollars. AI developers want to eventually sell and license their technology for a profit.

There are already hints about what this new generative AI industry could look like. OpenAI, which developed the DALL-E and ChatGPT systems, operates under a capped-profit model, and plans to receive $1 billion in revenue by 2024, primarily through selling access to its tech (outside developers can already pay to use some of OpenAI’s tech in their apps). Microsoft has already started to use the system to assist with some aspects of computer programming in its code development app. Stability AI, the Stable Diffusion creator, wants to build specialized versions of the technology that it could sell to individual companies. The startup raised more than $100 million this past October.

Some think ChatGPT could ultimately replace Google’s search engine, which powers one of the biggest digital ad businesses in the world. ChatGPT is also pretty good at some basic aspects of coding, and technologies like it could eventually lower the overall costs of developing software. At the same time, OpenAI already has a pricing program available for DALL-E, and it’s easy to imagine how the system could be turned into a way of generating advertisements, visuals, and other graphics at a relatively low cost.

Is this the end of homework?

AI tools are already being used for one obvious thing: schoolwork, especially essays and online exams. These AI-produced assignments wouldn’t necessarily earn an A, but teachers seem to agree that ChatGPT can create at least B-worthy work. While tools for detecting whether a piece of text is AI generated are emerging, the popular plagiarism detection software, Turnitin, won’t catch this kind of cheating.

The arrival of this tech has driven some to declare the end of high school English, and even homework itself. While those predictions are hyperbolic, it’s certainly possible that homework will need to adapt. Some teachers may reverse course on the use of technology in the classroom and return to in-person, paper-based exams. Other instructors might turn to lockdown browsers, which would prevent people from visiting websites during a computer-based test. The use of AI itself may become part of the assignment, which is an idea some teachers are already exploring.

“The sorts of professionals our students want to be when they graduate already use these tools,” Phillip Dawson, the associate director of the Centre for Research in Assessment and Digital Learning, told Recode in December. “We can’t ban them, nor should we.”

Is AI going to take my job?

It’s hard to predict which jobs will or won’t be eradicated by generative AI. Greg Brockman, one of OpenAI’s co-founders, said in a December tweet that ChatGPT is “not yet ready to be relied on for anything important.” Still, this technology can already do all sorts of things that companies currently need humans to do. Even if this tech doesn’t take over your entire job, it might very well change it.

Take journalism: ChatGPT can already write a pretty compelling blog post. No, the post might not be particularly accurate — which is why there’s concern that ChatGPT could be quickly exploited to produce fake news — but it can certainly get the ball rolling, coming up with basic ideas for an article and even drafting letters to sources. The same bot can also earn a good score on a college-level coding exam, and it’s not bad at writing about legal concepts, either. A photo editor at New York magazine pointed out that while DALL-E doesn’t quite understand how to make illustrations dealing with complex political or conceptual concepts, it can be helpful when given repeated prodding and explicit instructions.

While there are limits on what ChatGPT could be used for, even automating just a few tasks in someone’s workflow, like writing basic code or copy editing, could radically change a person’s workday and reduce the total number of workers needed in a given field. As an example, Conitzer, the computer science professor, pointed to the impact of services like Google Flights on travel agencies.

“Online travel sites, even today, do not offer the full services of a human travel agent, which is why human travel agents are still around, in larger numbers than many people expect,” he told Recode. “That said, clearly their numbers have gone down significantly because the alternative process of just booking flights and a place to stay yourself online — a process that didn’t exist some decades ago — is a fine alternative in many cases.”

Should I be worried?

Generative AI is going mainstream rapidly, and companies aim to sell this technology as soon as possible. At the same time, the regulators who might try to rein in this tech, if they find a compelling reason, are still learning how it works.

The stakes are high. Like other breakthrough technologies — things like the computer and the smartphone, but also earlier inventions, like the air conditioner and the car — generative AI could change much of how our world operates. And like other revolutionary tech, the arrival of this kind of AI will create complicated trade-offs. Air conditioners, for example, have made some of the hottest days of the year more bearable, but they’re also exacerbating the world’s climate change problem. Cars made it possible to travel extremely long distances without the need for a train or horse-drawn carriage, but motor vehicle crashes now kill tens of thousands of people, at least in the United States, every year.

In the same way, decisions we make about AI now could have ripple effects. Legal cases about who deserves the profit and credit — but also the liability — for work created by AI are being decided now, but could shape who profits from this technology for years to come. Schools and teachers will determine whether to incorporate AI into their curriculums, or discard it as a form of cheating, inevitably influencing how kids will relate to these technologies in their professional lives. The rapid expansion of AI image generators could center Eurocentric art forms at the expense of other artistic traditions, which are already underrepresented by the technology.

If and when this AI goes fully mainstream, it could be incredibly difficult to unravel. In this way, the biggest threat of this technology may be that it stands to change the world before we’ve had a chance to truly understand it.

This story was first published in the Recode newsletter. Sign up here so you don’t miss the next one!

Related articles

Inspecting claims Campbell’s Soup uses ‘bioengineered meat’ and ‘3-D printed chicken’

Campbell's called the claims about its ingredients "false" and "absurd."

‘We have started to see cracks’: Dem senator spills about GOP’s Hegseth ‘nervousness’



A Democratic lawmaker said Thursday that Republican lawmakers have begun to separate themselves from President Donald Trump.

Sen. Jeff Merkley (D-OR) told CNN anchors Wolf Blitzer and Pamela Brown that Republicans have voiced their concerns over the president's recent moves and have questions about Defense Secretary Pete Hegseth's most recent comments on the Sept. 2 strike on an alleged drug boat in the Caribbean, off the coast of Venezuela.

Merkley, who serves on the Foreign Relations Committee, argued that the administration's response to the killings is not a satisfactory response for him. He described what the lawmakers have learned about the second strike, where "two helpless men clinging to debris" were killed.

"If this was a legal action of war, which is still under dispute, then it would be a war crime," Merkley said. "If it was not, it was a murder. In either case."

The Democratic lawmaker said that the U.S. Coast Guard should have investigated this incident.

"Again, the right way to find out if there are drugs aboard a boat is you stop the boat, you board it, you investigate it, and in the process you learn if there are drugs, you learn about the strategies involved, which gives you information to help dismantle a broader operation," Merkley said. "Blowing a boat up, not even knowing much about what the boat is doing simply destroys that type of information. So it's not only extrajudicial, it is also stupid. And so this is this is vast concerns about judgment. And by the way, of course, this is all a prelude to the possible strikes on Venezuela itself."

Trump has signaled that the U.S. has planned to attack Venezuela in ground strikes, although those details have not yet been released publicly.

The recent revelations have prompted congressional leaders to request Admiral Frank M. “Mitch” Bradley brief lawmakers Thursday in Washington, D.C. It has also raised questions about whether GOP leaders are ready to face the president over the reports, among other lingering concerns.

“There has been such a sense, of my colleagues, that they are not ready to confront Trump over the mistakes of this administration but we have started to see cracks in that following the November election a month ago where they're starting to feel like they have hitched their wagon to a horse that is going to take them over a cliff and they better start separating themselves,” Merkley said.

Merkley said it will be interesting to see what Republicans say after the briefings Thursday and that he believes Hegseth should resign.

“My Republican colleagues in the Senate are getting very nervous about being tied — not just to Hegseth — but to the overall actions of the administration," he added.

Where the Bands Are: This Week in Live Music and Concert News

A Swingin’ Christmas with Louis Prima Jr. and The...