was barely a term in 1956, when top scientists from the sector of computing arrived at Dartmouth College for a summer conference. The pc scientist John McCarthy had coined the phrase within the funding proposal for the event, a gathering to work through easy methods to construct machines that would use language, solve problems like humans, and improve themselves. However it was a great alternative, one which captured the organizers’ founding premise: Any feature of human intelligence could “in principle be so precisely described that a machine may be made to simulate it.”
Of their proposal, the group had listed several “facets of the bogus intelligence problem.” The last item on their list, and in hindsight perhaps essentially the most difficult, was constructing a machine that would exhibit creativity and originality.
On the time, psychologists were grappling with easy methods to define and measure creativity in humans. The prevailing theory—that creativity was a product of intelligence and high IQ—was fading, but psychologists weren’t sure what to switch it with. The Dartmouth organizers had one in every of their very own. “The difference between creative pondering and unimaginative competent pondering lies within the injection of some randomness,” they wrote, adding that such randomness “should be guided by intuition to be efficient.”
Nearly 70 years later, following quite a few boom-and-bust cycles in the sector, we now have AI models that kind of follow that recipe. While large language models that generate text have exploded within the last three years, a unique kind of AI, based on what are called diffusion models, is having an unprecedented impact on creative domains. By transforming random noise into coherent patterns, diffusion models can generate recent images, videos, or speech, guided by text prompts or other input data. The most effective ones can create outputs indistinguishable from the work of individuals, in addition to bizarre, surreal results that feel distinctly nonhuman.
Now these models are marching right into a creative field that’s arguably more vulnerable to disruption than every other: music. AI-generated creative works—from orchestra performances to heavy metal—are poised to suffuse our lives more thoroughly than every other product of AI has done yet. The songs are prone to mix into our streaming platforms, party and wedding playlists, soundtracks, and more, whether or not we notice who (or what) made them.
For years, diffusion models have stirred debate within the visual-art world about whether what they produce reflects true creation or mere replication. Now this debate has come for music, an art form that’s deeply embedded in our experiences, memories, and social lives. Music models can now create songs able to eliciting real emotional responses, presenting a stark example of how difficult it’s becoming to define authorship and originality within the age of AI.
The courts are actively grappling with this murky territory. Major record labels are suing the highest AI music generators, alleging that diffusion models do little greater than replicate human art without compensation to artists. The model makers counter that their tools are made to help in human creation.
In deciding who is correct, we’re forced to think hard about our own human creativity. Is creativity, whether in artificial neural networks or biological ones, merely the results of vast statistical learning and drawn connections, with a sprinkling of randomness? In that case, then authorship is a slippery concept. If not—if there’s some distinctly human element to creativity—what’s it? What does it mean to be moved by something with out a human creator? I needed to wrestle with these questions the primary time I heard an AI-generated song that was genuinely incredible—it was unsettling to know that somebody merely wrote a prompt and clicked “Generate.” That predicament is coming soon for you, too.
Making connections
After the Dartmouth conference, its participants went off in several research directions to create the foundational technologies of AI. At the identical time, cognitive scientists were following a 1950 call from J.P. Guilford, president of the American Psychological Association, to tackle the query of creativity in human beings. They got here to a definition, first formalized in 1953 by the psychologist Morris Stein within the : Creative works are each novel, meaning they present something recent, and useful, meaning they serve some purpose to someone. Some have called for “useful” to get replaced by “satisfying,” and others have pushed for a 3rd criterion: that creative things are also surprising.
Later, within the Nineteen Nineties, the rise of functional magnetic resonance imaging made it possible to review more of the neural mechanisms underlying creativity in lots of fields, including music. Computational methods prior to now few years have also made it easier to map out the role that memory and associative pondering play in creative decisions.
What has emerged is less a grand unified theory of how a creative idea originates and unfolds within the brain and more an ever-growing list of powerful observations. We will first divide the human creative process into phases, including an ideation or proposal step, followed by a more critical and evaluative step that appears for merit in ideas. A number one theory on what guides these two phases known as the associative theory of creativity, which posits that essentially the most creative people can form novel connections between distant concepts.
“It may very well be like spreading activation,” says Roger Beaty, a researcher who leads the Cognitive Neuroscience of Creativity Laboratory at Penn State. “You’re thinking that of 1 thing; it just type of prompts related concepts to whatever that one concept is.”
These connections often hinge specifically on semantic memory, which stores concepts and facts, versus episodic memory, which stores memories from a specific time and place. Recently, more sophisticated computational models have been used to review how people make connections between concepts across great “semantic distances.” For instance, the word is more closely related to than to . Studies have shown that highly creative people may perceive very semantically distinct concepts as close together. Artists have been found to generate word associations across greater distances than non-artists. Other research has supported the concept that creative people have “leaky” attention—that’s, they often notice information which may not be particularly relevant to their immediate task.
Neuroscientific methods for evaluating these processes don’t suggest that creativity unfolds in a specific area of the brain. “Nothing within the brain produces creativity like a gland secretes a hormone,” Dean Keith Simonton, a frontrunner in creativity research, wrote within the.
The evidence as a substitute points to a number of dispersed networks of activity during creative thought, Beaty says—one to support the initial generation of ideas through associative pondering, one other involved in identifying promising ideas, and one other for evaluation and modification. A brand new study, led by researchers at Harvard Medical School and published in February, suggests that creativity might even involve the of particular brain networks, like ones involved in self-censorship.
Up to now, machine creativity—in case you can call it that—looks quite different. Though on the time of the Dartmouth conference AI researchers were fascinated with machines inspired by human brains, that focus had shifted by the point diffusion models were invented, a couple of decade ago.
The most effective clue to how they work is within the name. Should you dip a paintbrush loaded with red ink right into a glass jar of water, the ink will diffuse and swirl into the water seemingly at random, eventually yielding a pale pink liquid. Diffusion models simulate this process in reverse, reconstructing legible forms from randomness.
For a way of how this works for images, picture a photograph of an elephant. To coach the model, you make a replica of the photo, adding a layer of random black-and-white static on top. Make a second copy and add a bit more, and so forth lots of of times until the last image is pure static, with no elephant in sight. For every image in between, a statistical model predicts how much of the image is noise and the way much is admittedly the elephant. It compares its guesses with the suitable answers and learns from its mistakes. Over thousands and thousands of those examples, the model gets higher at “de-noising” the photographs and connecting these patterns to descriptions like “male Borneo elephant in an open field.”
Now that it’s been trained, generating a brand new image means reversing this process. Should you give the model a prompt, like “a completely happy orangutan in a mossy forest,” it generates a picture of random white noise and works backward, using its statistical model to remove bits of noise step-by-step. At first, rough shapes and colours appear. Details come after, and eventually (if it really works) an orangutan emerges, all without the model “knowing” what an orangutan is.
Musical images
The approach works much the identical way for music. A diffusion model doesn’t “compose” a song the way in which a band might, starting with piano chords and adding vocals and drums. As an alternative, all the weather are generated directly. The method hinges on the proven fact that the numerous complexities of a song may be depicted visually in a single waveform, representing the amplitude of a sound wave plotted against time.
Consider a record player. By traveling along a groove in a bit of vinyl, a needle mirrors the trail of the sound waves engraved in the fabric and transmits it right into a signal for the speaker. The speaker simply pushes out air in these patterns, generating sound waves that convey the entire song.
From a distance, a waveform might look as if it just follows a song’s volume. But in case you were to zoom in closely enough, you can see patterns within the spikes and valleys, just like the 49 waves per second for a bass guitar playing a low G. A waveform accommodates the summation of the frequencies of all different instruments and textures. “You see certain shapes start happening,” says David Ding, cofounder of the AI music company Udio, “and that type of corresponds to the broad melodic sense.”
Since waveforms, or similar charts called spectrograms, may be treated like images, you may create a diffusion model out of them. A model is fed thousands and thousands of clips of existing songs, each labeled with an outline. To generate a brand new song, it starts with pure random noise and works backward to create a brand new waveform. The trail it takes to achieve this is formed by what words someone puts into the prompt.
Ding worked at Google DeepMind for five years as a senior research engineer on diffusion models for images and videos, but he left to found Udio, based in Recent York, in 2023. The corporate and its competitor Suno, based in Cambridge, Massachusetts, at the moment are leading the race for music generation models. Each aim to construct AI tools that enable nonmusicians to make music. Suno is larger, claiming greater than 12 million users, and raised a $125 million funding round in May 2024. The corporate has partnered with artists including Timbaland. Udio raised a seed funding round of $10 million in April 2024 from distinguished investors like Andreessen Horowitz in addition to musicians Will.i.am and Common.
The outcomes of Udio and Suno to date suggest there’s a large audience of people that may not care whether the music they hearken to is made by humans or machines. Suno has artist pages for creators, some with large followings, who generate songs entirely with AI, often accompanied by AI-generated images of the artist. These creators will not be musicians in the standard sense but expert prompters, creating work that may’t be attributed to a single composer or singer. On this emerging space, our normal definitions of authorship—and our lines between creation and replication—all but dissolve.
The outcomes of Udio and Suno to date suggest there’s a large audience of people that may not care whether the music they hearken to is made by humans or machines.
The music industry is pushing back. Each corporations were sued by major record labels in June 2024, and the lawsuits are ongoing. The labels, including Universal and Sony, allege that the AI models have been trained on copyrighted music “at an almost unimaginable scale” and generate songs that “imitate the qualities of real human sound recordings” (the case against Suno cites one ABBA-adjacent song called “Prancing Queen,” for instance).
Suno didn’t reply to requests for comment on the litigation, but in a press release responding to the case posted on Suno’s blog in August, CEO Mikey Shulman said the corporate trains on music found on the open web, which “indeed accommodates copyrighted materials.” But, he argued, “learning will not be infringing.”
A representative from Udio said the corporate wouldn’t comment on pending litigation. On the time of the lawsuit, Udio released a press release mentioning that its model has filters to make sure that it “doesn’t reproduce copyrighted works or artists’ voices.”
Complicating matters even further is guidance from the US Copyright Office, released in January, that claims AI-generated works may be copyrighted in the event that they involve a substantial amount of human input. A month later, an artist in Recent York received what could be the primary copyright for a bit of visual art made with the assistance of AI. The primary song may very well be next.
Novelty and mimicry
These legal cases wade right into a gray area much like one explored by other court battles unfolding in AI. At issue here is whether or not training AI models on copyrighted content is allowed, and whether generated songs unfairly copy a human artist’s style.
But AI music is prone to proliferate in some form no matter these court decisions; YouTube has reportedly been in talks with major labels to license their music for AI training, and Meta’s recent expansion of its agreements with Universal Music Group suggests that licensing for AI-generated music could be on the table.
If AI music is here to remain, will any of or not it’s any good? Consider three aspects: the training data, the diffusion model itself, and the prompting. The model can only be nearly as good because the library of music it learns from and the descriptions of that music, which should be complex to capture it well. A model’s architecture then determines how well it might use what’s been learned to generate songs. And the prompt you feed into the model—in addition to the extent to which the model “understands” what you mean by “turn down that saxophone,” for instance—is pivotal too.
Is the result creation or just replication of the training data? We could ask the identical query about human creativity.
Arguably an important issue is the primary: How extensive and diverse is the training data, and the way well is it labeled? Neither Suno nor Udio has disclosed what music has gone into its training set, though these details will likely need to be disclosed in the course of the lawsuits.
Udio says the way in which those songs are labeled is crucial to the model. “An area of energetic research for us is: How can we get increasingly refined descriptions of music?” Ding says. A basic description would discover the genre, but then you can also say whether a song is moody, uplifting, or calm. More technical descriptions might mention a two-five-one chord progression or a selected scale. Udio says it does this through a mixture of machine and human labeling.
“Since we wish to focus on a broad range of goal users, that also implies that we want a broad range of music annotators,” he says. “Not only individuals with music PhDs who can describe the music on a really technical level, but additionally music enthusiasts who’ve their very own informal vocabulary for describing music.”
Competitive AI music generators must also learn from a relentless supply of recent songs made by people, or else their outputs can be stuck in time, sounding stale and dated. For this, today’s AI-generated music relies on human-generated art. In the longer term, though, AI music models may train on their very own outputs, an approach being experimented with in other AI domains.
Because models start with a random sampling of noise, they’re nondeterministic; giving the identical AI model the identical prompt will lead to a brand new song every time. That’s also because many manufacturers of diffusion models, including Udio, inject additional randomness through the method—essentially taking the waveform generated at each step and distorting it ever so barely in hopes of adding imperfections that serve to make the output more interesting or real. The organizers of the Dartmouth conference themselves really helpful such a tactic back in 1956.
In accordance with Udio cofounder and chief operating officer Andrew Sanchez, it’s this randomness inherent in generative AI programs that comes as a shock to many individuals. For the past 70 years, computers have executed deterministic programs: Give the software an input and receive the identical response each time.
“Lots of our artists partners can be like, ‘Well, why does it do that?’” he says. “We’re like, well, we don’t really know.” The generative era requires a brand new mindset, even for the businesses creating it: that AI programs may be messy and inscrutable.
Is the result creation or just replication of the training data? Fans of AI music told me we could ask the identical query about human creativity. As we hearken to music through our youth, neural mechanisms for learning are weighted by these inputs, and memories of those songs influence our creative outputs. In a recent study, Anthony Brandt, a composer and professor of music at Rice University, identified that each humans and enormous language models use past experiences to guage possible future scenarios and make higher selections.
Indeed, much of human art, especially in music, is borrowed. This often leads to litigation, with artists alleging that a song was copied or sampled without permission. Some artists suggest that diffusion models must be made more transparent, so we could know that a given song’s inspiration is three parts David Bowie and one part Lou Reed. Udio says there’s ongoing research to realize this, but at once, nobody can do it reliably.
For nice artists, “there’s that combination of novelty and influence that’s at play,” Sanchez says. “And I believe that that’s something that can be at play in these technologies.”
But there are a lot of areas where attempts to equate human neural networks with artificial ones quickly crumble under scrutiny. Brandt carves out one domain where he sees human creativity clearly soar above its machine-made counterparts: what he calls “amplifying the anomaly.” AI models operate within the realm of statistical sampling. They don’t work by emphasizing the exceptional but, quite, by reducing errors and finding probable patterns. Humans, then again, are intrigued by quirks. “Moderately than being treated as oddball events or ‘one-offs,’” Brandt writes, the quirk “permeates the creative product.”

He cites Beethoven’s decision so as to add a jarring off-key note within the last movement of his Symphony no. 8. “Beethoven could have left it at that,” Brandt says. “But quite than treating it as a one-off, Beethoven continues to reference this incongruous event in various ways. In doing so, the composer takes a momentary aberration and magnifies its impact.” One could look to similar anomalies within the backward loop sampling of late Beatles recordings, pitched-up vocals from Frank Ocean, or the incorporation of “found sounds,” like recordings of a crosswalk signal or a door closing, favored by artists like Charlie Puth and by Billie Eilish’s producer Finneas O’Connell.
If a creative output is indeed defined as one which’s each novel and useful, Brandt’s interpretation suggests that the machines can have us matched on the second criterion while humans reign supreme on the primary.
To explore whether that’s true, I spent a number of days fooling around with Udio’s model. It takes a minute or two to generate a 30-second sample, but when you have got paid versions of the model you may generate whole songs. I made a decision to select 12 genres, generate a song sample for every, after which find similar songs made by people. I built a quiz to see if people in our newsroom could spot which songs were made by AI.
The typical rating was 46%. And for a number of genres, especially instrumental ones, listeners were incorrect as a rule. After I watched people do the test in front of me, I noticed that the qualities they confidently flagged as an indication of composition by AI—a fake-sounding instrument, a weird lyric—rarely proved them right. Predictably, people did worse in genres they were less acquainted with; some did okay on country or soul, but many stood no likelihood against jazz, classical piano, or pop. Beaty, the creativity researcher, scored 66%, while Brandt, the composer, finished at 50% (though he answered accurately on the orchestral and piano sonata tests).
Keep in mind that the model doesn’t deserve all of the credit here; these outputs couldn’t have been created without the work of human artists whose work was within the training data. But with just a number of prompts, the model generated songs that few people would pick as machine-made. A number of could easily have been played at a celebration without raising objections, and I discovered two I genuinely loved, whilst a lifelong musician and usually picky music person. But sounding real will not be the identical thing as sounding original. The songs didn’t feel driven by oddities or anomalies—actually not on the extent of Beethoven’s “jump scare.” Nor did they appear to bend genres or cover great leaps between themes. In my test, people sometimes struggled to choose whether a song was AI-generated or just bad.
How much will this matter in the long run? The courts will play a job in deciding whether AI music models serve up replications or recent creations—and the way artists are compensated in the method—but we, as listeners, will resolve their cultural value. To understand a song, do we want to picture a human artist behind it—someone with experience, ambitions, opinions? Is an important song not great if we discover out it’s the product of AI?
Sanchez says people may wonder who’s behind the music. But “at the tip of the day, nevertheless much AI component, nevertheless much human component, it’s going to be art,” he says. “And persons are going to react to it on the standard of its aesthetic merits.”
In my experiment, though, I saw that the query really mattered to people—and a few vehemently resisted the concept of having fun with music made by a pc model. When one in every of my test subjects instinctively began bobbing her head to an electro-pop song on the quiz, her face expressed doubt. It was almost as if she was trying her best to picture a human quite than a machine because the song’s composer. “Man,” she said, “I actually hope this isn’t AI.”
It was.