And it works pretty well. In January 2021, Supertone revealed its Singing Voice Synthesis technology.
The big party trick was to present Kim Kwang-seok, a Korean folk superstar who sold millions of records at home, singing a new brand new song. Pretty cool, considering that he died in 1996.
Using artificial intelligence, Supertone’s SVS tech “learned” about 100 songs by 20 different singers in order to develop a style. Then it learned 10 songs by Kim. Putting everything together, the AI was able to create something that was more than a reasonable facsimile.
Why would HYBE in interested in such technology? Because it’s the company behind some major K-pop acts, including BTS. This past year, the boy band shocked their global fanbase by announcing that they were going to take a break from music. Fair enough, given the insane ride they’ve been on for the last number of years.
This, however, created some serious problems. First, under South Korean law, every member of BTS is now required to complete compulsory military service (they had been exempt under exceptions given to artists and athletes).
This will take BTS out of the spotlight for at least 18 months. And since the seven members are of different ages, the times when they are supposed to start their time in the army will be staggered. BTS could be MIA for years. Not good for an enterprise that has revenues of US$3.6 billion a year.
Might HYBE use Supertone to create new BTS material while the boys are in the army? It appears possible.
From a sheer capitalistic point of view, this seems brilliant. No more temperamental musicians who get drunk, high, and end up doing stupid #MeToo things with fans. New music can be summoned on cue so no more waiting for inspiration to strike. Talk about music that’s cheap to produce, too. No salaries, royalties, per diems, or any of those other expenses eaten up by real human beings.
This begs the question: If the fake is indistinguishable from the real, will fans fall for it? Maybe.
Music synthesis by machine has been a dream of scientists for decades. In 1961, Max Matthews, a programmer at IBM, was playing around with a 7905 mainframe and managed to get it to sing. No computer had ever done this before.
Stanley Kubrick would later use this as the basis for a pivotal scene 2001: A Space Odyssey in 1968.
What were cutting-edge research and science fiction are now very, very real. And while machines aren’t totally autonomous composers and performers yet, we’re headed in that direction. Right now, though, the focus is on AI-powered music creation software as a tool. Call it software-assisted composing.
In 2020, Grimes (the Canadian singer who is Elon Musk’s ex) worked with a startup called Endel to create a new piece of music she called an “AI lullaby.” She created “stems” (short distinct clips) of both music and vocals and then let the software do the rest. Endel has also been used to create music that helps people sleep and driving music to keep Mercedes-Benz drivers focused on the road.
Google is working on a system called AudioLM which can both create natural-sounding speech/singing and create music. All it needs is a few seconds of original audio and it’ll take it from there. Its piano pieces are smooth, fluid, and reasonably nuanced. No piano is necessary, either.
Harmonai is a project of a company called Stability AI, which describes itself as “a community-driven organization releasing open-source generative audio tools to make music production more accessible and fun for everyone.” It also has a tool called Dance Diffusion (currently in beta) which can generate new original short clips of music based on its knowledge of a catalogue of music. Some artists are using the software as a way of thought-starting a new composition.
Then there’s Amazon, which is working on software that can mimic the voices of dead people. Has Grandma passed on? Give Alexa a few seconds of her voice and your device will be able to speak to you like Grandma for as long as you can stand it. I’m sure getting Grandma to sing your favourite Rage Against the Machine song is just a few lines of code away.
Another area that sees potential is the industry that composes incidental music for film, TV, and commercials. This arena needs thousands and thousands of new, short pieces of music every day, a need that’s currently filled by human musicians. AI-generated production music will be cheaper, faster, and royalty-free.
In December 2015, a British company called Jukedeck started offering users the ability to create five songs a month for free and then US$7 for every track thereafter. It was followed by Amper Music, an American company, in 2017 and Amadeus Code in 2019.
There’s also Germany’s Loudly (royalty-free production tracks), AVIA (“the artificial intelligence composing emotional soundtrack music”), Infinite Album (AI-generated music for video games), and DAACI (“composes, arranges, orchestrates, and produces” original music.) And if you’re a podcaster in need of theme music, you might want to check out India’s Beatoven. There are dozens more startups that are making real progress, including AI Music!, which is now the property of Apple.
Speaking of royalties, there’s an interesting debate going on. If a machine creates a new piece of intellectual property like music, who owns the copyright on that work? The machine? The programmer? The maker of the hardware/software? That still needs to be worked out in some territories.
There are those who believe that machines will never be able to equal the emotion imparted to music. In order to really get across the trauma of a D-I-V-O-R-C-E, you have to go through one. Then again, maybe all that’s needed is a few extra lines of code.
Alan Cross is a broadcaster with Q107 and 102.1 the Edge and a commentator for Global News.