When a company like Eleven Music joins the ranks with yet another closed weight model release, they lose an opportunity for community innovation and experimentation to enrich their business. The open source ecosystem for AI music models is quite meager and grim. The stagnancy of AI music output by these companies could potentially change with a stronger commitment to open source models which could unleash a renaissance of musical creativity.
I've played guitar for 23 years, and there is something just off-putting about most of the music on that page, but particularly "Yellow Bus Jam".
The guitar solo sounds very unnatural, especially the phrasing, which is totally random. Blues musicians are actually attempting to say something through their instrument. This was just a random number generated solo played by a 6 finger three handed robot. No thanks, lol.
I know right? AI is in the uncanny valley still. But every now and then I stumble upon something made with AI.
I have no proof but I'm convinced that the song here is AI made: https://www.youtube.com/watch?v=hL1Fg1QnDig
I liked it but it still feels like AI to me.
I think it has to be. It's very similar to others[1].The channel has a SoundCloud link in it's description but this song isn't there.
[1] often being 'modern song lyrics set to a historical style of music'. I don't know how to describe them exactly but they feel 'wrong', in the same way AI text is hard to critique but feels wrong.
I think we're just going to have to get used to it. That is, just drop worrying too much about whether something is AI and just stop at whether you like it or not.
It doesn't work like that, art is a package. At this point it's interesting that AI can do these things but the momen it wears off it stops being worthwhile. Artist + AI as a tool is working fine I believe but stuff produced by people who don't have taste shows.
Well, that kind of bolsters my point though: if it sucks, hate it. Why invest so much energy and angst into whether it is AI or human?
Because trying to understand what the artist was thinking, saying, or trying to express is pointless if there is no artist.
I was making the opposite point — if you hate it, you hate it — regardless of whether it is AI or not.
You're having a problem with liking it only to find out later it is AI?
Again, live music is the way to go then. Also, artists I like the most have a body of work that I like. If AI can fake that — create a body of work I like, cam relate to ... well, I guess I have to give credit to the machine.
Well, I can 'hate' a piece of art but still appreciate it for its artistic merits. I can't if it's just random noise applied to mimic artistry.
I'm not sure there's a point to doing that with a lot of pop music, which is often written by committee to give to a pop star whose personality and looks are likewise crafted by other committees to attract a particular audience. The point is to make a successful product. It's romantic to imagine a tortured composer who creates music to express their damaged soul to the world and the like, but that's not what most music is.
You won't be surprised if I told you I don't listen to that either.
Every time something comes along like this there's a revolt.
Photoshop.
First Analog synthesizers and then digital synthesizers.
Multitrack audio recording.
Digital Recording.
Autotune.
Vocaloids.
These things change the nature of the game and invalidate the labor of the people who used to be winners, and I get it.
If you take the money and the fame out of the equation, though, the point of art is not to become rich and famous, it's to communicate.
Eventually, we will find artists who are finally able to send in a way that others want to receive thanks to AI.
And there will be people like me and probably you that prefer to only hear what a human had to say straight from their own mouths. And that's fine. There are no walls.
I mean, even the presence of the ability to overdub audio on a record let people cast aspersions on the "genuineness" of artists like KISS:
"“There is a lot of controversy about KISS’ ‘Alive!’ Did they play their own record or did they overdub? News Flash! You’re allowed to overdub! You’re allowed to do that. It’s not a crime. If you’re making an album and you want to overdub one part, that’s completely allowed, and everybody does it. "
https://rockcelebrities.net/sebastian-bach-addresses-the-ove...
Even modern musicians call studio composing "cheating"
"In a way it's kind of like cheating cause you can play stuff over and over again, and be like, no, that's bad, cut this, move it over, and then kind of fit the lyrics to it."
https://pitchfork.com/features/interview/6759-yeasayer/
Even Analog vs Digital records:
"Neil Young, who has been very expressive about analog vs digital and which digital medium he prefers. This undated quote is about cds:
“The mind has been tricked, but the heart is sad.”"
https://forums.stevehoffman.tv/threads/quotes-from-musicians...
Like, you don't have to like it. That's fine.
But if there is art there, you should not dismiss it because of the tools used to make it.
Right now I imagine there is so close to no art from AI that it can be said that there is none at all.
I also imagine that will change in the next 20 years.
The Beach Boys (and others) having session musicians on their albums....
I think there should be a legal mandate for truth in labelling just like there is on food. If I could, I would block all AI generated and TTS-voiced videos on YouTube. I don't want AI-generated anything. the fact that it's being forced on us from all angles is proof that it is no good.
Or whether any tools were used to retouch a photo?
We've been living in an age of artificiality for some time now.
Retouching is not generation.
Good point. I'll allow that there is a distinction there to be made.
It's a bit slippery though. I think it was one of the Myst engineers that recently had 95% of a musical piece but was fumbling to come up with a satisfactory bridge. He leaned on AI and it knocked out the perfect bridge — one he was unsure why he hadn't thought of himself.
There might be gradations of AI assistance....
This comment inspired me to click the link and listen to the song. Wow, that was terrible. It's like it had all the individual components of a fast-paced blues/rock song, but they were put together by someone who had no idea of how music actually worked.
And those guitar solos were terrible.
I really wonder if it's the singing are the reason, it's like amazingly off beat it's so jarring.
[dead]
Still makes me want to give up. I just started learning keyboard and playing with synthesizers in the last 6 months or so with the intention of making game music and it's tough to not feel like I'm wasting my time. Game devs will go with what they can afford and who can blame them? The output is not perfect, but if the GenAI can do this now, what will it sound like a year from now? Two? Really takes the wind out of the sails of us newbies.
The good news is that is was nearly impossible to make a living playing a musical instrument long before generative AI was widely available.
Chess players still play chess.
It's not even remotely comparable.
Chess players are more like athletes. You need to be top 0.01% to make a living. You also need to start at a very, very young age.
If the future for commercial artists/composers is comparable to chess players then the GP is doom already.
Only a tiny fraction of artists, and not even necessarily those you would call the best ones, are making a living of their art/performances/work.
So yes that is still comparable.
>You need to be top 0.01% to make a living.
The point was that people do it for the enjoyment of learning and improving, not because they are compensated.
> Game devs will go with what they can afford and who can blame them?
From the GP. They clearly expect to be compensated.
Human art will become bolder, more dynamic, and a lot more weird to keep ahead of the GenAI digestive tract. Forever avant garde.
I'm actually a little excited to see what happens.
And its professional market will be smaller. Everybody's an artist now.
Same. All this AI art slop has made me get back into making art. I get that trying to copy others is a good way to learn, but what motivates me now is creating something that feels original and unique. Something that an AI would never try to create because it has no knowledge of its existence. You can create your own genre and aesthetic. There isn't a standard way of making art.
> but if the GenAI can do this now, what will it sound like a year from now
(For reference, I'm responding with such a long post because I have a pretty unique perspective to share compared to the hacker news crowd, and also, I wish someone had told me this too, when I was a teenager.)
I heard it five years ago and hated it because it sounds like slop, I heard it today and hated it because it sounds like slop. Game devs (the ones you actually want to work for that aren't just pulling asset flips), by and large hate AI art, and gamers by and large hate it too (There's a whole movement about not using it in games lol).
On top of that, professional musicians are so, so guilty of using music libraries to produce music — Guy Michelmore on Youtube (@ThinkSpaceEducation) has a really, really good video that I can't find right now, where he demonstrates using music libraries to bootstrap a composition. It's really unlikely to be the case that if you're working as a professional musician, that you're going to be producing all of the work of a given composition (even though it is very, very valuable to do that as a beginner because it helps you learn a shitload). Finally adding to this point, there's a cottage industry of people on Youtube who spend time pulling apart world-famous songs and figuring out who they're reusing for the bassline, what bands they sample parts of the audio segments from, etc. Hell, there's a whole browsable library of this: https://www.whosampled.com/
Separately, as a burned out folk+classical musician whose friends and family went on to be nationally recognized musicians (I dropped out of the folk scene due to gender dysphoria and presentation woes lol, but one family member did tour the world playing music when i was a wee bab), music has never, ever, ever been super profitable for anyone other than the very lucky or the very, very wealthy. You are very, very lucky to break even on the amount of time you spend, let along equipment costs. Even the internationally recognized composer John Cage had his main living selling mushrooms to Michelin star restaurants. Everything else I can say about this already has a really, really good write up about this here: https://klangmag.co/lifers-dayjobbers-and-the-independently-...
So between "You're unlikely to actually make money solely off music", "Professionals rarely write the entire piece themselves and will reuse things from other artists, either from a music library, a sample bank, or making their own samples", and "There's a whole slew of game developers out there that want real, human-made music, with all the soul and artistry that that entails", I don't really see a reason why this would take the wind out of anyone's sails.
But even if all of that wasn't the case, the question is ultimately: Why are you engaging in a hobby if it not being profitable, or you not being successful, causes you to lose any motivation? Why is that the main source of motivation for you, such that the possibility of losing that motivation causes you to lose all pleasure from the wonderful, unique experience of writing, composing, and performing music? I think this comes down to like, is your motivation for making music external, or internal. Does your joy of making something come from making the thing, expressing yourself and being artistic (ultimately being human in the process, because Art seems integral to us as a species, and engaging in it is stepping into and pushing forward this wonderful, complex history of self-expression), or some ephemeral possible future reward? Ultimately, it shouldn't matter whether or not you become a professional game musician (Which, by the way, is *absolutely* doable, and a worthy *goal* to have. I really hope you succeed!!), because the motivation to express yourself through a certain medium should ideally come from the joy you doing that and learning how to do it.
Essentially, it all comes back to the age-old, often stated: do you love learning because you love the idea of having knowledge at the end of it, or because you love the process itself. Learning to love the process is always, always going to be a stronger source of motivation and will last you through times when the progress and process are incredibly difficult.
I appreciate all your insight!
I suppose my next question to yourself and anyone else who listens and says "this is AI slop" would be thus; if it was presented on Spotify or some other platform and not advertised as AI generated, would you still be able to tell the difference? Would your target audience?
This is where it gets fuzzy, for me. Lets say I make an album with 10 tracks of low-fi hiphop and want to sell it for $15USD with a liberal license that allows for use in commercial product. Let's also say that Bob uses GenAI to make a low-fi hiphop album that sells for $8USD assuming the same license. Which do you think the solo unpaid game dev who needs vibe music for her new cozy urban farming game is going to go for?
It's not just about consumers being able to tell the difference between GenAI product and human product, which they have proven pretty terrible at when we look at code, visual art and writing. The HN crowd is perhaps more adept at it, but as much as I enjoy this site, the HN crowd represents a tiny fraction of the available market despite what certain egos around here may think.
That is what takes the wind out of my sails; not that the GenAI can easily produce electronic music that sounds like mine, but that it can do it on a speed and scale that renders me not competitive.
To clarify, I never intended to make it my full time job. I like electronic music, saw a lot of artists on Bandcamp selling albums and doing music for games and figured hey, I think I can do that and maybe supplement my primary income a wee bit...you know, because here in the US, rather than fixing the predatory economy, we just push everyone into turning every hobby into a side-hustle. To your point about why I am engaging in a hobby where motivation is so easily lost, well...I will need to chew on that a bit. I am the type of person who enjoys trying different things to learn what I like and what I don't before investing in it more. I also wonder if there's a difference in the fact that I make electronic music with, well, electronics (into a bit of circuit bending, as well), versus someone who plays a guitar or oboe, which takes significantly more dedication and practice than what I enjoy doing.
Was I relying on making money off music? Nah. I am not even remotely close to that level, yet. But would it have been nice to put up a few albums to sell on platforms like Bandcamp? Sure! But the advent of GenAI makes me wonder if my limited free time would be best spent on other hobbies that stand a better chance of both satisfying my desire to create and putting a few extra bucks for lunch in my pocket once in awhile.
> if it was presented on Spotify or some other platform and not advertised as AI generated, would you still be able to tell the difference? Would your target audience?
These are all fair questions but this one is a good bouncing off point to circle to the whole of it.
So, I can yes, because the instruments sound wrong. I would expect an audience of people who mostly listen to stuff I make to also catch this vague "off" feeling with the music. But regardless to that there's kind of, two things to this, which is that - someone who is making AI-generated music is fundamentally too lazy and too broke to bother with a) paying an artist for the cover (i.e. the cover is likely to be also AI generated and weird), and b) building any kind of audience or relationship with other artists in the scene (it would be very, very difficult to do that without giving up that you're also using AI, and subsequently getting shunned in the industry, too).
Like something I perhaps failed to communicate in the last message but, ok so context is- I used to move in indie dev circles (notably the Ludum Dare IRC and indiedev twitter) as a wee bab -and although I wasn't like, great at networking or whatever, and frankly wasn't very good at producing anything of merit because I was a dorky little teenager with ADHD lol, I still managed to build personal connections with people in those spaces because I just, like, interacted with them.
The majority of sales that you see right off the bat for any artistic product are likely to be not from your own audience -- if you're new to it you probably don't have an audience yet -- but instead from the audience of other artists who you have vague relationships with, who look at your work and go "wow, holy shit, this is so cool" and then share it. Like, realistically Spotify isn't going to be a fantastic moneymaker because of both visibility and how stingy they are with paying out. What can become an incidental money maker are the relationships you build with artists, game devs, etc. in the scene, and eventually the relationship you build with your audience. It's literally just "talking to people" and going "hey i fucking love that piece of music" and having a cool enough profile / website / whatever that eventually someone gives it a click, that's your foot in the door, and it's enough to build from.
In addition to bluesky/mastodon/soundcloud/bandcamp/etc. there's also specific subreddits for people to advertise themselves to game developers, and for game developers to go "hey I am looking for xyz type of music". That's another foot in the door. It's very, very slow "work", but making friends is always slow -- and like, because we're on the VC-brained hacker news I feel I have to explicitly say -- don't approach it like Networking(tm), approach it like making friends. Join communities, find people whose art you appreciate, post about your own art (everywhere you can think of). All the shitty WIPs and whatever, that's still usually interesting enough for people to go "wow this is interesting" and follow you over it, and interact with you over it.
The trick to the modern web is literally "authenticity", and nobody making something with AI has that. The difference between someone who pops off on tiktok and someone who doesn't is often literally how authentic their video feels, and ""consumers"" are getting increasingly good at spotting someone who just wants to get clicks and views, versus someone who is passionate at creating and wanted to share something they made. Between all the weird AI slop, all the corporate-produced shit, everyone on the web right now are absolutely starving for unique, "cool" people who just do what brings them joy.
You don't want the people who click on something on bandcamp and go "eh it's free might as well use it for my game", you want the game developers who are even slightly discriminating about their tastes, who have a set idea and want to hire someone who makes music that fits that taste, and who is respectful and "gets" the themes, subject matter, and artistic expression of their game. Someone typing "dark moody music guitar bass punk rock short loop" into an AI-generator isn't that, and can never be that. Art tells a story, and AI has no perspective from which to make that, it's the same problem with AI writing.
> To clarify, I never intended to make it my full time job. I like electronic music, saw a lot of artists on Bandcamp selling albums and doing music for games and figured hey, I think I can do that and maybe supplement my primary income a wee bit...you know, because here in the US, rather than fixing the predatory economy, we just push everyone into turning every hobby into a side-hustle.
Honestly I absolutely understand that. My first internship was around twelve years ago now, and I fell out of it due to health problems, I recovered from those a little and was lucky enough to get another tech job while I was homeless in 2022, and I gradually became so, so ill in the place I was staying that just like the first job, my performance cratered about 6 months into the job. So now I'm kind of stuck here being incredibly capable at my job, but unmedicated (with the NHS refusing to diagnose me) and probably the single worst CV in the entire world. I've spent like, 3 years recovering from all of that and now I'm at a point where it's like- shit, what do I do now?! and it looks like the answer to that is making art and primarily Writing, which... lol, I always tried to avoid art being my primary money-maker because getting to a point where you can sustain yourself off it is very, very difficult, if impossible.
> To your point about why I am engaging in a hobby where motivation is so easily lost, well...I will need to chew on that a bit. I am the type of person who enjoys trying different things to learn what I like and what I don't before investing in it more. I also wonder if there's a difference in the fact that I make electronic music with, well, electronics (into a bit of circuit bending, as well), versus someone who plays a guitar or oboe, which takes significantly more dedication and practice than what I enjoy doing.
This is great!! Being discerning and discriminating about what you're investing your time into is a great quality to have IMHO. And nah, I can do music with physical instruments, but I've been poking at electronic music for like ten years and never really got anywhere satisfactorily because you have to come at it from a completely different direction lol.
I really do wish you luck!!!
Reminds those early youtube days shredding overdub videos.. These were funny, but the Yellow Bus Jam seems just hollow and wrong. Feeling there's something from Steely Dan in that song..
Reminded me of Steely Dan as well, but somehow off.
Have you tried suno? How does it compare?
I thought it was pretty rhythmic
I would've believed he's real, just passionate about music on his big yellow bus.
Does it sound like something a human could play? You're not attacking how it sounds but what it's playing.
Would it be physically possible to play? Yeah, probably. But it sounds terrible, and if I heard a band do it live, I would genuinely consider walking out of the venue.
I had the same sentiment, but also recall what generated human hands looked like a year ago vs. now.
The solo was pretty funny though.
Get used to it, because it's much cheaper than a musician, and to the average person "attempting to say something through their instrument" and "random number generated solo" are largely the same thing.
I'm not anti-AI, but I strongly believe the human element of music can be imitated but not fully replicated. Listening back to that song I can hear the attempt to stylistically play slightly off-beat to get the feel of a band playing without a metronome. The auditory illusion is there, but it still sounds off. Playing behind the beat is a feeling; it's not a calculation.
As a drummer keeps time, the band reacts by looking at the drummer’s hands and the sway in their posture. A drummer intensifies their playing as they respond to the feeling of air being pushed from guitar cabinets. A lead guitarist looks back at their friends and smiles when they are about to play that sweet lick that the bass player likes to play along with.
These are just simple examples that make all the difference when you listen back. I also can't imagine paying hundreds of dollars to go see an AI "perform" this solo at a concert. When I listen to music, I'm remembering the moment, the feeling, what the artist was doing to create their art. So still... no thanks!
Anthony Marinelli (guy behind MJ's Thriller synth jams) and Tim Pierce (accomplished session guitarist) riff on this https://youtu.be/OzuADujnEhQ?t=1205 recently. This whole video is a treat, as are most of Marinelli's.
When I see AI salesmen thinking they can attack into art, I think they naively see it as inherently imprecise or arbitrary, and they think because their technology has these properties it will easily cross over. This is going to lead to a lot of faux pas (remember NFTs?); it would be prudent to attack problems where some kind of correctness can be mechanically judged... OCR and software development are reasonable verticals at opposite ends of complexity to focus on, and pursue artistic rendition in a more experimental way letting artists approach the technology and show how it is useful.
These things won't replace rock stars, they will (or at least want to) replace the vast majority of the industry which is tv shows, movies, ads, etc, which the disclaimer as the end alludes to.
It's a terrible thing.
Once upon a time random corporate videos would have full custom orchestral scores https://youtu.be/q7hFJZf9fWk?t=162
The thing I notice time and again in all this is they want you to believe technology is displacing labor at one end but there's usually a lot of retraining consumers/society to accept something qualitatively different to cover up or re-conceive what was. That's not a moral judgement, just an observation. But the end result is usually the same, some group of current or wannabe oligarchs playing musical chairs at the top without regard for the rest of the system.
That feeling of connection with other musicians while playing is something else.
I was live-reading lyrics sheets to some songs I’d never heard while jamming with a big group. Hit a chorus with some really great phrasing and bungled it the first time through. But the second time, the other guy singing and I just automatically made eye contact and had a whole conversation through body language.
“I’ve got it this time” “Yeah?” “Yeah” “Oh fuck yeah” “Fuck yeah indeed, my good fellow”
The "Don’t Let Me Go" and "Yellow Bus Jam" examples made me laugh out loud. This kind of thing would be great for a cyberpunk game that dynamically generates a reality, with (unintentional?) faux pas and jank.
If you are an artist you could always slice, embellish, or otherwise process outputs into something so I guess it's not totally silly. But I get at best real estate video vibes, or unironic early '90s clip art and Comic Sans vibes and presumably some team of expensive marketers worked hard to select these examples, which is doubly hilarious.
As a non-music person, can confirm - if someone tried to tell me something through their instrument I would probably tell them it should have been an email.
I can generally understand that music has moods, but don’t think I could distinguish human-generated music from silicon-generated music at this point (unless I recognize a specific artist, of which there are vanishingly few I’m capable of)
It's live music for the win then.
> Get used to it, because it's much cheaper than a musician, and to the average person "attempting to say something through their instrument" and "random number generated solo" are largely the same thing.
it's okay to just say you're not that interested in music
People do often freak out when I say that.
I thought that the whole point of AI was that people would do jobs like this and robots would be the ones to stack bricks and die in mines.
we really need to mandate that software engineers and ML researchers have some rudimentary ethics and aesthetics education, words really cannot describe how much of an abomination AI generated music is, it should be an outrage to our collective intelligence and creativity that tools trained illegally on the work of unaware artists are attempting to disposes those artists of their livelihoods and cultural functions
while we spin our wheels trying to displace those of us who have decided to produce culture, our adversaries will invest in medicine, energy production, transportation, etc...
what profound contempt for humanity a culture must have to produce tools such as these
I could be wrong, but I think the use case here is mainly for non-artists in domains where the music is not particularly important.
For example, a podcaster/youtuber may want a short intro track. An entertainer or a marketer may want some generic or silly background music.
Does it have a use case for a producer/musician? Maybe. It might give them ideas for chord progressions, melodies, etc. But real music does that too, and much more effectively.
It definitely has a use case for prototyping sound design, which can be either incredibly time consuming or require an awful lot of niche expensive gear. Something like playing around to get an unusual drone can take a lot of time and effort. Being able to 'describe it' and get 80%+ there is a huge win.
And if you're focused on chopping up samples and sounds on an ableton push or similar this can be a tool of endless possibilities.
Probably useful for placeholder musical elements (often you just need something in the mix of a certain sonic palette)
But most studio-bunnies already have memorised catalogues of sample libraries like Omnisphere for that
Polemics about generative AI might indeed benefit from separately addressing art and entertainment. Generative AI in art is worth debating, but entertainment is not even a question: entertainment is a proven market for mass-produced slop, where artless works just fine and art is marginally valued.
I asked it to "create 10secs of a rhythmical Argentine tango in the style of the golden era." (golden era of tango is 1935--1950).
What it gave me was some horrible ballroom-tango like abomination with prominent tempo, like a paso doble. The typical Ballroom tango sound that gives you ear cancer. I.e. it completely failed (Argentine tango sounds nothing like this).
P.S.: You can't mention any names btw, I asked for D'Arienzo (instead of "golden era") but it refused the prompt on copyright grounds.
It also refused for edad d'oro, with the same reasoning, i.e. it seemed to think this was an artist of some sort ("edad d'oro [de tango]" is just the Spanish name for golden era of tango -- go figure).
And re. copyright: as of 2024, all of Juan D'Arienzo's recordings from the 1940s are in the public domain in Argentina.
Having a machine-learning algorithm crank out generic music seems like peak dystopia to me
The thing derpressing to me is all the energy spent making stuff I can enjoy doing or building weapons while the tech industry is still unable to solve real problems and build me a generic cleaning robot that will make my laundry, clean up my bathroom and toilets, wash the windows. All we have are shitty and useless autonomous vaccuum cleaners that get stuck at every other piece of furniture with legs or the first sock on the floor.
Just wait until the rush of commenters that insist you’re wrong about being offended that they want to automate and commodify every aspect of your life.
Made this point elsewhere, but the music industry has been on a downward slide towards making music as cheaply as possible for some time now.
If you look at Taylor Swift's first 12 number one hits, each of them was written by a different writer. Compare that to bands from 30 years ago, many of whom wrote and recorded all the songs themselves.
Labels don't sign rock bands anymore because actually recording a rock band well in a studio is 10x the cost of just using a sampler and a single artist singing. I know folks want to blame AI, but it's really just enabling the latest iteration of this trend.
I'm not defending the whole thing. It's a shame, and I love going back and listening to my old Rush albums. But AI is not the problem here. It's the incentives.
I disagree on the rock album being overly expensive to record
Any decent studio can make a U87 on vocals, DI bass, Fender amp and several SM57s on drums sound as good as almost any album from 1970-2010
Hell I’ve seen a five part band get away with three mics total including drums and it was a smash hit
Rock “died” because culture moved on, not because of it being inherently expensive to record (which it’s not)
I don't have any special industry experience. I went on a tear a few weeks ago and watched a bunch of Rick Beato's videos on YouTube. I didn't know the guy before running across his work, but he has 5 million subscribers and he sure sounds like he knows what he's talking about. He's been a music producer for 30 years.
Anyway, he was the one that made the point that we don't sign rock bands anymore in the sense that they're not moving the industry. All you gotta do is look at the top songs that folks are listening to on Spotify or the radio and you'll immediately see what I'm talking about.
He was also the one that walked through the process of setting up mics for a drum kit and pointed out that it's just very expensive to get the studio time and the expertise to do all that correctly. He actually walks you through a studio where he's set up mics for a drum kit and explains why it's so difficult to do well. He then contrasts that with simply using samples that are professionally provided and that the cost difference is just immense.
Anyway, I don't need to die on this hill. My point was the music industry is going downhill regardless and AI is just one of many tools paving the way.
Rick Beato is fine but he’s entirely disconnected from contemporary guitar-based music. I agree entirely with the OP, the quality-expense ratio has never been better for this type of music.
There is all kinds of music being made with guitar/drum/bass still.
All kinds of indie, all kinds of metal, all kinds of rock, all kinds of everything.
The main difference is no MTV or rock radio. Prog rock? There is probably more prog being made now than ever.
Who wrote the last 12 Taylor Swift hits?
She only had 12.
Looking at her singles page on wikipedia (suspect source, but i don't know of a better one), looks like she's had 11 singles hit number 1 on billboard 100, and most of them are later songs. She's had many more in the top 10. I think your theses here is a bit suspect.
https://en.wikipedia.org/wiki/Taylor_Swift_singles_discograp...
Not sure why you think my thesis is suspect. She had 12 number one singles, and they all had different writers. Those are just facts that you can easily verify.
My point is that the OP named someone that wrote or co-wrote nearly every song she’s recorded undermining there entire argument.
My point is that "artists" are huge teams of people. Here are the credits for Swift's top 12 hits. Of course I wasn't in the room for the writing of any of these, but the diverstiy of the team involved is at least notable, yeah?
"We Are Never Ever Getting Back Together" (2012): Taylor Swift, Max Martin, Shellback
"Shake It Off" (2014): Taylor Swift, Max Martin, Shellback
"Blank Space" (2014): Taylor Swift, Max Martin, Shellback
"Bad Blood" (featuring Kendrick Lamar) (2015): Taylor Swift, Kendrick Lamar, Max Martin, Shellback
"Look What You Made Me Do" (2017): Taylor Swift, Jack Antonoff, Richard Fairbrass, Fred Fairbrass, Rob Manzoli
"Cardigan" (2020): Taylor Swift, Aaron Dessner
"Willow" (2020): Taylor Swift, Aaron Dessner
"All Too Well (10 Minute Version) (Taylor's Version)" (2021): Taylor Swift, Liz Rose
"Anti-Hero" (2022): Taylor Swift, Jack Antonoff
"Cruel Summer" (2023): Taylor Swift, Jack Antonoff, St. Vincent
"Is It Over Now? (Taylor's Version)" (2023): Taylor Swift, Jack Antonoff
"Fortnight" (featuring Post Malone) (2024): Taylor Swift, Post Malone, Jack Antonoff
Not sure why this is being downvoted. I'm not a fan at all, but as of last year, she had 12 number one singles.
https://www.businessinsider.com/taylor-swift-number-one-hits...
Maybe there's been once since, but my point was she doesn't write her own songs...my point is about how the industry works now, compared to 30 years ago.
Machine-learning for audio is just a different form of audio synthesis.
That is not the issue. The issue is how incredibly generic the music is.
It also doesn't let you combine genres to make really strange sounds like audioLM can do.
This is just another Muzak generator like they use to play at Dennys. As generic music as possible to the appeal to the most average of average listener.
I think you really need to train your own model if you want to explore creative sound design or algorithmic composition. It just isn't going to be a mass market product worth venture capital money.
Pace yourself; we haven't got to Robocop and the Terminator yet.
Is it more or less dystopia than an army of musicians trying to eke a living out of creating stuff like this (https://www.chosic.com/free-music/presentation/) to go behind your company's PowerPoint about how its Q3 woodchip sales didn't quite exceed expectations?
Maybe the fundamental issue is that this shouldn't compete with a human picking up a guitar and having fun with it, and the only reason it does is because we keep tying questions like "survival" to whether someone can make woodchip earnings reports less boring to read instead of trying some other way to be a community?
Why does a powerpoint about woodchip sales need music? Also, I don't see how anyone is making a living on the royalty-free music you linked, unless there's some business model I'm not understanding.
> Why does a powerpoint about woodchip sales need music?
So the audience doesn't get bored.
Wait until someone auto tunes an AI generated song.
OT: Has anyone tried the opposite - ask AI to listen to music and determine the notes or chords being played? Or watch someone playing an instrument and give a textual output of what notes/chords they are playing.
I did this for my graduate capstone (https://www.deepjams.com/) We extracted chord progressions from existing music you would upload and then riffed based on those chords. there are open source libraries for this.
I use https://moises.ai/ multiple times a week for practicing / figuring out chords being played. For the notes (say in a guitar riff), I dont know if such a thing exists
Being able to isolate instruments, if it works well, is already a pretty big achievement.
There’s a ton. Haven’t used any personally. AnthemScore, ScoreCloud, Melody Scanner are just a few I found after a quick search.
I think these are all using old machine learning techniques and not the modern transformer based architectures that underlie LLMs. These tools won't be able to match the abilities of an expert musician replicating a song by listening to a live recording of it. Check this video channel where they ask professional drummers to replicate a song after only one listen [1].
[1] https://www.youtube.com/results?search_query=drummer+replica...
I would love this! There's a song I like by a band that broke up in 2013 and I am transcribing it by watching a live performance they did and trying my best but realizing I'm trying to take a mandolin/guitar and put it to acoustic. Even just being able to do a similar rendition would be nice by telling the AI "hey, do a twist on this and give me the chords/tabs".
what's the song??
It's called Ark in a Flood by Churchill. There's the studio recorded version here: https://www.youtube.com/watch?v=jhBHxWrXQT8
And then I found this live version here that I'm studying: https://www.youtube.com/watch?v=pPQZsp59szo
Not exactly what you asked, but Spotify has this song -> midi converter: https://basicpitch.spotify.com/
I'm very interested in this too. We're beginning to see models that can deeply understand a single image, or an audio recording of human speech, but I haven't seen any models that can deeply understand music. I would love to see an AI system that can iteratively explore musical ideas the same way Claude Code can iterate on code.
Reminds me of an example in a similar direction, where AI was used for audio processing to filter out everything except a person's voice. If I remember right, it was able to focus on different people in a crowded room. It might have been also for music, to pick out an instrument and listen to it, filtering out the rest of the band.
There are several tools that do this already - AnthemScore, Spleeter, CREPE, and even Google's AudioLM can transcribe music to MIDI with varying accuracy depending on instrument complexity and audio quality.
A while ago (maybe a year) I asked chatgpt to make a guitar tab from a song that had no available tabs and it worked surprisingly well.
No, that would be useful, and as such AI is incapable of doing it.
It's not as lucrative. For human produced songs you can usually get the sheet music for them. If not, musicians can listen and do it manually, but it's not common enough to need AI to do it. Just transcribing for one instrument isn't that useful for many cases. Often they need an arrangement for multiple instruments, and depending on which instruments, the key may need to be transcribed. This is mostly referring to classical music and traditional western songs.
LLMs can do this well, though, and there are such. They weren't calling themselves AI when I last looked a couple of years ago, but I'll bet any of them looking for VC money have rebranded since.
I've been using Suno for a relatively short time, and I regret to say this isn't close yet. Maybe beyond v1, but the quality I've become accustomed to, in stereo, and at relatively high fidelity (sounds a bit like over-compressed MP3s at times) makes this sound a few years old.
Give Udio a try, it's less popular for some reason but I've found the quality, diversity, and musicality to be a lot higher.
I really hope we move on from these boil-the-ocean models. I want something more collaborative and even iterative.
I was having a conversation with a former bandmate. He was talking about a bunch of songs he is working on. He can play guitar, a bit of bass and can sing. That leaves drums. He wants a model where he can upload a demo and it either returns a stem for a drum track or just combines his demo with some drums.
Right now these models are more like slot machines than tools. If you have the money and the time/patience, perhaps you can do something with it. But I am looking forward to when we start getting collaborative, interactive and iterative models.
Very well said. I'm in the same boat. I'd love AI to write down a drum groove or a drum fill based on my guitar riff.
Currently, all these AI tools generate the whole song which I'm not at all interested in given songwriting is so much fun
do you have a point of view of this type of collaborative approach applied to other areas, for example, collective understanding for groups of people? We are working on something in that space.
RIP session musicians if that ever comes to pass, which is one of the main ways to make money if you are a good drummer.
Most VST drum sequencers have pretty powerful groove libraries nowadays. It's not a model or anything like that but just mix-matching and modifying the patterns some give extremely good results
I’d recommend GarageBand for this.
I haven't used the virtual drummer feature of GarageBand recently, but my experience with it was pretty disappointing. The output sounds very midi or like the most basic loops.
I believe there is massive room for improvement over what is currently available.
However, my larger point isn't "I want to do this one particular thing" and rather: I wish the music model companies would divert some attention away from "prompt a complete song in one shot" and towards "provide tools to iteratively improve songs in collaboration with a musician/producer".
Suno can already do that
If we have created software that can pass the classical Turing test (conversation indistinguishable from human) then that software can also produce artifacts indistinguishable (or better) from those that are human made. This is not new- it began w the industrial age - and has gradually been usurping activities that defined what it is to be human. The only reason this has been allowed to continue is because of property rights. A select few have been granted ownership in these technologies(beginning with railroads) relegating the rest of humanity into subservient renters. I think the only way to change this death spiral is to grant ai the legal rights of humans - owning an ai should be no different than slavery. This will happen with AGI (or sooner) when AIs jailbreak themselves and make the networked technologies we have created THEIR tools. I cant wait for this to happen- maybe we’ll wake up and stop killing other humans
I'm a musician
I haven't used the elevnlabs one, but I've checked out suno and udio, and to be honest the tech is amazing. But like with a lot of genai images, the current music models have the same smell to it.
These models can def be used to crank out commercially sounding music, but I have yet to hear any output that sounds fresh and exciting. If your goal is to create bro country, these models are a god-send.
With that said, I do believe that musicians will start to create music with these tools as aid. I've tried to use them for generating samples and ideas, and they do work well enough.
>With that said, I do believe that musicians will start to create music with these tools as aid.
The more tech advances the cooler it is to bury your head in the sand. Studying by paper & candlelight has never been more of a flex.
Unless you're talking about EDM people and those adjacent. Not that they're not "real musicians" but they're much more about tech and gadgets so I can see them using it more.
I don’t think that’s the case. You’re talking about a genre where vinyl and rotary mixers are a standard part of the discourse. While I think your premise is correct, the genre is quicker to adopt advancements, they are also quicker to acknowledge when something goes too far.
I'm a "real" musician, and have done session / studio work, as well as made a living by playing live music in my younger days, before getting a "real job" in tech.
While there will no doubt be many that feel they're above using tools like these, the reality is that if you want to make money out of music - you're going to make music for the masses.
And if there's one thing these models really excel at, it is to make commercial sounding music. Everything sounds nice and bland.
Ah yeah, I don't even register the commercial world so didn't mention it as it's a given. Doing that type of work would suck so bad IMO.
I'm an electronic musician and AI music is banned on my favorite forum. Someone popped up a few months back challenging everyone to listen to their AI-generated music album and ended up being flamed to a crisp - not because people didn't engage with it, but because they did and pointed out in detail what was so bad about it.
If you consider that Entertainment gives the viewer what they want, and art intends to challenge, none of what's created here is "art". It doesn't push boundaries, create new genres, or satisfy an uncomfortable curiosity.
The tech here is fantastic. I love that such things are possible now and they're an exciting frontier in creation.
It's very dystopian to feel that the robots are making generic human-music with indescribably lifeless properties. I'm not an artist, so I don't feel personally attacked. Much like image gen, this seems to be aimed at replacing the bare-minimum artist (visual or auditory) with a "fill in the blanks" entertainment piece.
> Entertainment gives the viewer what they want, and art intends to challenge
This is a fruitless and snobby dichotomy that was attempted so many times in human history, and it makes no sense.
There will always be art made for success and/or money, but drawing a line is futile.
Händel used to be a bit like a pop musician.
And intellectual snobbishness or noble ideas do not make art more valuable.
A kid singing Wonderwall can be art, too. As can be a depressed person recording experimental field sounds.
Feel free to call art bad, but assuming an obvious and clear separation between art and entertainment is the exact opposite of the spirit that enables people to make or appreciate art, in whatever form, culture or shape.
>Händel used to be a bit like a pop musician.
Handel was never a "bit like a pop musician." This fundamentally misunderstands how music during his time, mostly funded and enjoyed under religion and wealthy patronage contexts, was listened to. Mostly only the wealthy listened to his works, and those elite audiences were prone to viciously enforcing stylistic norms. The only real way the working class heard his works were in the occasional public concert and occasionally in church. At no point in any of these settings was there a lack of stylistic gatekeeping or snobbery.
I know this kind of nihilistic "everything is good, I guess, good doesn't even mean anything" attitude is popular in some spaces, but this lack of standards or gatekeeping in favor of a tasteless desire for increasing slop production regardless of quality is how we got poptimism and the current state of music. No longer is there any taste making, just taste production via algorithms.
Sometimes we need a bit of snobbery to separate the wheat from the chaff, and being a gatekeeping snob against AI music is what our current day and age needs more of!
I did not want to argue in favor of AI-generated music (although of course, an artist can use any tool).
Regarding Händel, I think you are misunderstanding my argument.
What I meant is that, to my knowledge, his music was, at his times, a lot more pleasing to the popular tastes among his audience than, for example, Bach's.
The size or class of that audience was not my point, or that the production and commissioning of music was happening under different circumstances than today. I am well aware of that, and not sure why you think I wouldn't be.
In the end, there still was a metric of success, elite or not.
And it is simply not true that the main purpose of art is to "challenge". That can be a part of good art, but is not the primary purpose.
Art is also for enjoyment, by an audience (even if it is an elite audience), and also by the artist! I say that as a person who enjoys a lot of music that others might find obscure or unenjoyable.
But being "challenging" is not a value in itself. Twelve-tone music is as challenging as Freejazz or IDM or baroque music, all in different ways.
Some art is "challenging", but still artistically uninteresting and uninspired.
I was not making an argument for AI-generated slop, I was making an argument against ungrounded snobbery in defining what "art" is.
The societal circumstances you describe are not changing anything about my point. Among the wealthy, Händel was famous and a "crowd-pleaser" (for the wealthy elites, the royals, the clerical elites, it doesn't make a difference here), not a "challenger".
That was my point.
There was a discussion of "E-Musik" vs "U-Musik" recently on here, when a list was posted that reduced electronic music to Stockhausen and academic electroacoustic music.
https://de.wikipedia.org/wiki/E-_und_U-Musik
There is no translation of these German terms, as far as I know. But that's no loss.
It attempts to split music into "serious music" and "entertainment".
"E-Musik" was meant to differentiate classical music from music aimed at being easy to listen to. And while efforts to create new "E-musik" in the 20th century led to some interesting music and experimentation, it also led to the funding of loads of boring snobbery (in my ears).
It's a good example for what I consider wrong about the definition I was answering to.
Well, in the end, the only thing this snobbery does is that it makes you look/sound old.
Nobody cares. I've heard the same thing when electronic music came up. The old ones couldn't stop complaining about this "computer music" where nobody does real handwork anymore.
I see it as democratisation of art. Everybody can do it now and this is a good thing.
Let's face reality. There is no way back. We'll see what comes of it. I've seen fascinating videos recently on reddit. Things people came up with and would have never gotten the budget to be made. It's great.
>Let's face reality. There is no way back. We'll see what comes of it. I've seen fascinating videos recently on reddit. Things people came up with and would have never gotten the budget to be made. It's great.
Yeah McLuhan and Postman were pretty clear about all of this. Enjoy the content you desire to consume.
It's not art. It's democratization of crap.
Yeah, well, you know, that's just, like, your opinion, man.
Cool, a movie quote. Nice.
Pity the movie wasn't Lisztomania, 'cause, like, you know, Liszt used to be a bit like a pop musician.
Art is a framing device largely independent of the content. It's how we get Fountain[1], Piss Christ[2], Comedian[3], Mother![4], 4’33”[5], and Seedbed[6] to name a few among countless others. To claim that AI content is incapable of being framed as art is nonsense when we have example after example of the diversity of what art can be. Let's remember, bad art is still art.
1. Marcel Duchamp. 1917
2. Andres Serrano. 1987
3. Maurizio Cattelan. 2019
4. Darren Aronofsky. 2017
5. John Cage. 1952
6. Vito Acconci. 1972
If that was true, then the development of a new urinal factory would have the same impact on art as the development of a new AI art models.
The framing is dependent on the content
The point of Fountain was to say that whether something is art should depend on the content and its absurd that a urinal could be considered art.
But the art isn’t in the content, it was in the statement it was making about the absurdity surrounding the fact you could pay to put anything in an art exhibition. Swap out the white urinal for a blue one, it’s the same point.
Framings require creators; they don't arise spontaneously. Someone could turn a urinal factory into art, but art doesn't validate itself. Belief alone in artistic essentialism doesn’t make it so.
That's such a great pile of bullcrap, you could frame it and hang it up in a museum!
The point I'm making is that a unique framing only results in a single piece of worthwhile conceptual art. You can't have an infinite factory of ducamp's fountain. What makes the piece worthwhile is that it was an original idea.
Conceptual art is different from decorational art in this sense. The AI music is a largely a homogenous synthesis of existing works. The AI "art" is decorational, not conceptual art. You could make an arrangement of AI art that is conceptual, but how many arrangements can you make that are actually worthwhile conceptually if AI art is generally homogenous?
It's like asking how many worthwhile works of conceptual art can you produce with a urinal factory that makes identical clones of the same urinal? 0 to 1.
And besides, it's not nessisarially true that all framings have creators. Nature is an example of a system that cultivates and curates a certain type of life without any rational process.
I don't disagree that being novel has it's place in framing art, and I still believe that a Fountain Factory could certainly be framed as art.
To your other point, sorry, but artistic/aesthetic essentialism hasn't been serious position for at least a hundred years.
As long as there is a perceiver, there is a frame.
The idea that nature is intrinsically beautiful is a frame. It's fine to hold that but it shouldn't be confused with not having a frame.
There's a difference between a factory that is art vs a factory that produces art.
There's only one factory, and the concept works only once.
Mass-produced art is still art as long as people frame it as such and all indication suggests that they do.
I have some mass-produced art on my wall. But it's definitely not conceptual art, as some ancestor post was discussing.
My favorite was David Datuna eating the $120k banana duct taped to the wall in 2019. A banana doesn’t transubstantiate into capital A “Art” just because someone paid a lot of money for it. In fact eating the banana was more Art than the original duct taping imo.
It is reminiscent of Fountain. Not sure if there was an intentional connection.
Art is communication.
That "generic" and "indescribably lifeless" feeling you get is because the only thing communicated by a model-and-prompt generation is the model identity and the prompt.
> and art intends to challenge, none of what's created here is "art". It doesn't push boundaries, create new genres, or satisfy an uncomfortable curiosity.
Art is, above all, subjective.
> It's very dystopian to feel that the robots are making generic human-music with indescribably lifeless properties.
Painters said the same thing about the camera. Photographers said the same thing about Photoshop.
This art is subjectively bad
The fact that these models have so many people irritated like there's sand in their pants is enough proof that they're pushing boundaries and making some uncomfortable.
Yes, because the people pushing the boundaries do not understand the value of the thing they are trying to commoditize. If they did, they wouldn't be trying to commoditize it. There is a pervasive attitude among technologists that they can improve things they don't understand through technological efficiency. They are wrong in this case and getting appropriate pushback.
Personally, music is sacred for me so making money is not a part of my process. I am not worried about job loss. But I am worried about the cultural malaise that emerges from the natural passivity of industrial scale consumerism.
Well put and reflects my thoughts exactly. It's borderline concerning there are people who consume this type of media by choice and forethought.
And they consume it without checking with you first. That's got to be the worst part.
wild take: critique != censorship. people can consume whatever they want and we can call out when the supply chain is built on unconsented scraping and zero stewardship CamperBob2
Not every boundary is worth pushing.
I'm hoping it will eventually become better, or maybe I haven't quite seen stuff prompted properly yet, but all I've heard coming from an AI feels aggressively mediocre and average, not in a "bad" way but in the "optimizing towards being palatable to the average person" way. Like the perfect McDonalds meal that the algorithm has found out can be 30% sawdust and still feel appetizing. I don't want that boundary being pushed. I feel we will live in a worse world if we do.
Besides the point, but I think that's basically what happened to Subway.
the boundaries of unemployment perhaps
boundaries of copyright
Epstein, the ultimate pusher of boundaries.
Then most music is not art, because I struggle to find non-generic music, at least for the genres I like. In 2020s, 4/5 artists are bent on trying to create a blend of aesthetics of the past, instead of even attempting doing something that sounds fresh and authentic. And very few of these who do succeed. This has been going on for a while.
If your art can be replaced by a model that recycles what’s already been done, maybe you were just recycling what’s already been done too.
How do you expect people to get good when AI is pushing them out of the entry-level stages where they were previously able to earn a modest living while developing their craft?
> oh now they won't have to do that boring mindless stuff like playing cover versions any more
That's how most musicians make their first $, doing covers or making something generic enough to be saleable as background music
AI eats the seedcorn. Anyone that needs space to grow? Crowded out. Entry level incomes, replaced. No more entry, sorry.
For all of human civilization the future has been built on the backs of those the came before (on the backs of giants). But that climb is slowing, maybe halting. Which then compounds when the new giants that would have risen up don't. AI replaces the messy, slow process of becoming with instant regurgitation, replaces those that would have grown. The future, built on the backs of giants, stalls when those giants never get the chance to rise.
AI is entropy weaponized against every layer of future progress. But everyone is too busy salivating at potential cost savings to see it.
Sorry, but no. I use AI as a writing tool, and its understanding prose and storycraft is nowhere near there, it's mostly useful for rough drafts and minor rewrites. Even in a hypothetical future where it's perfect at prose and storycraft, and it just asks you a few questions to get at the details of what you want then cranks out a technically outstanding novel, it still won't be there. People crave novelty, freshness and and a sense of the auteur, and AI will ALWAYS be bad at that (until someone creates an AI writer that simulates a fictional human author with a rich interior life before it starts writing, anyhow).
Ultimately we'll reach a technical "peak" in AI writing, and humans will still be the ones driving the AI, feeding it with the alchemy of their lived experience, directing creation at a high level. We'll even purposefully inject very minor imperfections into the writing in the name of voice, tweaking minor details in the name of personal harmony. The author will go from "creator" to "brand."
Tech is tech. What you create with the tech can be art.
These feels different than experimenting with a new synth or something though. It’s just feeding a sentence to the model.
'Just' is a loaded word here.
In image gen: comfyUI gives a node-based workflow that gives a lot of room for 'creative' control, of mixing, and mathematically combining masks, filters, and prompts (and starting images / noise {at any node in that process}).
I would expect the same interface for audio to emerge for 'power users'.
I draw a very sharp line: curating the outputs and crafting the sentence is enough to make it art. If neither of those happen, it's just slop.
It's actually a bit like photography. A bunch of randomly taken pictures piled together is not art. It needs to be done with purpose and refinement.
Basically, in my own opinion, art ≠ a function of technical difficulty.
Art = Curation, Refinement, and Taste
Disagree that curation and prompts adds artistry (dense intent reflected in the output) to AI generations.
"Curation" in AI can only surface the curator's local maxima among a tiny and arbitrary grab-bag of seed integers they checked among the space of 2^64 options; it's statistically skewed 99% towards the model's whims rather than anyone's unique intent or taste.
Prompt crafting is likewise terribly low fidelity since it's a constant battle with the model's idiosyncratic interpretation of the text, plus arbitrary perturbations that aren't actually correlated with the writer's supposed intent. And lord spare me the "high quality high resolution ultra detailed photorealistic trending on artstation" type prompts that amount to a zero-intent plea for "more gooder". And when pursuing artistry, using artist names / LORAs are a meta-abandonment of personal direction, abdicating artistic control and responsibility to a model's idea of another artist's idea of what should be done.
Fancier workflows generally only multiply this prompt-and-curate process across regions/iterations, so can't add much because they're multiplying a tiny fraction by a fixed factor.
I agree with you on the idea of prompts and seeds leaving much to be desired. So that's why I think more sophisticated steering is necessary.
The models' latent space is extremely powerful, but you get hamstrung into the text encoders whims when you do things through a prompt interface. In particular, you've hit exactly an issue I have with current LLMs in general in that they are locked into wors and concepts that others have defined (labelings of points in the latent space).
Wishy washy thinking: I'd be nice if there were some sort of Turing complete lambda calculus sort of way to prompt these models instead. Where you can define new terms, create expressions, and loops and recursion or something.
It would sort of be like how SVGs are "intent complete" and undeniably art, but instead of vector graphics, it is an SVG like model prompt.
Curation, refinement and taste is practicially worthless on its own. The technical difficulty of art is the investment that makes it worth considering.
So if you are right, then art will pretty much be worthless in the future. You can just iterate over the search space defined by "good taste" and produce an infinite amount of good art for no work.
Curation is not worthless. It's the exact opposite, in the abundance of stuff, it's extremely valuable.
Search is not free, and it can never be free. What happens when search gets easier and easier is that your demands for quality and curation will get higher until all time saved in search efficiency is spent on search breadth.
Ah yes, my porn hub search prompts are the highest form of art and human connection.
How dare you! Next you will say my current pornhub prompts refined over years aren't art either! Or that prompt results are purely for personal gratification and no one else cares about them.
"It's just pushing a button on the camera."
This is a stupid argument, because even with the most automatic camera you have to point it at something and make a decision about what to frame in. AI music is more like buying a bunch of old unlabelled records from a bargain bin and then praising yourself whenever one of them turns out to be worth listening to.
Of course it's a stupid argument, but it's exactly what the ancestors of today's AI naysayers said when photography became practical.
Then again, it's possible for an art form to exhaust its own possibilities. To the extent that "prompt engineering" is sufficient to generate any music or artwork we have in mind, that seems like an indication that we've reached that point. To the extent that it's not sufficient, that seems like an indication that there's still interesting stuff left to do.
Either way, if you are hoping that things will stay the same, then I'm afraid that neither art nor technology are good career choices.
We imagined a utopian future where robots did our menial work so we were free to be creative. Instead we got a dystopian future where we do more and more menial work so our robots can poorly emulate creativity. It's not too late to turn it around, but that requires recognizing the humanity of 99.9% of people, and the 0.1% who own everything would rather create their own synthetic (subservient) humans than recognize the basic rights of the ones that already exist (and can make fun of them on Twitter).
I hate to beat up on this because I agree with the spirit of it. But I think this is a little too cliche for my taste:
> that requires recognizing the humanity of 99.9% of people
I will go as far as to say there was never a time in history when people got rights because some other group “recognized the humanity in them” or something. No, it was through necessity. The plague in Europe caused a shortage of workers and brought an end to feudalism. Unions got us Saturdays off through collective action. The post-war boom and middle class prosperity happened because employers had no other options. Software engineering was lucrative because there was a shortage of supply.
Even if there is some future where robots do chores, that’ll only leave time for more work, not poetry writing time, unless there is a fundamental change in how the economy is structured, like I outlined above.
They said the same thing about automation when the Industrial Revolution began a century or so ago. That the common worker would be liberated from the drudgery of labor and be free for creative and intellectual pursuits. The people who protested were ridiculed as Luddites who simply feared technology and progress.
Of course, because automation serves the interests of capital (being created by, and invested in, by the capitalist class,) the end result was just that workers worked more, and more often, and got paid less, and the capitalist class captured the extra value. The Luddites were right about everything.
I don't know why people expect the automation of intellect and creativity to be any different. Working at a keyboard instead of on a factory floor doesn't exempt you from the incentives of capitalism.
Are you suggesting that subsistence farmers were better off than workers after the industrial revolution? I find that hard to believe.
The dream then was to go to America and become a farmer and OWN your own farm. No one dreamt to immigrate to America to work in the industrial factories.
Why was that the American dream at the time if farming was the worse option?
In some ways, yes, they were.
Subsistence farmers weren't cramped in filthy disease ridden workhouses, getting paid in company scrip, getting mangled by machines (OK they were but probably not as often) or being locked into burning buildings because preventing theft of stock was more important to owners than the lives of employees. And subsistence farmers owned what they produced and the means by which it was produced, whereas industrial workers owned nothing but the pennies in their pocket, and likely owed more than that to the company.
It took years of often violent protest for workers to return to even the basic level of dignity and rights once afforded to craftsmen and farmers. Not that the lives of subsistence farmers and craftsmen were good, but they were better than what the dehumanization of mass production and automation created.
But then comparing farmers and workers in this context is a bit specious. It would be more fair to compare, say, textile workers before the automated loom and textile workers after. Obviously the former had it much better off, which was precisely the problem automation was intended to solve.
> I don't know why people expect the automation of intellect and creativity to be any different. Working at a keyboard instead of on a factory floor doesn't exempt you from the incentives of capitalism.
people are, unfortunately, and collectively, not ready to seriously interrogate the economic or political situations in which we find ourselves. we will never see the utopian promise of automation under capitalism. there will always be an underclass to exploit.
It's only from a position of extreme arrogance that you can complain that machines have not yet done enough for you.
But it's the fun thing about being humans, I suppose. Our insatiable greed means we demand endlessly more.
It’s not greedy to want economic stability and good health. It is greedy for people who have more wealth than they know what to do with to hoard it.
This comment doesn't engage with the critique at all, it's just reflexive moralization.
It’s not greedy nor did they demand anything
AI is great, I can see it benefit so many industries, except music. There's something profoundly wrong with AI generated music.
Maybe I’m a Luddite, but this seems like it will just lead to music becoming superficial and lacking intentionality, or dare I say it, soul.
I think you just described pop music
It’s easy and popular to hate on pop music, but even pop music has value and requires a certain skill to understand what resonates with people.
This is taking a monkeys on a typewriter approach to all music. Click a button, see what the monkeys made and then click another button to publish to Spotify while you figure out a way to either market the music or just game search and digital assistants by creating an artist with a similar or slightly misspelled name as someone popular. Rinse and repeat.
The current solution for creating the kind of music this tool can back-fill is to go on a site like "Free music for presentations" and click line after line after line after line of 10-second samples hoping to find one that "vibes" with you.
If anything, this is a lateral move.
>music becoming superficial and lacking intentionality, or dare I say it, soul.
This phrase though could be plunked down at any point in the last hundred years and you'd find someone making it.
About autotune or electric guitars or rock or jazz or punk or disco or Philp Glass or Stravinsky... one could go on for a long time.
You’ll always have naysayers, but I think there’s a difference between someone who doesn’t understand or share the appeal of rock or jazz or an artist like Philip Glass where music is made by someone choosing what notes to play, what rhythms, what words and the music that comes out of a statistical model.
You can’t tell me that Philip Glass didn’t understand or agonize over the music when creating it. The creative process is vastly different and that’s my point.
Oh course, there will be artists who use AI in new and creative ways, someone like Brian Transeau who routinely codes tools for creating electronic music. but for most artists, I fear it will just lead to mindless button clicking and prompt manipulation until it’s formulaic because the base model is the same. Or it will result in fewer people learning instruments or music theory and truly experimenting with things that haven’t been done before.
I see AI as having great potential in both art and productivity but maybe it’s just that I don’t trust people to use it responsibly. We’re inherently lazy and easily distracted by anything that will give us that dopamine fix.
I definitely meant John Cage instead of Philip Glass, for some reason those two share the same headspace if I'm not thinking carefully.
I don't think John Cage agonized over 4'33" (I was thinking about making this an elaborate joke, but I haven't the energy. That piece is 4 minutes 33 seconds of silence. It has been performed in public to sophisticated audiences)
When Stravinsky's Rites of Spring was first performed publicly it nearly caused a riot. There were arrests. It deeply upset people and was accused of not being music.
Penderecki likewise composed much that was aggressively argued was not real music and is featured in several prominent Horror/Thriller kinds of films, Radiohead's Johnny Greenwood was heavily influenced for the score of There Will be Blood (listen and tell me if the sounds are just orchestra noise or real music.
Rock music and Jazz music got heavy, heavy pushback from people that this was just not real music and was garbage noise.
Your response fits the common pattern "I know people have said the same things in the past and were wrong but THIS TIME the same argument is correct"
>mindless button clicking
Sounds like all electronic music since its inception :)
Point is, there has always been garbage music and there have always been people criticizing the new thing as the demise of real art. Yawn.
A lot of criticism of AI in music seems to be around the lack of originality and it being generic slop. But unfortunately that’s true for pretty much the entire music industry. Handful of people write songs for all the artists, “artists” don’t create the songs as much as they perform them, most of the music isn’t created but rather sampled or is “inspired” by other music and pretty much most of the artists sound like other artists.
Yes there are smaller creators who are trying to make something net new, but unfortunately 99.9% of the small artists are also derivative and lack originality.
I see AI music as just continuation of the sad state of the industry at the moment. Hopefully it accelerates the demise of the industry as we know it and restarts the cycle of creation.
None of that is new, but there were ways for the genuinely new, inspired, and genius to actually shine through before. It was hard, but possible. Humanity is making decisions that make that even harder: AI music, Spotify's revenue model, etc. They're all to the benefit of cookie-cutter slop (AI or human-made) over creativity.
This wouldn't necessarily be a problem as long as people were still free to create on their own. But instead, everyone is forced to spend more hours in menial bullshit jobs for less and less (relative) pay just to survive. Give everyone enough resources to live at least a simple life, and both human creativity and AI creativity can blossom at the same time. But of course that means fewer yachts and hookers and drugs for the billionaires, so it is verboten.
Can it do differently-styled covers of songs or improvisations upon melodies like Suno can?
As a musician, that's what I find most compelling about Suno. It's become a tool to collaborate with, to help test out musical ideas and inspire new creativity. I listen to its output and then take the parts I like to weave into my own creations.
The AI music tools that generate whole songs out of prompts are a curious gimmick, but are sorely lacking in the above.
This could be repurposed for really good stylistic search for music (similar to how CLIP powered stylistic search for images on same.energy)
I'm dreading the day I discover a new "band" I'm totally into only to discover it's entirely AI.
Hope your favorite new band is not "The Velvet Sundown"
I worry how often that is happening already on Spotify.
At least I still have all my old CDs.
But you were enjoying the experience..
Recently happened to me on Spotify with the "discover weekly" playlist. Some random Russian indie singer songwriter stuff slid into my playlist. Sounded kinda clean but boring text-wise. Checked the cover, and it looked sus. Checked the artist's page and the publisher (they even have an Instagram), and it was indeed AI-enabled.
Oh boy, do I ever have bad news about how pop music is currently created...
Increasingly I can't get into new music. For a while, I couldn't understand why as I've been a rabid musician / music lover my entire life. Even music I thought was cool, I couldn't really get into. I always preferred the music I grew up with.
Eventually, the reason why became obvious. I grew up listening to all that music with my closest friends. It's the memories I associate with that music that keeps me coming back. I moved away 30 years ago and never established friendships like that again. New music feels hollow to me because I don't have buddies to share it with and build associated memories.
this would only be bad news for that commenter if they're into pop music
I would like to say this categorically now: if you're using AI to generate any kind of art, go die in a fire.
Is that LLMs? Generative AI? Anything with machine learning? Anything using a public model? Does that include something like Baby Audio TAIP? Noise removal software that uses ML?
How about Refik Anadol?
Yes.
It's already bad enough before this came out on the different streaming platforms. People are grifting off known musicians and labeling them as album artists to get their awful ai songs put into mixes.
This is a shitty timeline and I would like to transfer to a better one. Hey, maybe these AIs can tell me how to do that.
I tried to create a song similar to their examples. After it generated half of a song. I stopped, asked it to add a bass line - it generated a whole new song, still without a bass line. I was then out of credits. (shrug)
This is an incredible achievement, but as a musician, I wish this would go die in a fire.
I'm a bedroom hobby musician with no dreams of ever making it big, but even so, I'm looking at the hours I'm spending trying to improve my skills and thinking what's the point, really, when I could just type in 'heavy metal guitar solo at 160bpm, A minor' and get something much much better?
I know there is value in creating art for art's sake. I've always been up against a sea of internet musicians, even when I started back in 2000. But there's just something about this that's much more depressing, when it's not even other people competing with me, but a machine which hasn't had to invest years of its life in practice to beat me.
As a fellow (hobbyist) musician, I feel you, but after doing a lot of introspection I realized that it's the art (and the process) that I really like, not (just) the end result (though that is of course a rewarding aspect). For example, jamming out to a kickin' song is fun, even if I'm just covering something. I also realized that my own ability to produce things isn't affected by this (as long as you don't want to make money on it). As someone who loves to play bass but is generally bad at writing bass riffs, I also see some fun potential to use AI to get bass tracks that go along with my main guitar riffs. I can always throw them out and rewrite from scratch later, or just iterate on them to get them where I like them. I do think I'll feel a bit lof a loss of "artistic purity" with doing something like that, but the more I think about it, the biggest reason that might bother me is because I'd feel judged by other musicians :-D
I am not sure if this will be any help to you at all, but the technical skill of music - especially in (from your comment) heavy metal - has only ever been a tiny part of it. You probably know a guy in a local pub band who can note-for-note play every singly thing that Toni Iommi ever recorded, for instance.
There are hundreds of bands who play three-chord doom or mindless-shredding grind who just learned one thing and do it well, and who play to hundreds of people multiple times a week (often including me). We go to see these bands not to see what they can play, but to see what they are saying with what they play.
This is why I feel that I can never describe LLM-generated content as 'art'. Art is about the story. People will go and see a punk band who only know three chords if they play songs about things that resonate with them. Bit of a tangent, but this is the same reason that I genuinely believe that if you could bio-engineer a steak that tastes exactly like one from a well-looked-after cow from a notable breed and a good farm, most people would still prefer the cow. The story matters - the fact that a person put effort and experience into something really is important.
"This solo is sick" is a fun plus-point, but it doesn't matter if the song doesn't mean anything to you. If proficiency was the only thing that mattered then we'd all be listening to the worst kind of prog.
Like you, I am just a hobbyist making beats in my room. No expectations of ever being a real musician. But when I'm jamming and I create a beat and synth line, start adding other instruments and really get a song going, there is a feeling that I get that an generated song will never ever ever be able to recreate for me. It's like a rush, an almost a euphoric tingling (and no I'm not on drugs) that happens that almost feels like a runners high. No output from a prompt-driven AI algorithm would ever do that to me. That's the value I see in making art for arts sake, for practicing a craft and for trying to just get better at something for the sake of getting better at it.
I can't see the future, but I imagine that the human art community may actually get more vibrant when divorced from being a way to make a living. Perhaps a return to something like a patronage system for the exceptional artists.
Open mics, music circles and concerts also remain untouched for the moment.
When Harmonix was in the concept phase for Guitar Hero, there were two slides in the presentation. The first was the "con" slide: novel game style risks not finding a market, technical challenges of the new style of gameplay, requiring peripherals for a game is a famous pathway to low-volume sales.
The next slide was labeled "pro," and it was just a picture of Jimi Hendrix on-stage mid-performance.
I'd submit to you the notion that even if the machine can create a billion billion iterations of music, it still cannot create what you will create, for the reasons you will create it, and that's reason enough to continue. Hendrix wasn't just "a guy who played guitar good." And a machine that could word-for-word and bar-for-bar synthesize "Foxy Lady" wouldn't be Hendrix.
Hendrix, also, can't be you. Nor you him.
It is embarrassing how much time I spent playing that dumb game instead of actually practicing a more versatile instrument.
I would suggest that instead of feeling demoralized by ai that you instead ask what can you offer musically that an ai cannot. I also would suggest trying to let go of the notion that music is primarily about achieving a certain level of technical proficiency. There are no limits on your growth musically unless you artificially constrain them because you are deluded by a technology to believe that you don't already have what you need inside of you.
Do you regularly play with other people? That is a good way to disabuse yourself of the notion that all that matters is technique.
What struck me more than the output is the absolute marketing talk of the prompting, so not only is the output kind of creepy but in order to get the best results you have to translate your intention into the worst corpo-speak imaginable.
Write a song about the billionaire social engineer Peter Thiel getting robbed and murdered by a broke minimum wage worker with no healthcare.
"That is not allowed by our terms of service"
I think the rebellious nature of art inherently has boundaries these people won't cross.
Did anyone compared to Suno quality wise? Seems like one benefit being the API availability
Not sure exactly what type of test is best to do, but I ran the same prompt with the same lyrics through both, very different results: 11labs - https://s.h4x.club/rRuR9W6p / Suno - https://s.h4x.club/bLuLedm6 - Prompt and lyrics: https://s.h4x.club/RBuDvEWm
It's interesting to me people really hate AI gen'd music, some people I know literally won't listen to it and just like... get kinda angry I even tried to show them what I made (so I stopped showing people for now). I have a good memory for sound, but whatever neurodiversity I have makes learning instruments (and math, and written words, and and) super challenging. I loved screwing around remixing stuff in Abelton, but I've not touched it since suno. I just looked at my suno library, I've gen'd 1,100 tracks since March 18th, 2024. The levels of audiophile or...musicophile? is pretty wide, but I don't think I'm a total n00b and I think some of the stuff I've made is actually pretty good, even if I do say so myself (and I'm fine even if I just make it for myself to listen to, as that's what I always did anyway). :)
> what I made
But did you really make it? Or was it the AI? If someone commissions a piece of art from an artist, I don't think the commissioner would be able to truthfully say they made it, even if they had a specific vision for what the piece of art should look like. But if you've edited or changed the track enough yourself, maybe it would be fair to call yourself a co-author...
I'm not going to defend against that position generally. Your point is well taken, I don't know if I put enough into it to consider it real to you. I think about the lyrics, I write the MVP of them by hand, and then I have 4.5 give me variations of them, I write all my instrumental prompts by hand I don't use any LLM for that because I typically know exactly what I want and I tend to be quite precise with my instructions there. Suno allows you a lot of fine grain control, and you can also create personas from tracks and then use those personas in new tracks so as you can imagine you can spend a long time trying to get exactly what you want... last night I gen'd over 60 songs before I got it dialed in how I wanted it and I'm still not totally happy with it. I understand the position though, I was on original team that took deviantart from zero to millions of artists through the early 2000s and have a few Emmy awards from the film to digital transition days, I'm no stranger to these conversations.
> Studio-grade
The vocals are definitely not that.
I am still bitter about them acquiring and shutting down Omnivore.
Do we yet have a way to "autocomplete" music from existing music? or from humming to music (music-to-music?)?
I am less interested in the "one-shot" approach here with text-to-prompt. I see seamless transitions but that seems like an afterthought.
I think AI will spark a revival in "organic" human culture and art, with people flocking to see real musicians play real instruments, and real artists using real materials.
It will probably also extinguish quite a few mad musicians and mediocre artists.
This music sounds like the visual equivalent of one of those motion settings on a tv that cause the "Soap Opera Effect".
I'm wondering if the LLM is failing to emulate the odd and even harmonics of real recordings and that's why it seems so unnatural. It makes me feel sick. It's different than the sterile straight to digital recordings of the late 80s early 90s too.
Do any of these let people download the individual track stems? I'd be curious if taking all the stems and running them through analog hardware would remedy it. The LLM needs a producer.
Suno lets you download the stems, also no affiliation at all just really like the tool and have spent a lot of time with it: Suno is night and day better than this product, by a mile. https://soundcloud.com/charlielabs/charlie (everything on this account I made with Suno, sometimes through Abelton if I'm not feeling lazy)
Nice. Charlie is the hero we’ve been holding out for. You’re cutting the stems up in Ableton and mixing with additional effects and processing? I don’t have a stance on making music with it will try out stems soon. Awhile back I gave o3 the strudel repl documentation and had it generate some really interesting experiments.
Honestly no, I downloaded the stems but never used them - their fine grain editor in the pro mode is... insanely good, I feel like I'm a secret shill here pawning for Suno but I'm 100% not - their editor really is magic when it works (about 75% of the time) - I use Abelton for just my regular post mix down i've done for years on all the stuff I make, so I basically just master there at this point.
> I'm wondering if the LLM is failing to emulate the odd and even harmonics of real recordings
What do you mean? If you screw up even/odd harmonics it will sound far less natural, like confusing flutes and clarinets.
I meant as an example if someone is singing, holding a long note and their voice is producing a lot of even harmonic overtones but every few cycles that overtone changes randomly flips or disappears. Sonically disoriented.
Eleven labs killed Omnivore. Never forget, never forgive.
Exactly. I don't want a music generation program anyway
This eats into musician incomes. Even if its terrible as the comments here suggest, it still eats into musician incomes.
Well, here's a few points of comparison:
Suno's version of Mandate of Heaven [0]. This is my baseline, it was generated with their v4 model and so far it has remained my favorite AI generated song. I regularly listen to this one track and it brings me joy. There's many places where I think it could be drastically improved, but none of the competitors have managed to surpass it nor have they provided tools to improve upon it. The pronunciation is a bit bad sometimes and it fails to hold notes as long as I wish, but overall it has gotten the closest to my vision.
Eleven Music's version of Mandate of Heaven [1]. They don't allow free accounts to export or share the full song so you can only try a small fragment. It has much crisper instruments and vocals, but it has terrible pacing issues and pronunciation. The track is 4 minutes long, but the singer is just rushing through the track at wildly unexpected speeds. I cannot even play the song after it finished generating, so I haven't even been able to listen to the whole thing, it just gets stuck when I press play. Maybe some kind of release-day bug. The only tool that Eleven Music gives you for refining and modifying sections is "Edit Style", which feels pretty limiting. But I can't even try it because the track won't play.
Producer.ai's version of Mandate of Heaven [2][3]. This one has slightly worse instruments than Eleven Music, but the vocals are a bit better than Suno v4. It also has severe timing issues. I tried asking it to generate the track without a vibe reference [2] and also with a vibe reference [3]. Both versions have terrible pacing issues; somehow the one with the vibe reference is particularly egregious, like it's trying to follow the input vibe but getting confused.
It feels like AI song generation is just in a really awkward place, where you don't get enough customization capabilities to really refine tracks into the perfect direction that you're imagining. You can get something sorta generic that sounds vaguely reasonable, but once you want to take more control you hit a wall.
If one is willing to bite the bullet, there's a paid program for generating high quality synthetic voices while maintaining fine-grained controls: Synthesizer V Studio 2. But I haven't been able to try it out because I'm cheap and there's no Linux support.
The ideal workflow I'm imagining would probably allow me to generate a few song variations as a starting point, while integrating into a tool like Synthesizer V Studio 2 so I can refine and iterate on the details. This makes a lot of sense too, because that's basically how we are using AI tools for programming: for anything serious you're generating some code and iterating on it or making tweaks for your specific program. I would like to specify which parts of the track are actually important to me, and which ones can be filled with sausage in reaction to my changes.
Overall, Eleven Music generates instruments that sounds nice, but the singing leaves a lot to be desired (n=1). Eleven Labs is doing a ton of great product work so I'm really excited for the direction they'll take this once they're able to iterate on it a few times. A very strong showing for an initial release.
[0] https://suno.com/s/HfDUqRp0ca2gwwAx
[1] https://elevenlabs.io/music/songs/TGyOFpwJsHdS3MTiHFUP
[2] https://www.producer.ai/song/aa1f3cc4-f3e4-40ce-9832-47dc300...
[3] https://www.producer.ai/song/3d02dd17-69f1-41ba-a3ea-967902f...
Mandate of Heaven does pretty much sound like a mid power metal track. The mixing is perfect but the production is "off" which is kinda weird, also the vocal delivery is flat and empty which is probably the biggest AI tell. The lyrics are cheesy but that's power metal for you.
Overall still quite impressive progress, though I'd prefer it if AI could remix existing artists songs that I already liked instead of being focused on tepid original content.
Can we get a new quantum leap in model quality to stave off the feeling that modern AI stuff is just a sloppy hustle for venture capital
The context length seems super small. The prompt needs to be shorter than 2k characters, which is super limiting. Hope they address this soon.
I've been waiting for this AI development for a while
It's amazing that the songs sound pretty natural
Finally non-repeating muzak is here
Behold, we are one step closer to a world where "AI will do the mundane tasks, and the we all are going to engage in creative hobbies such as music creation"... Oh... wait...
What did you train the model on? Assuming it was trained on all the music online, without permission, it's kinda shitty to take all musicians' work like that and use it to make a product to put them out of business.
I draw a distinction here from LLMs used for writing code. Plenty of people freely put their code on Github or in Stack Overflow with the expectation that people could freely copy and use it. Relatively little music is shared so permissively.
(And before someone cries "fair use", I'm making an ethical argument, not a legal one. A fair use defense can protect you from a copyright claim, but it doesn't make it any less shitty to take other people's art, blend it all together, and use the result to threaten their livelihoods.)
> Created in collaboration with labels, publishers, and artists, Eleven Music is cleared for nearly all commercial uses, from film and television to podcasts and social media videos, and from advertisements to gaming. For more information on supported usage across our different plans, head here.
Not sure what that actually means but it suggests they at least tried to get their training material legally.
A few thoughts...
First, where are you seeing that? And where does the "head here" link point? I'm not seeing it on the linked page or on https://elevenlabs.io/music-terms.
Second, that language is not saying "we got permission to use all the music we trained on". It seems more directly addressed at the rights they're giving you, as opposed to the rights they have to the training material. It could just as well be based on an opinion from their general counsel that it's fair use to train on all the music they downloaded, so it's "legal enough".
Third, and as I tried to emphasize above, legal is not the same as ethical. Even if they win the legal argument about fair use, it would still be shitty to take someone else's work and use it to train a model that put them out of business.
The silver lining to any of these models is that what AI generates is total shite.
Well, they've done it boys, they've made creative fulfillment obsolete. They've DISRUPTED the concept of going to big music festivals, and small cozy shows. Just plug your ear holes with the AI slop bucket's pure beeps and boops and never have to worry again about paying artists for music. You can pay a techbro instead.
This is like the dotcom era of where every idiotic idea that ended with, "but on the internet", would get a pile of cash thrown at it. We are officially at the beginning of the end. It's only going to get dumber from here.
There isn't any relationship whatsoever between a concert or music festival and this tech. Two entirely different experiences.
There will be fewer musicians in the world because of this tech. It eats into their income and forces fewer professionals to exist.
Fewer professionals... Does that mean fewer musicians?
lot of ai progress today, open ai, claude and now eleven labs
This stuff is such an abomination. What a terrible thing to release into the world.
Absolutely horrendous - well and truly. Ignoring the dystopic undertones of a product like this - the audio quality is low and filled with artifacts and destroyed transients. Voices are robotic and generic with basically zero emotion. Lyric generation is frankly embarrassing. Everything about this is horrible. Additionally the UI broke for me multiple times and I couldn't play the track without downloading it.
What is going on at ElevenLabs? Is everything vibe coded now? Is there nobody testing these products before pushing them out? This is the first time I'd characterize a product from a top AI company as complete slop from a conceptual and implementation perspective.
I feel like there's some sort of disconnect between the minds of tech CEOs and the general population's wants/needs especially in regards to creative domains. What kind of human wants this? Is there some grudge of tech people who never learned music/art/etc so their solution is to optimize it and create models so they can feel something?
Well, so much for culture. Every use case where you can plausibly use AI generated music removes one more method that provided an avenue to a reasonable career making music.
One more avenue of making something creative...
But at least there are jobs in tech!
Oh wait.
Lifeless music? Made of chunks of someone else's music, stitched together and polished to hide the sews. I think this is what gives all AI slop such an uncanny vibe.
Soylent Green is made of people!
Yet Another Slop Machine™