Be advised: this thread is not for serious in-depth discussion of weighty topics (we have a link for that), this thread is not for anything Culture War related. This thread is for Fun. You got jokes? Share 'em. You got silly questions? Ask 'em.
- 111
- 1
What is this place?
This website is a place for people who want to move past shady thinking and test their ideas in a
court of people who don't all share the same biases. Our goal is to
optimize for light, not heat; this is a group effort, and all commentators are asked to do their part.
The weekly Culture War threads host the most
controversial topics and are the most visible aspect of The Motte. However, many other topics are
appropriate here. We encourage people to post anything related to science, politics, or philosophy;
if in doubt, post!
Check out The Vault for an archive of old quality posts.
You are encouraged to crosspost these elsewhere.
Why are you called The Motte?
A motte is a stone keep on a raised earthwork common in early medieval fortifications. More pertinently,
it's an element in a rhetorical move called a "Motte-and-Bailey",
originally identified by
philosopher Nicholas Shackel. It describes the tendency in discourse for people to move from a controversial
but high value claim to a defensible but less exciting one upon any resistance to the former. He likens
this to the medieval fortification, where a desirable land (the bailey) is abandoned when in danger for
the more easily defended motte. In Shackel's words, "The Motte represents the defensible but undesired
propositions to which one retreats when hard pressed."
On The Motte, always attempt to remain inside your defensible territory, even if you are not being pressed.
New post guidelines
If you're posting something that isn't related to the culture war, we encourage you to post a thread for it.
A submission statement is highly appreciated, but isn't necessary for text posts or links to largely-text posts
such as blogs or news articles; if we're unsure of the value of your post, we might remove it until you add a
submission statement. A submission statement is required for non-text sources (videos, podcasts, images).
Culture war posts go in the culture war thread; all links must either include a submission statement or
significant commentary. Bare links without those will be removed.
If in doubt, please post it!
Rules
- Courtesy
- Content
- Engagement
- When disagreeing with someone, state your objections explicitly.
- Proactively provide evidence in proportion to how partisan and inflammatory your claim might be.
- Accept temporary bans as a time-out, and don't attempt to rejoin the conversation until it's lifted.
- Don't attempt to build consensus or enforce ideological conformity.
- Write like everyone is reading and you want them to be included in the discussion.
- The Wildcard Rule
- The Metarule
Jump in the discussion.
No email address required.
Notes -
The first song I consider to be really bad.
The second has a nice bass line that I would be happy to steal wholesale. Much more impressive.
The third one didn't seem any good, either. Sure, it's a simple beat and chord changes, but that's much less impressive than a melody, or funky base line.
Can it output sheet music? Can it output notation? Part of the problem is the synthesizer isn't that great. Ideally it would output notation is something like Frescobaldi.
Can it write a melody for me? Can I give it a melody and have it write counterpoint? While it's interesting to give it words and get out sounds, I'm more interested in including music (notation) and getting music (notation) back.
No, and it fundamentally can't right now. Those models are trained on raw music, not on notation. During creation, the model isn't "composing" like a human would, in the same sense that an image model isn't actually sketching, drawing and painting - the final image is directly condensed from the diffusion process.
But this is clearly the next step in the value chain. Once audio creation models can input and output notation, they will completely change the creative process - in the same way that video models will become valuable once they can input and output an entire 3D scene into/from Blender. But this step is difficult, there is orders of magnitude less training data in all those cases (you need specific sets of music + notation, video + 3D models, ect.
Music is, of course, simpler than 3D in this aspect. You can run AI audio creation through the usual transcription aids or quickly rebuild a beat you like in Abelton by ear/hand.
Do you know how viable it would be for an AI model to be able to "reverse engineer" sheet music from an audio file? Knowing very little about music myself, my intuition is that one could train a model with lots of sheet music-audio file pairs and then feed it the latter to generate the former, but I could easily be missing some hurdle that would prevent this from being viable.
Yeah, that's the way. Once you run out of training data, you can probably also do self learning by transcribing music without available sheet music, transforming the generated notation into sound through a synthesizer, compare the results (this needs another model) and then try again. Once you run out of music, you can continue with synthetic data (since current models can already make fresh sound files of high enough quality).
The devil is in the details, of course, e.g. current software transcription aids work much better for solo piano than for any other instruments (there not many different ways to modify the sound of a note on a piano). Guitars, on the other hand, are notoriously hard to transcribe. They kind of make up for it by having tabs available for a million songs, so at least there's a lot of training data. But the relationship between tabs and final sound is much less straight forward than for piano.
More options
Context Copy link
More options
Context Copy link
Text -> Image -> Spritesheet -> 3D Model models are actually already here. They're just pretty bad at giving you usable topology, but you'll probably begin seeing AI generated assets in production games inside a few months. Not big or moving stuff, but static medium poly assets like crates or plants.
There's a few on huggingface, and an integration called BlenderGPT they're working on.
More options
Context Copy link
More options
Context Copy link
Different strokes I guess. I'm also not primarily evaluating if the exhibited technical/music theory prowess of the songs in question are particularly impressive - most music isn't particularly rich in complex composition, and mediocre music is inevitably going to represent a large part of Udio's dataset. I consider all of the linked songs to be about on par with a lot of the music that gets released. Instead, I’m evaluating on the basis of “could this be a song that I’d hear out in the wild?”
Ideally, that'd be the goal of a machine learning-driven plugin. Unfortunately I'm not aware of any notation-producing ones worth their salt yet, but I do know that there are a number of very competent plugins which have focused on the generation of sound design.
More options
Context Copy link
More options
Context Copy link