timlod
a year ago
The title is a bit confusing as open-source separation of ... reads like source separation, which this is not. Rather, it is a pitch detection algorithm which also classifies the instrument the pitch originated with.
I think it's really neat, but the results look like it could take more time to fix the output than using a manual approach (if really accurate results are required).
earthnail
a year ago
Thanks for clarifying.
In fairness to the author, he is still at high school: https://matthew-bird.com/about.html
Amazing work for that age.
TazeTSchnitzel
a year ago
Is “source separation” better known as “stem separation” or is that something else? I think the latter term is the one I usually hear from musicians who are interested in taking a single audio file and recovering (something approximating) the original tracks prior to mixing (i.e. the “stems”).
timlod
a year ago
Audio Source Separation I think is the general term used in research. It is often applied to musical audio though, where you want to do stem separation - that's source separation where you want to isolate audio stems, a term referring to audio from related groups of signals, e.g. drums (which can contain multiple individual signals, like one for each drum/cymbal).
Earw0rm
a year ago
Stem separation refers to doing it with audio playback fidelity (or an attempt at that). So it should pull the bass part out at high enough fidelity to be reused as a bass part.
This is a partly solved problem right now. Some tracks and signal types can be unmixed easier than others, it depends on what the sources are and how much post-processing (reverb, side chaining, heavy brick wall limiting and so on)
dylan604
a year ago
> This is a partly solved problem right now.
I'd agree with the partly. I have yet to find one that either isolates an instrument as a separate file or removes one from the rest of the mix that does not negatively impact the sound. The common issues I hear are similar to the early internet low bit rate compression. The new "AI" versions are really bad at this, but even the ones available before the AI craze were still susceptible
mh-
a year ago
I'm far (far) from an expert in this field, but when you think about how audio is quantized into digital form, I'm really not sure how one solves this with the current approaches.
That is: frequencies from one instrument will virtually always overlap with another one (including vocals), especially considering harmonics.
Any kind of separation will require some pretty sophisticated "reconstruction" it seems to me, because the operation is inherently destructive. And then the problem becomes one of how faithful the "reproduction" is.
This feels pretty similar to the inpainting/outpainting stuff being done in generative image editing (a la Photoshop) nowadays, but I don't think anywhere near the investment is being made in this field.
Very interested to hear anyone with expertise weigh in!
nineteen999
a year ago
I won't say expertise, but what I've done recently:
1) used PixBim AI to extract "stems" (drums, bass, piano, all guitars, vocals). Obviously a lossless source like FLAC works better than MP3 here
2) imported the stems to ProTools.
3) from there, I will usually re-record the bass, guitars, pianos and vocals myself. Occassionally the drums as well.
This is a pretty good way I found to record covers of tracks at home, re-using the original drums if I want to, keeping the tempo of the original track intact etc. I can embellish/replace/modify/simplify parts that I re-record obviously.
It's a bit like drawing using tracing paper, you're creating a copy to the best of your ability, but you have a guide underneath to help you with placement.
Earw0rm
a year ago
It's not really digital quantisation that's the problem, but everything else that happens during mixing - which is a much more complicated process, especially for pop/rock/electronic etc., than just "sum all the signals together".
There's a bunch of other stuff that happens during and after summing which makes it much harder to reliably 100% reverse that process.
mh-
a year ago
I didn't mean to say that quantization was the problem, just that you're basically trying to pick apart a "pixel" (to continue my image-based analogy) that is a composite of multiple sounds (or partially-transparent image layers).
I was sincere when I said:
> I'm really not sure how one solves this with the current approaches.
I was hoping someone would come along and say it is, in fact, possible. :)
popalchemist
a year ago
Source separation is a general term, stem separation is a specific instance of source separation.
emptiestplace
a year ago
No, it doesn't read like that. The hyphen completely eliminates any possible ambiguity.
ipsum2
a year ago
The title of the submission was modified. It you read the article it says:
Audio Decomposition [Blind Source Seperation]
croes
a year ago
Maybe added later by OP? Because there is no hyphen in the article’s subtitle.
>Open source seperation of music into constituent instruments.
emptiestplace
a year ago
The complaint:
> The title is a bit confusing as open-source separation of ... reads like source separation, which this is not.