NewAgeRipper
Superstar DJ
Did you leave it going overnight and how were the results?
Everything's fine brother. Even with country songs that it hasn't even been trained for yet.
Did you leave it going overnight and how were the results?
[MENTION=39673]Anjok[/MENTION] Will we see the baseline updated to use commands or have menu option in the GUI for specific stems? Like drums, bass, guitars, keys, etc? Yes lower end machines will take longer but it'll damn be worth it.
I don't think this exact method will work for separate stems, as it's only used for instrumentals. In that case, use Demucs - it's by far the best tool to extract bass, drums, synth and vocals.
Can you point me there? I currently use RX7. But if it's better than that I'll give it a whirl.
I can make a guide on how to install it on Google Colab, since it won't require coding there, otherwise this is the GitHub page and there are a few examples and comparisons https://github.com/facebookresearch/demucs
It's my go-to way for DIY stems. Downside is on some tracks it can ignore background vocals and leave them as synths, rather than part of the vocals track. But depends from song to song.
I can make a guide on how to install it on Google Colab, since it won't require coding there, otherwise this is the GitHub page and there are a few examples and comparisons https://github.com/facebookresearch/demucs
It's my go-to way for DIY stems. Downside is on some tracks it can ignore background vocals and leave them as synths, rather than part of the vocals track. But depends from song to song.
I need the background vocals anyways. LOL.
Anjok........is there anyway that the AI has been used on a speaking part of a song......
does it struggle with just plain talking as opposed to singing due to the amount of reverb that some singers use
Just speaking from my own experiences here, I have found it handles speaking exceptionally well in general.
I've heard some fantastic results on rap/hip-hop tracks etc. The flip side is that when the instrumentation is bare, every tiny little missed detail stands out. It's easier to scrub in spectral editing since there's not much sound on the spectrum to dig through ... but it all sticks out.
But really, since it's trained on voice, so much depends on how well it recognizes a specific type of voice sound, and how much it mistakes certain instrumentation for voice. That's why having multiple models could potentially prove very useful. Reverb kinda fits into that category as well. This model is trained on different music than the primary AI over in the other thread. The primary one can't handle reverb nearly as well as this one.
I will be doing A LOT of experiments. This will take some time!
Just speaking from my own experiences here, I have found it handles speaking exceptionally well in general.
I've heard some fantastic results on rap/hip-hop tracks etc. The flip side is that when the instrumentation is bare, every tiny little missed detail stands out. It's easier to scrub in spectral editing since there's not much sound on the spectrum to dig through ... but it all sticks out.
But really, since it's trained on voice, so much depends on how well it recognizes a specific type of voice sound, and how much it mistakes certain instrumentation for voice. That's why having multiple models could potentially prove very useful. Reverb kinda fits into that category as well. This model is trained on different music than the primary AI over in the other thread. The primary one can't handle reverb nearly as well as this one.
**UPDATE**
I've made a lot traction on the GUI and should have it released by the end of the first week of June, along with an updated model that so far will be the best one I've made so far. On par with the one used to create the instrumentals in the other thread!
Can you link to it when this is ready? Would greatly appreciate it.I can make a guide on how to install it on Google Colab,
[MENTION=39673]Anjok[/MENTION] would it be possible to train a model to preserve backing vocals within allowed boundaries? I work with music on the karaoke side of things and sing and record. Backing vocals would be very helpful. I realize they can't be preserved in all songs due to how the mix is done. but your A.I. has really helped to make instrumentals from older tracks I thought would never be possible. It could essentially just be another command that tries to filter and export another track but instead of being acapella it's the backing vocals.
That's a good question and it's something I would have to experiment with. I think one way I can potentially make this happen is to train the AI on a dataset consisting of only full mixes paired with their official TV track counterparts. To have a model as effective as the ones I've shared, it would have to consist of at least 200 pairs as well. The GUI that's being developed will have a drop-down of models to choose from, so you'll be able to toggle between a karaoke model and a full vocal removal model.