Honest opinions?
I want to release my songs with my own singing voice. However although I am usually a decent singer in terms of amdram and on stage and in pubs, my singing in recording is not great. I get choked up and my recordings sound cringe. For starters I don't have great recording equipment. No matter what editing I do in Bandlab, when i record songs, it just sounds awful to me (I'm using my phone to record, my mic is even worse). It could be I'm my own worse critic but I'm worried I'll ruin my own songs with my own voice.
So I thought about voice cloning but both Weights and Jammable, I dunno, just sounds so.....fake....I can tell so easily its not really me. Its my voice but something is off. Someone suggested I try Kits which is way more expensive. Does Kits really sound realistic? It looks more expensive than Jammable and Weights but I worry I don't wanna spend a fortune on Kits and then find out it still sounds so obviously fake. My friends and family I wanna share my songs with I don't want it to be too obvious its not me really singing it. How is people's experience with voice cloning from Kits?
Videos
Using Udio with Kits.ai
Name origin for Kit's AI, Western Amber. : Kitboga
[Creation Kit] NPC not responding to AI Package
Arpeggi Labs launches Kits AI for artists to license AI voices
Gotta check that out
More on reddit.comI saw a video on YouTube of someone showing this workflow (i.e. using Kits.ai to replace the vocal generated in Udio or Suno), but so far I'm not finding it at all usable. Even when I separate the stems on kits.ai and then do the conversion from the vocal track by itself, the vocals results in awful glitches. Kits.ai just doesn't read the base audio cleanly enough. It's not just a problem with my voice model, as their own royalty-free models have the same issue.
Is anyone else using this workflow, and if so, how are you managing to get something that works? I did try separating stems outside of kits.ai, but so far the results don't appear to be better.
I do get somewhat better results with conversion using the Replay app instead of Kits.ai, but it is extremely slow. It took a lot of hours to train my first model based on just a couple minutes of audio (much longer than it took kits.ai to train on the ~20 minute audio file I recorded later), and then it's extremely slow processing the conversion once that's done. (I'm sure this is related to the fact that I don't have a really high end GPU.)