How To Make an AI Voice Model That Actually Sounds Good
A practical guide to building AI voice models that don't suck. Covers vocal isolation, audio cleanup, training platforms, and the mistakes most people make on their first attempt.
Posted by
Related reading
How to Make Music With AI (2026 Guide)
A practical walkthrough of AI music generation in 2026. Covers Suno, Udio, ElevenLabs, songwriting tools, mixing workflows, and the lyric swap process.
The Best Way To Change A Vocalist With AI in 2026
I tested two methods for changing a song's vocalist with AI: Suno's persona feature vs RVC pre-processing. Note: Suno has since disabled personas for audio uploads. ChangeLyric V3 now offers a similar vocal cloning workflow.
How To Change Lyrics In a Song (3 Methods That Actually Work)
A step-by-step breakdown of how to change the lyrics in any song using AI tools, vocal cloning, and manual editing. Real workflows from someone who has done it 600+ times.

Voice cloning has gone from a research lab novelty to something you can do on a laptop for under ten bucks. The problem is that most people rush through the process and end up with a model that sounds like it was recorded inside a washing machine.
I have trained dozens of AI voice models for client projects through ChangeLyric's done-for-you service and built the vocal processing pipeline behind ChangeLyric. The difference between a good model and a garbage one almost always comes down to the same handful of mistakes.
This guide walks through the entire process from source audio to finished model. No fluff, no theoretical nonsense. Just what actually works.
Start Simple Before Going Deep
If you have never built a voice model before, do not start with a 20-minute training dataset and custom hyperparameters. Start by taking an existing vocal and running it through a pre-trained RVC model. This gives you instant feedback on what voice conversion actually sounds like.
Tools like Applio let you run inference on pre-trained models without training anything yourself. Download a model, feed it a vocal, and listen to the result. This teaches you what artifacts sound like and what clean output should be.
Once you understand the basics of voice conversion, building your own model becomes way less mysterious. You know what you are aiming for. I covered some of these voice conversion fundamentals in my post about the best way to change a vocalist with AI in 2026.
Choosing Your Source Audio
The quality of your model is capped by the quality of your training audio. Full stop. If you feed it noisy, reverb-drenched vocals with bleed from drums and guitars, your model will learn all of that noise as part of the voice.
You want dry, clean, isolated vocals. The ideal source is an acapella recording or a studio session where the singer recorded with no effects chain. But most of us do not have access to raw studio stems, so we work with what we have.
Here is what to look for in source tracks:
- Songs with prominent, upfront vocals and minimal layering
- Recordings with less reverb and delay (drier mixes are better)
- Multiple songs from the same artist to increase your dataset
- Tracks where the singer performs in the style you want to clone (breathy vs. belting matters)
That last point trips people up constantly. If you want to clone someone's powerful belt, do not train on their soft acoustic tracks. Models are style-specific. A Freddie Mercury model trained on soft ballads will not sound like Freddie Mercury doing rock anthems. Build targeted models for specific vocal deliveries.
Vocal Isolation: The Make-or-Break Step
Unless you are starting with raw acapella recordings, you need to isolate the vocals from the instrumental. This is the single most important step. Bad isolation equals a bad model. No amount of training will fix it.
Ultimate Vocal Remover (UVR) is the go-to free tool for this. It runs locally, supports multiple AI separation models, and gets better results than most cloud services. I recommend using the MDX-Net or Kim Vocal 2 models for initial separation.
Here is the trick that most tutorials skip: run the isolation twice. Take your separated vocal, then run it through a different model. The first pass removes the bulk of the instrumental. The second pass catches the residual bleed that the first model missed.
For paid options, LALAL.AI delivers excellent results with less effort. It handles edge cases better than UVR in some scenarios, particularly with heavy reverb or complex arrangements.
If you have done any work with removing explicit content from songs or vocal processing for lyric swaps, you already know how critical clean separation is. Same principle applies here.
Cleaning Up Your Isolated Vocals
Once you have your isolated vocals, do not just dump them straight into a training platform. Take ten minutes to clean them up in a DAW. This is boring work that pays massive dividends.
Import the isolated audio into any DAW (Ableton, Logic, Reaper, whatever you use). Then go through and remove:
- Silent sections longer than half a second
- Breath sounds that are louder than the singing
- Any remaining instrumental artifacts or clicks
- Sections where the isolation clearly failed (muddy or phasey audio)
If you have access to Clear by Supertone (around $70), it does an incredible job removing reverb and delay from isolated vocals. This is especially useful when your source tracks were mastered with heavy effects.
For levels, keep your vocals sitting between -10dB and 0dB peak. Apply a limiter if needed to catch stray peaks, but do not compress the dynamics out of the vocal. The model needs to learn the singer's natural dynamic range.
Export as both WAV (for quality) and MP3 (some platforms have upload size limits). Label your files clearly because you will be uploading multiple clips.
Ready to Transform Your First Song?
Join hundreds of music producers who are using ChangeLyric.
✓ Free trial available ✓ No content moderation ✓ Cancel anytime
Building a Better Dataset from Multiple Sources
Two to three minutes of clean audio is the minimum for training a usable model. But more data almost always equals better results, up to a point. Five to ten minutes of diverse vocal performances is the sweet spot.
Pull vocals from multiple songs by the same artist. This gives the model exposure to different vowel shapes, consonant patterns, and emotional deliveries. A model trained on a single song will overfit to that specific performance.
One technique I use: if you have isolated vocals from different songs that sound slightly different (one is drier, one has more room tone), you can mix them down to a consistent tonal profile using light EQ matching. This prevents the model from learning inconsistent room characteristics. Match the brighter recording to the darker one, not the other way around.
Choosing a Training Platform
For most people, cloud-based training is the way to go. You do not need a beefy GPU or any command-line experience. Upload your audio, click train, and wait.
Jammable (formerly Voicify) is the most beginner-friendly option right now. Upload your cleaned audio, set a few parameters, and the platform handles the rest. Training typically takes under an hour on their cloud GPUs.
For local training, Applio is the community standard. It is free, open source, and actively maintained. The tradeoff is that setup requires some technical comfort. You need a compatible NVIDIA GPU (8GB+ VRAM recommended) and a willingness to troubleshoot Python environments.
I wrote about how I used Applio through the Dione platform which makes local RVC training much easier than setting it up from scratch. If you want local control without the command-line hassle, that is a solid path.
Training Settings That Actually Matter
Most training parameters can stay at defaults and you will be fine. But a few settings make a real difference:
- Epochs: Start with 150-200 epochs. Going higher is not always better. Overtraining makes the model sound robotic and lose natural variation.
- Sample rate: Train at 48kHz if your source audio supports it. Downsampling to 32kHz or lower loses high-frequency detail that makes voices sound realistic.
- Batch size: Larger batch sizes train faster but use more VRAM. Start at 8 and increase if your GPU can handle it.
- Dataset length: If your cleaned audio is under two minutes, consider augmenting it by including slightly different EQ versions. This is a hack, but it works.
Monitor your training loss. If it flatlines early, you probably need more diverse training data. If it starts going back up after dropping, you are overtraining. Stop the run and use the checkpoint from the lowest point.
Testing Your Model (Inference)
Training is done. Now you need to test the model by running inference. This means feeding it a vocal it has never heard and seeing how well it converts.
Pick a test vocal that matches the style your model was trained on. If you trained on pop vocals, test with pop vocals. Do not test your pop model with a death metal growl and wonder why it sounds bad.
During inference, adjust the pitch shift and index ratio. The index ratio controls how much the model relies on its training data versus the input pitch. Higher values produce a closer voice match but can introduce more artifacts. Start at 0.5 and adjust from there.
If your converted vocal has a weird warble or metallic quality, it usually means one of three things: not enough training data, too many epochs (overtraining), or your source audio had artifacts the model learned. I discuss these kinds of vocal artifacts in depth in my article about why AI vocals fail to match the original singer.
The Mistakes Everyone Makes
After watching dozens of people go through this process, the same errors come up repeatedly. Avoid these and you are already ahead of most people building voice models.
Using Dirty Source Audio
This is the number one killer. People grab a YouTube rip with compression artifacts, run a mediocre vocal isolation, and wonder why their model sounds metallic. Garbage in, garbage out. Spend the extra time finding clean sources and running proper isolation.
Training a Generic Model
A single model cannot capture every vocal style a singer has. Whitney Houston whisper-singing is a fundamentally different voice from Whitney Houston belting. Train separate models for separate styles if you need both. It is more work but the results are dramatically better.
Overtraining
More epochs does not mean better results. At some point the model starts memorizing your specific training clips instead of learning the general voice characteristics. When this happens, inference output starts sounding flat and lifeless. Watch your training loss and stop when it plateaus.
Skipping the Cleanup Step
I get it, trimming silence and removing artifacts is tedious. But every second of garbage audio in your training set degrades the model. Ten minutes of cleanup saves hours of retraining. If you are serious about quality, do not skip this.
Beyond Voice Models: What Comes Next
Once you have a working voice model, the question becomes what to do with it. The most common use cases I see are vocal covers, content creation, and custom song projects.
For lyric swapping specifically, having a custom voice model is powerful but not always necessary. ChangeLyric handles vocal replacement and lyric changes without requiring you to build a model first. The platform uses its own processing pipeline to match vocals to your new lyrics.
If you want to combine a custom voice model with lyric changes, the workflow is: generate your lyric swap, extract the AI vocal, run it through your RVC model, and comp the result back into the instrumental. I walked through a similar workflow in my post about getting started with changing lyrics in songs.
For personalized projects like custom birthday songs or event music, a voice model trained on a friend or family member adds a personal touch that generic AI voices cannot match.
The Real Cost Breakdown
You can build a usable voice model for under ten dollars. Here is the realistic breakdown:
- Vocal isolation: Free with UVR, or a few dollars with LALAL.AI credits
- Audio cleanup: Free if you already have a DAW (Reaper is free to evaluate)
- Training: Free locally with Applio if you have a GPU, or pay-per-model on Jammable
- Optional reverb removal: ~$70 for Clear by Supertone (one-time purchase)
The expensive part is your time, not the tools. Expect to spend two to four hours on your first model from start to finish. That time drops significantly once you have a workflow dialed in. For a deeper look at what professional lyric swapping costs, check out my breakdown of how much it costs to change song lyrics.
Bottom Line
Building a good AI voice model is not hard. It is just tedious in the right places. Clean source audio, proper isolation, careful cleanup, and sensible training parameters. That is 90% of the battle.
Skip the shortcuts on audio quality. Be style-specific with your training data. Do not overtrain. Test thoroughly before using the model in a real project.
The tools are accessible and mostly free. What separates good models from bad ones is the attention to detail in the preparation steps, not the training itself. Do the prep work right and the training practically takes care of itself.
Copyright Reminder
Training voice models on copyrighted performances raises legal questions that are still being settled. Commercial rights from AI platforms only apply to ORIGINAL songs they generate. Cloning a real artist's voice for commercial use without permission carries real legal risk. Personal use exists in a legal gray area. Users are responsible for understanding applicable laws in their jurisdiction.
Frequently Asked Questions
A minimum of two to three minutes of clean, isolated vocals. For better results, aim for five to ten minutes from multiple songs by the same artist. More diverse training data helps the model generalize across different vowels, consonants, and delivery styles rather than overfitting to a single performance.
Ultimate Vocal Remover (UVR) is the community standard. It runs locally, supports multiple AI separation models, and produces results competitive with paid services. Use the MDX-Net or Kim Vocal 2 models, and run isolation twice through different models for the cleanest results.
Usually one of three causes: dirty source audio with artifacts the model learned, overtraining (too many epochs causing the model to memorize clips instead of learning the voice), or a mismatch between training style and inference style. Clean your audio better, reduce epochs, or train a style-specific model.
Yes, but it is a multi-step process. Generate your lyric swap (ChangeLyric handles this), extract the AI vocal, run it through your RVC model for voice matching, then comp the result into the original instrumental. Or skip the custom model entirely and use ChangeLyric's built-in vocal processing pipeline.
The legal landscape is still evolving. Training models for personal, non-commercial use exists in a gray area. Commercial use of a cloned voice without the artist's permission carries significant legal risk, especially in jurisdictions with right-of-publicity laws. Always check your local regulations and when in doubt, get permission.
Cloud platforms like Jammable are easiest for beginners with no setup required. Local training with Applio gives you more control and is free if you have an NVIDIA GPU with 8GB or more VRAM. Local training is better for batch processing and iterating on settings. Start with cloud to learn, then move local if you need more control.
Want to Change Lyrics Without Building a Model?
ChangeLyric handles vocal replacement and lyric swapping without requiring custom voice models. Upload a song, edit the lyrics, and get your modified track back. No GPU, no training, no command line.
Try ChangeLyric Free