Once upon a time, I believed that RMS normalization was the holy grail.
I love fat harmonies that blend so well that you can't even pick out individual voices. (For reference, listen to the chorus on "
Sweet Little Lies" by Fleetwood Mac.) For a long time it was a mystery to me how they achieved that. Logically, it made sense that if each voice was RMS-normalized to one another, that would do the trick.
So I took a short piece out of a 4-part harmony section and very carefully adjusted their average RMS values - by hand - to within 0.5 dB of one another. Looking back, what I discovered should have been obvious, but nevertheless surprised me at the time. The vocal balance was
worse than when I'd just tweaked them by ear. Duh. We don't perceive all frequencies the same, even if they're RMS-matched. A high harmony part might have to be as much as 6 dB below the main part to sound balanced.
I didn't give up, though, but kept on trying stuff and reading about classic techniques. In the intervening years I've figured out that there simply is no way to automate this kind of thing. Turns out, vocal levels are highly subjective, depending on the vocalist's timbre, pitch, even lyrical content. And of course, what's going on behind the vocals that may or may not be masking them. This is why any automated leveling is only ever going to be partly successful, and will never obviate the need for manual tweaks.
So how did they do it in "Sweet Little Lies"? Heavy compression, triple-tracking and manually-programmed volume automation.