Balancing two very different remotely-rrec tracks to produce an even sound quality podcast mixdown
Hi Guys
We have used a remote online recording service (Squadcast) to conduct interviews for a new podcast that I have been editing. The advantage is that this gives us split tracks, which I can edit and process individually before doing my mixdown. The disadvantage is there is no embedded timecode, so the tracks can slip out of sync and I have to be very cognizant of my edits in the multitrack session. There is also the risk of a significant disparity in audio quality, with our local host track, recorded on a high-quality microphone and pro headphones, is much better than the remote subject, who may be using a crappy computer mike and headphones that allow some bleed-through from the host, which sounds like an echo. In trying to even these out, I fear that I am messing up the quality of the subject's voice. I have used Normalization, Match Loudness, and Speech Volume Leveler, as well as some RX7 tools like Voice De-click and de-reverb, but may have overdone that processing.
I would really welcome your guidance. I've enclosed a link here to the most problematic interview; am I being paranoid, or is this uneven and hard to listen to?
https://www.dropbox.com/s/zsfzbeut9pd3mzf/DALE_BELL.wav?dl=0
Thanks!
Rob
