hi all,
we are working on a stream solution based on Janus as SFU/MCU and
coturn as our TURN servers. Currently, the setup is hosted on AWS.
The video-source to destination is one-way, and data-channel is two-way.
our current latency is ~400ms, which is alright for now, as the source-to-destination distance is half-the-globe.
But the challenge we are facing that on source there's usually two persons speaking, one is loud while another one is too quiet due to distance from the single mic. In the source scenario, we have very little control over the environment. However, on the receiving end, both voices are important. So, i think, we need to apply Dynamic range compression(DRC) or, volume normalization on the audio stream to make both voices reasonably loud, at least bring the difference down.
How to approach the problem?
Does it make sense to do that on MCU (through Janus plugin?) or to do it on the client-side (browser)? If on the MCU, what impact it may have on the latency?
as a note, we are currently using VP8+PCM encoding to make the RTP stream.
any pointer or suggestion will be highly helpful.
// khaled