On Wed, Jun 18, 2014 at 11:03 AM, AirMike <
airm...@gmail.com> wrote:
> I have task of recording webrtc local stream audio and webrtc remote
> stream audio.
> I succeeded in recording audio with MediaRecorder (using timeSlice) in
> which I get recorder chunk Blob (audio/ogg) for local and remote audio.
>
> Now, before I send this to server I would like to mix recorded local and
> remote audio chunks to one chunk using Web Audio API and this is where I
> have some problems.
>
It sounds like you're using MediaRecorder to compress the local and remote
audio chunks on the client, and then trying to uncompress them on the
client, mix them, recompress them and send the result to the server. Is
that right? If so, why are you doing the first compression step instead of
just leaving them uncompressed?
I used this steps:
>
> 1. when both local and remote audio blobs are available I'm using
> FileReader to get ArrayBuffer for each
>
> 2. using AudioContext decodeData to get AudioBuffer (here I get error:
> The buffer passed to decodeAudioData contains an unknown content type. and
> The buffer passed to decodeAudioData contains invalid content which cannot
> be decoded successfully.)
>
Is your initial compression step using timeSlice to produce multiple Blobs
from a single MediaRecorder? If so, those Blobs must be concatenated to get
a single resource which you can decode successfully. E.g. you can't pass
just the second Blob created by a MediaRecorder to
AudioContext.decodeAudioData and expect it to work.
Rob
--
Jtehsauts tshaei dS,o n" Wohfy Mdaon yhoaus eanuttehrotraiitny eovni
le atrhtohu gthot sf oirng iyvoeu rs ihnesa.r"t sS?o Whhei csha iids teoa
stiheer :p atroa lsyazye,d 'mYaonu,r "sGients uapr,e tfaokreg iyvoeunr,
'm aotr atnod sgaoy ,h o'mGee.t" uTph eann dt hwea lmka'n? gBoutt uIp
waanndt wyeonut thoo mken.o w