similar to: Anyone knows how microsoft AEC can deal with mismatches between clocks of capture and render streams?

Displaying 20 results from an estimated 4000 matches similar to: "Anyone knows how microsoft AEC can deal with mismatches between clocks of capture and render streams?"

2011 Apr 12
1
Anyone knows how microsoft AEC can deal with mismatches between clocks of capture and render streams?
Hi Shridhar, Sample rate conversion is not enough to solve this problem. I have tried this method several months ago. The first step is to measure the difference between sample rate of capturing and rendering. Then resampling (by what you said "sinc interpolation") one signal to eliminate the difference. The frequency step in my experiment is less than 0.1Hz. I have tried speex AEC
2011 Apr 13
1
Anyone knows how microsoft AEC can deal with mismatches between clocks of capture and render streams?
On 04/13/2011 02:58 AM, Shridhar, Vasant wrote: > I am doing this right now with no problem. I am not using speex for this at the moment though. Group delay is the biggest problem. I implemented a version where the input and output sample rates are known up front. The routine than interpolates between the jitter. This should solve the problem. The crystals used to clock the input and
2011 Apr 15
0
Anyone knows how microsoft AEC can deal with mismatches between clocks of capture and render streams?
On 04/14/2011 07:26 PM, LiMaoquan2000 wrote: > Hi All, > Many Thanks to Underwood for her excellent review of our big trouble > which prevent LMS-based AEC algorithms to be used in most computer. > Maybe it can be summaried as follows: > 1. Different sample rate of sampling and rendering does exists in most > low-cost soundcards (In my experiments over more than 20 soundcards,
2011 Apr 12
0
Anyone knows how microsoft AEC can deal with mismatches between clocks of capture and render streams?
I am doing this right now with no problem. I am not using speex for this at the moment though. Group delay is the biggest problem. I implemented a version where the input and output sample rates are known up front. The routine than interpolates between the jitter. This should solve the problem. The crystals used to clock the input and output have very fine tolerances on most standard audio
2011 Apr 17
0
Speex-dev Digest, Vol 83, Issue 10
Hi Steve, Have you read this paper? (Heping Ding, David I. Havelock, Drift-Compensated Adaptive Filtering for Improving Speech Intelligibility in Cases with Asynchronous Inputs. EURASIP J. Adv. Sig. Proc. 2010:) Let me call is paper-Drift. It provided a method to evaluate Relative Sample Offset (RSO, d[i]) which is omitted in the microsoft paper (Challenges and Solutions for Designing Software
2011 Apr 16
0
Speex-dev Digest, Vol 83, Issue 10
Hi Steve, > I don't know if this has only recently been put on line, but I never > noticed it until today - > www.iwaenc.org/proceedings/*2008*/contents/papers/9044.pdf > > That paper is from people at MS describing, in some detail, what the > Windows kernel echo canceller does to handle synchronisation issues. It > tracks both time varying sample clock drift and hiccups
2011 Apr 12
4
Anyone knows how microsoft AEC can deal with mismatches between clocks of capture and render streams?
Hi all, We all know that mismatch between clocks of ADCs of far-end voice and near-end voice is not allowed in a time-domain or frequency-domain LMS based AEC system. It means that capture and render audio streams must be synchronized to a same sample rate. However, I found that this restriction is removed in microsoft AEC from Windows XP SP1. Anyone knows how microsoft AEC do it? This technology
2011 Apr 11
2
lpcSize
Okay, Not exactly the answer I was looking for. This sounds like a big change. I don't mind re-writing the LSP quantizer but re-training code books and breaking compatibility is not what I want to do. I am working on an optimization for an ARM cortex-A8. It is desirable to process things in 4 element blocks. Is there a simpler approach you could recommend? Vasant Shridhar
2011 Apr 11
0
lpcSize
On 11-04-11 06:48 PM, Shridhar, Vasant wrote: > Okay, > > Not exactly the answer I was looking for. This sounds like a big > change. I don't mind re-writing the LSP quantizer but re-training > code books and breaking compatibility is not what I want to do. I am > working on an optimization for an ARM cortex-A8. It is desirable to > process things in 4 element blocks.
2011 Jan 19
3
About Sampling Rate Correction in acoustic echo cancellation
Hi all, We have discussed so many about sampling rate asynchronous (or offset) between rendering (D/A converter) and capturing (A/D converter) of most PC soundcards. It seems all acoustic echo cancellers, include AEC in speex, can not deal with this trouble, because it causes a drift of echo path and also buffer overflow and underflow which jumps the delay of echo path seriously. Unfortunately,
2011 Jul 11
0
Playing captured speex frames
Speexdec plays ogg files. You can simply modify it to play your raw file or even rtp capture. Jozsef -----Original Message----- From: speex-dev-bounces at xiph.org [mailto:speex-dev-bounces at xiph.org] On Behalf Of speex-dev-request at xiph.org Sent: Saturday, July 09, 2011 12:00 PM To: speex-dev at xiph.org Subject: Speex-dev Digest, Vol 86, Issue 3 Send Speex-dev mailing list submissions to
2006 Jun 26
1
Re: AEC frame size
Jean-Marc Valin <Jean-Marc.Valin <at> USherbrooke.ca> writes: > > Le mercredi 07 juin 2006 ? 10:03 +0000, shridhar desai a ?crit : > > > > hi all, > > i am using the Acoustic Echo Cancellation from "Speex 1.1.12 version" > > in my VOIP application. Is it that the frame length to be chosen > > should always be 20ms or can i have
2009 Jul 06
2
AEC with different soundcards
The problem with different sound cards is that their clocks are not usually synchronized, and therefore the clock drift adds a non-linear factor to the audio path. The AEC can only cancel linear changes to the audio path, and so the AEC never converges.One solution is to measure the clock drift and resample either the input or output signal so that they *are* synchronized, and then the AEC
2009 Jul 07
1
AEC with different soundcards
AFAIK, that's a common point for all AECs. But some of them solve the problem by resampling on of the end to keep it in sync with the other. On Tue, Jul 7, 2009 at 5:14 PM, ggb<ggb at tid.es> wrote: > Thank you John. > > On 07/06/2009 11:03 PM, John Ridges wrote: > > ly synchronized, and therefore the clock drift adds a non-linear > factor to the audio path. The AEC
2007 Jul 22
2
Server Side AEC
Hi Jean-Marc, Regarding you points: 1) Is it ok if the audio is encoded (using Nelly Moser ASAO) and sent to the client and decoded when it is recevied so the AEC is always performed on raw PCM16 8KHZ ? 2) The audio is moved in 32ms (512 byte) chunks and the reading and writing to the AEC code will be done by separate threads at regular 32 ms intervals. 3) Occasionaly audio is
2007 Jul 22
1
Server Side AEC
The client is the adobe flash player. No install and on 98% of all desktops but we can't change it. It works ok if people use headphones but we need to stop the howl than can build up if more than one person in a conference has mic to close to speakers. Any ideas? Jean-Marc Valin <jean-marc.valin@usherbrooke.ca> wrote: > 1) Is it ok if the audio is encoded (using
2011 Apr 11
2
lpcSize
I would like to make lpcSize a multiple of 4. In the current library for narrow band mode lpcSize is currently set to 10. I would like to increase this 12 for optimization reasons. Is it a simple matter of just changing lpcSize in the structures in mode.c or are there other implications to doing this? I did try and change the value but I seem to be getting some runtime errors indicating a
2010 Jun 09
3
Sound card problem in acoustic echo cancellation
Then why ONE sound card have different capture and playback rate? It must be ONE single physical clock generator which is used by both ADC and DAC in the sound card, isn't it? If you are a hardware engineer. Will you design two different physical clock for ADC and DAC seperately? What on earth causes this problem? Who knows its intrinsic real reason? Isn't there any other solutions? For
2011 Apr 21
3
Acoustic echo cancellation
Simply to say, in a quiet room, you can play a impulse signal and then find it's impulse response signal from the microphone. For example, if the delay between the impulse signal and its response signal range from 500 to 3000 cycles, you can buffer the far-end signal to 0-300 cycles and set the filter length to 4000. It is also called to align far-end signal and near-end signal. BTW: Speex
2007 Jul 20
2
Server Side AEC
Hi, I am looking for AEC software which can be run on the server side. This means there will be a fairly constant 600ms or so gap between sending out an audio frame and getting it back with echo. Could Speex AEC be configured to handle these conditions? If so, how good can I expect it to be? Thanks --------------------------------- Yahoo! Mail is the world's