From: maury on
On Jul 14, 9:35 am, Vladimir Vassilevsky <nos...(a)nowhere.com> wrote:
> maury wrote:
> > If you look at the normalized LMS (NLMS), then the maximum mu is
> > limited to the sum of the number of filter coefficients.
>
> IMO what does matter is optimim rather then maximum. You need to
> minimize the total error of the process. This error is because of
> imperfect adaptation, noisy gradients, nonlinearity, ambient noise and
> numeric artifacts. It also depends on the statistics and power of the
> reference signal. So the algorithm must adapt Mu in pseudo Kalman way.
>
> Vladimir Vassilevsky
> DSP and Mixed Signal Design Consultanthttp://www.abvolt.com

Hi Vlad,
With respect to the NLMS, the effect of the power of the reference
depends on the implmentation. For example, in a system idemtification
implementation, the power of the reference is not important since the
normalization of the LMS is performed over the signal power. The
algorithm is self-adjusting for imput power.

However!!! The performance is very dependent on the model assumed. By
nonlinearity, I take you to mean that a linear model will have errors
if nonlinearities are present in the reference signal (the reference
being some nonlinear inner-product of the signal and system vector).
This just means the wrong model was chosen. If your model matches the
system, then the squared error will not be great. No matter what the
model, the algorithm will give the minimum squared error for the
model. This doesn't mean that the coefficients you end up with are
close to the desired coefficients.

The other important parameter is the condition of the input auto-
correlation matrix. An ill-conditioned matrix will result in not being
able to determine an optimum estimate of the system impulse response.

This is the problem with the example used by the OP. An input of 50 Hz
(single signal) will result in an ill-conditioned input auto-
correlartion matrix. There is NO optimum solution in this case. In
fact, there are an infinite number of solutions to this problem.

But, your point is well taken that the maximum mu used will be based
on the optimum performance required from the problem. That performance
may be minimum misadjustment noise, or maximum speed of convergence.
Each will ditate a different "maximum" mu.

What I think the OP was looking for, is how to determine the maximum
mu for convergence (often equated with stability). This value is
different for LMS versus NLMS.

Maurice
From: Vladimir Vassilevsky on


maury wrote:

> On Jul 14, 9:35 am, Vladimir Vassilevsky <nos...(a)nowhere.com> wrote:
>
>>maury wrote:
>>
>>>If you look at the normalized LMS (NLMS), then the maximum mu is
>>>limited to the sum of the number of filter coefficients.
>>
>>IMO what does matter is optimim rather then maximum. You need to
>>minimize the total error of the process. This error is because of
>>imperfect adaptation, noisy gradients, nonlinearity, ambient noise and
>>numeric artifacts. It also depends on the statistics and power of the
>>reference signal. So the algorithm must adapt Mu in pseudo Kalman way.
>>
>>Vladimir Vassilevsky
>>DSP and Mixed Signal Design Consultanthttp://www.abvolt.com
>
>
> Hi Vlad,
> With respect to the NLMS, the effect of the power of the reference
> depends on the implmentation. For example, in a system idemtification
> implementation, the power of the reference is not important since the
> normalization of the LMS is performed over the signal power. The
> algorithm is self-adjusting for imput power.

Yes, but the input is catching the ambient noise and double talk besides
the reflected signal. Hence if the input power is low, the algorithm
should not crank Mu all the way up to infinity, as simple NLMS would do,
but do the exactly the opposite.


> However!!! The performance is very dependent on the model assumed. By
> nonlinearity, I take you to mean that a linear model will have errors
> if nonlinearities are present in the reference signal (the reference
> being some nonlinear inner-product of the signal and system vector).
> This just means the wrong model was chosen. If your model matches the
> system, then the squared error will not be great.

System is given to you. Your algorithm has to work with whatever you got.


Vladimir Vassilevsky

DSP and Mixed Signal Design Consultant

http://www.abvolt.com
From: maury on
On Jul 14, 2:07 pm, Vladimir Vassilevsky <nos...(a)nowhere.com> wrote:

> System is given to you. Your algorithm has to work with whatever you got.

Kobayashi maru solution!
From: Jerry Avins on
On 7/14/2010 4:38 PM, maury wrote:
> On Jul 14, 2:07 pm, Vladimir Vassilevsky<nos...(a)nowhere.com> wrote:
>
>> System is given to you. Your algorithm has to work with whatever you got.
>
> Kobayashi maru solution!

Hah! another Trekke!
--
Engineering is the art of making what you want from things you can get.
From: Manny on
On Jul 14, 8:07 pm, Vladimir Vassilevsky <nos...(a)nowhere.com> wrote:
> maury wrote:
> > On Jul 14, 9:35 am, Vladimir Vassilevsky <nos...(a)nowhere.com> wrote:
>
> >>maury wrote:
>
> >>>If you look at the normalized LMS (NLMS), then the maximum mu is
> >>>limited to the sum of the number of filter coefficients.
>
> >>IMO what does matter is optimim rather then maximum. You need to
> >>minimize the total error of the process. This error is because of
> >>imperfect adaptation, noisy gradients, nonlinearity, ambient noise and
> >>numeric artifacts. It also depends on the statistics and power of the
> >>reference signal. So the algorithm must adapt Mu in pseudo Kalman way.
>
> >>Vladimir Vassilevsky
> >>DSP and Mixed Signal Design Consultanthttp://www.abvolt.com
>
> > Hi Vlad,
> > With respect to the NLMS, the effect of the power of the reference
> > depends on the implmentation. For example, in a system idemtification
> > implementation, the power of the reference is not important since the
> > normalization of the LMS is performed over the signal power. The
> > algorithm is self-adjusting for imput power.
>
> Yes, but the input is catching the ambient noise and double talk besides
>   the reflected signal. Hence if the input power is low, the algorithm
> should not crank Mu all the way up to infinity, as simple NLMS would do,
> but do the exactly the opposite.
>
> > However!!!  The performance is very dependent on the model assumed. By
> > nonlinearity, I take you to mean that a linear model will have errors
> > if nonlinearities are present in the reference signal (the reference
> > being some nonlinear inner-product of the signal and system vector).
> > This just means the wrong model was chosen. If your model matches the
> > system, then the squared error will not be great.
>
> System is given to you. Your algorithm has to work with whatever you got.

What if you discovered by sheer luck that doing something wanky like
skipping a beat works wonders to your signal? Sure this is uncool,
inelegant, and can't be analyzed, only characterized. Do you scrap all
this and wake up next day pretending that nothing has happened?

-Momo