Title: LMS Algorithm in a Reproducing Kernel Hilbert Space
1LMS Algorithm in a Reproducing Kernel Hilbert
Space
- Weifeng Liu, P. P. Pokharel, J. C. Principe
- Computational NeuroEngineering Laboratory,
- University of Florida
- Acknowledgment This work was partially supported
by NSF grant ECS-0300340 and ECS-0601271.
2Outlines
- Introduction
- Least Mean Square algorithm (easy)
- Reproducing kernel Hilbert space (tricky)
- The convergence and regularization analysis
(important) - Learning from error models (interesting)
3Introduction
- Puskal (2006) Kernel LMS
- Kivinen, Smola (2004) Online learning with
kernels (more like leaky LMS) - Moody, Platt (1990s)Resource allocation
networks (growing and pruning)
4LMS (1960, Widrow and Hoff)
- Given a sequence of examples from UR
- U a compact set of RL.
- The model is assumed
- The cost function
5LMS
- The LMS algorithm
- The weight after n iteration
(1)
(2)
6Reproducing kernel Hilbert space
- A continuous, symmetric, positive-definite kernel
,a mapping F, and an inner
product - H is the closure of the span of all F(u).
- Reproducing
- Kernel trick
- The induced norm
7RKHS
- Kernel trick
- An inner product in the feature space
- A similarity measure you needed.
- Mercers theorem
8Common kernels
- Gaussian kernel
- Polynomial kernel
9Kernel LMS
- Transform the input ui to F(ui)
- Assume F(ui) ?RM
- The model is assumed
- The cost function
10Kernel LMS
- The KLMS algorithm
- The weight after n iteration
(3)
(4)
11Kernel LMS
(5)
12Kernel LMS
- After the learning, the input-output relation
(6)
13KLMS vs. RBF
- KLMS
- RBF
- a satisfy
- G is the gram matrix G(i,j)?(ui,uj)
- RBF needs regularization.
- Does KLMS need regularization?
(7)
(8)
14KLMS vs. LMS
- Kernel LMS is nothing but LMS in the feature
space--a very high dimensional reproducing kernel
Hilbert space (MgtN) - Eigen-spread is awfuldoes it converge?
15Example MG signal predication
- Time embedding 10.
- Learn rate 0.2
- 500 training data
- 100 test data point.
- Gaussian noise
- noise variance .04
16Example MG signal predication
MSE Linear LMS KLMS RBF (?0) RBF (?.1) RBF (?1) RBF (?10)
training 0.021 0.0060 0 0.0026 0.0036 0.010
test 0.026 0.0066 0.019 0.0041 0.0050 0.014
17Complexity Comparison
RBF KLMS LMS
Computation O(N3) O(N2) O(L)
Memory O(N2NL) O(NL) O(L)
18The asymptotic analysis on convergencesmall
step-size theory
- Denote
- The correlation matrix
- is singular. Assume
- and
19The asymptotic analysis on convergencesmall
step-size theory
20The weight stays at the initial place in the
0-eigen-value directions
21The 0-eigen-value directions does not affect the
MSE
It does not care about the null space! It only
focuses on the data space!
22The minimum norm initialization
- The initialization gives the
minimum norm possible solution. -
23Minimum norm solution
24Learning is Ill-posed
25Over-learning
26Regularization Technique
- Learning from finite data is ill-posed.
- A priori information--Smoothness is needed.
- The norm of the function, which indicates the
slope of the linear operator is constrained. - In statistical learning theory, the norm is
associated with the confidence of uniform
convergence!
27Regularized RBF
- The cost function
- or equivalently
28KLMS as a learning algorithm
- The model with
- The following inequalities hold
- The proof(H8 robust triangle inequality
matrix transformation derivative )
29The numerical analysis
- The solution of regularized RBF is
- The reason of ill-posedness is the inversion of
the matrix (G?I)
30The numerical analysis
- The solution of KLMS is
- By the inequality we have
31Example MG signal predication
weight KLMS RBF (?0) RBF (?.1) RBF (?1) RBF (?10)
norm 0.520 4.8e3 10.90 1.37 0.231
32The conclusion
- The LMS algorithm can be readily used in a RKHS
to derive nonlinear algorithms. - From the machine learning view, the LMS method is
a simple tool to have a regularized solution.
33Demo
34Demo
35LMS learning model
- An event happens, and a decision made.
- If the decision is correct, nothing happens.
- If an error is incurred, a correction is made on
the original model. - If we do things right, everything is fine and
life goes on. - If we do something wrong, lessons are drawn and
our abilities are honed.
36Would we over-learn?
- If the real world is attempted to be modeled
mathematically, what dimension is appropriate? - Are we likely to over-learn?
- Are we using the LMS algorithm?
- What is good to remember the past?
- What is bad to be a perfectionist?
37- "If you shut your door to all errors, truth will
be shut out."---Rabindranath Tagore