Tải bản đầy đủ (.pdf) (22 trang)

Regularized least squares and Gauss Newton Method

Bạn đang xem bản rút gọn của tài liệu. Xem và tải ngay bản đầy đủ của tài liệu tại đây (164.57 KB, 22 trang )

EE263 Autumn 2007-08

Stephen Boyd

Lecture 7
Regularized least-squares and Gauss-Newton
method
• multi-objective least-squares
• regularized least-squares
• nonlinear least-squares
• Gauss-Newton method

7–1


Multi-objective least-squares
in many problems we have two (or more) objectives
• we want J1 = Ax − y

2

small

• and also J2 = F x − g

2

small

(x ∈ Rn is the variable)
• usually the objectives are competing


• we can make one smaller, at the expense of making the other larger
common example: F = I, g = 0; we want Ax − y small, with small x
Regularized least-squares and Gauss-Newton method

7–2


Plot of achievable objective pairs
plot (J2, J1) for every x:

J1

x(1)
x(2)

x(3)

J2
note that x ∈ Rn, but this plot is in R2; point labeled x(1) is really
J2(x(1)), J1(x(1))
Regularized least-squares and Gauss-Newton method

7–3


• shaded area shows (J2, J1) achieved by some x ∈ Rn
• clear area shows (J2, J1) not achieved by any x ∈ Rn
• boundary of region is called optimal trade-off curve
• corresponding x are called Pareto optimal
(for the two objectives Ax − y 2, F x − g 2)


three example choices of x: x(1), x(2), x(3)
• x(3) is worse than x(2) on both counts (J2 and J1)
• x(1) is better than x(2) in J2, but worse in J1

Regularized least-squares and Gauss-Newton method

7–4


Weighted-sum objective

• to find Pareto optimal points, i.e., x’s on optimal trade-off curve, we
minimize weighted-sum objective
J1 + µJ2 = Ax − y

2

+ µ Fx − g

2

• parameter µ ≥ 0 gives relative weight between J1 and J2
• points where weighted sum is constant, J1 + µJ2 = α, correspond to
line with slope −µ on (J2, J1) plot

Regularized least-squares and Gauss-Newton method

7–5



PSfrag

J1

x(1)
x(3)
x(2)
J1 + µJ2 = α
J2

• x(2) minimizes weighted-sum objective for µ shown
• by varying µ from 0 to +∞, can sweep out entire optimal tradeoff curve

Regularized least-squares and Gauss-Newton method

7–6


Minimizing weighted-sum objective
can express weighted-sum objective as ordinary least-squares objective:
Ax − y

2

+ µ Fx − g

where
A˜ =


2

A

µF

A

µF

=
=

˜ − y˜
Ax

,

y˜ =

x−

y

µg

2

2


y

µg

hence solution is (assuming A˜ full rank)
x =
=

A˜T A˜

−1

A˜T y˜

AT A + µF T F

Regularized least-squares and Gauss-Newton method

−1

AT y + µF T g

7–7


Example
f

• unit mass at rest subject to forces xi for i − 1 < t ≤ i, i = 1, . . . , 10
• y ∈ R is position at t = 10; y = aT x where a ∈ R10

• J1 = (y − 1)2 (final position error squared)
• J2 = x

2

(sum of squares of forces)

weighted-sum objective: (aT x − 1)2 + µ x

2

optimal x:
T

x = aa + µI
Regularized least-squares and Gauss-Newton method

−1

a
7–8


optimal trade-off curve:
1

0.9

0.8


J1 = (y − 1)2

0.7

0.6

0.5

0.4

0.3

0.2

0.1

0

0

0.5

1

1.5

2

J2 = x


2.5

2

3

3.5
−3

x 10

• upper left corner of optimal trade-off curve corresponds to x = 0
• bottom right corresponds to input that yields y = 1, i.e., J1 = 0
Regularized least-squares and Gauss-Newton method

7–9


Regularized least-squares
when F = I, g = 0 the objectives are
J1 = Ax − y 2,

J2 = x

2

minimizer of weighted-sum objective,
x = AT A + µI

−1


AT y,

is called regularized least-squares (approximate) solution of Ax ≈ y
• also called Tychonov regularization
• for µ > 0, works for any A (no restrictions on shape, rank . . . )
Regularized least-squares and Gauss-Newton method

7–10


estimation/inversion application:
• Ax − y is sensor residual
• prior information: x small
• or, model only accurate for x small
• regularized solution trades off sensor fit, size of x

Regularized least-squares and Gauss-Newton method

7–11


Nonlinear least-squares

nonlinear least-squares (NLLS) problem: find x ∈ Rn that minimizes
m

r(x)

2


ri(x)2,

=
i=1

where r : Rn → Rm
• r(x) is a vector of ‘residuals’
• reduces to (linear) least-squares if r(x) = Ax − y

Regularized least-squares and Gauss-Newton method

7–12


Position estimation from ranges
estimate position x ∈ R2 from approximate distances to beacons at
locations b1, . . . , bm ∈ R2 without linearizing

• we measure ρi = x − bi + vi
(vi is range error, unknown but assumed small)
• NLLS estimate: choose x
ˆ to minimize
m

m

2

ri(x)2 =

i=1

Regularized least-squares and Gauss-Newton method

i=1

(ρi − x − bi )

7–13


Gauss-Newton method for NLLS
m

NLLS: find x ∈ Rn that minimizes r(x)
n

r:R →R

m

2

ri(x)2, where

=
i=1

• in general, very hard to solve exactly
• many good heuristics to compute locally optimal solution

Gauss-Newton method:
given starting guess for x
repeat
linearize r near current guess
new guess is linear LS solution, using linearized r
until convergence
Regularized least-squares and Gauss-Newton method

7–14


Gauss-Newton method (more detail):
• linearize r near current iterate x(k):
r(x) ≈ r(x(k)) + Dr(x(k))(x − x(k))
where Dr is the Jacobian: (Dr)ij = ∂ri/∂xj
• write linearized approximation as
r(x(k)) + Dr(x(k))(x − x(k)) = A(k)x − b(k)
A(k) = Dr(x(k)),

b(k) = Dr(x(k))x(k) − r(x(k))

• at kth iteration, we approximate NLLS problem by linear LS problem:
r(x)
Regularized least-squares and Gauss-Newton method

2

≈ A

(k)


x−b

(k)

2

7–15


• next iterate solves this linearized LS problem:
(k+1)

x

= A

(k)T

A

(k)

−1

A(k)T b(k)

• repeat until convergence (which isn’t guaranteed)

Regularized least-squares and Gauss-Newton method


7–16


Gauss-Newton example
• 10 beacons
• + true position (−3.6, 3.2); ♦ initial guess (1.2, −1.2)
• range estimates accurate to ±0.5
5

4

3

2

1

0

−1

−2

−3

−4

−5
−5


−4

−3

−2

Regularized least-squares and Gauss-Newton method

−1

0

1

2

3

4

5

7–17


NLLS objective r(x)

2


versus x:

16

14

12

10

8

6

4

2
5
0
5

0
0
−5

−5

• for a linear LS problem, objective would be nice quadratic ‘bowl’
• bumps in objective due to strong nonlinearity of r
Regularized least-squares and Gauss-Newton method


7–18


objective of Gauss-Newton iterates:
12

10

r(x)

2

8

6

4

2

0

1

2

3

4


5

6

7

8

9

10

iteration
• x(k) converges to (in this case, global) minimum of r(x)

2

• convergence takes only five or so steps
Regularized least-squares and Gauss-Newton method

7–19


• final estimate is x
ˆ = (−3.3, 3.3)
• estimation error is x
ˆ − x = 0.31
(substantially smaller than range accuracy!)


Regularized least-squares and Gauss-Newton method

7–20


convergence of Gauss-Newton iterates:
5
4
3

4
56
3

2
2

1

0

1

−1

−2

−3

−4


−5
−5

−4

−3

−2

Regularized least-squares and Gauss-Newton method

−1

0

1

2

3

4

5

7–21


useful varation on Gauss-Newton: add regularization term

A(k)x − b(k)

2

+ µ x − x(k)

2

so that next iterate is not too far from previous one (hence, linearized
model still pretty accurate)

Regularized least-squares and Gauss-Newton method

7–22



×