Shared Flashcard Set

Details

Nonlinear Equations and Optimization
Convergences; Single equations; Systems of equations; One-dimensional optimization; Multidimensional unconstrained optimization
33
Computer Science
Graduate
03/03/2014

Additional Computer Science Flashcards

 


 

Cards

Term
Intermediate Value Theorem
Definition
If f is continuous on a closed interval [a,b] and c lies in between f(a) and f(b) then there is some value x in [a,b] such that f(x) = c.
Term
Inverse Function Theorem
Definition
For a continuously differentiable function f, if the Jacobian matrix Jf is nonsingular at a point x then there is a neighborhood of f(x) in which f is invertible.  That is, the equation f(x) = y has a solution for any y in the neighborhood of f(x).  Nonsingular J = locally invertible everywhere ≠ globally invertible
Term
Contraction Mapping Theorem
Definition

If g is contractive on a closed set S and g(S) is a subset of S then g has a unique fixed point in S. 

Contractive: function g on set S if there is a constant γ with 0< γ<1 such that

||g(x)-g(z)|| < γ||x-z||

Term
Coerciveness
Definition

A continuous function f on an unbounded set S if lim(as ||x|| goes to infinity) of f(x) = +infinity.

For any constant M there is an r > 0 (dependent on M) such that f(x) >=M for any x in S such that ||x|| >= r.

Term
Convexity
Definition

A set S is convex if it contains the line segment between any two of its points

i.e. for all x,y ε S, 0 < α < 1, αx + (1 - α)y ε S

Term
Gradient Vector
Definition
The derivative of f with respect to every variable, in form of a vector.  The gradient points uphill while the negative gradient points downhill towards points having lower function values than f(x).
Term
Hessian Matrix
Definition
The Jacobian of the gradient of f.  hij is the second partial derivative of f(x) with repect to both xi and xj.
Term
Optimality conditions
Definition
For constrained optimization, the solution often occurs on the boundary of the set.  So, consider only feasible directions.  A nonzero vector s is a feasible direction at a point x in S if there is an r > 0 such that x + αs is in S for all α in [0,r].
Term
Multiple root
Definition

f(x) = 0 and f'(x) = 0

The curve defined by f has a horizontal tangent on the x axis.

Root of multiplicity 2.

Term
Convergence Rates
Definition
An iterative method is said to converge with rate r if lim(k→∞)||ek+1||/||ek||= C where C is a constant.
Term
Interval Bisection Method
Definition

Begins with an initial bracket and successively reduces its length until the solution has been isolated as accurately as desired.

At each iteration, f is evaluated at the midpoint of the current intervalue and half of the interval is discarded.

Term
Fixed-point Iteration
Definition

The point x=g(x) is a fixed point of the function g since x is unchanged when g is applied.  For fixed-point iteration, g is a function chosen so that its fixed points are solutions for f(x) = 0.

xk+1=g(xk)

Term
Newton's Method in One Dimension
Definition

xk+1=xk-f(x)/f '(x)

Converges if started close enough to root

Quadratic if not multiple root

 

Term
Convergence Rate of Newton's Method in One Direction
Definition

g(x) = x - f(x)/f '(x)

g '(x) = 1 - (f(x)f''(x) - (f '(x))2)/(f '(x))2

=  f(x)f '(x)/(f '(x))2 = 0 if simple root

 

Term
Secant Method
Definition

xk+1=xk-(xk-xk-1)f(xk)/(f(xk)-f(xk-1))

because f '(xk) ≈ (f(xk)-f(xk-1))/(xk-xk-1)

Term
Inverse Interpolation
Definition
Fit the values xk as a function of the values yk=f(xk) by a polynomial p(y) so that the next approximate solution is simply p(0).  At each iteration we have three approximate solution values.  The next approximate solution is found by fitting a quadratic polynomial to these values and then evaluating it at 0.
Term
Safeguarded Methods
Definition
Rapidly convergent methods are unsafe -- may not converge unless they are started close enough to the solution.  Hybrid methods safe and fast.  Could use a rapidly converent method but maintain a bracket around the solution and if outside, use safe method to return to interval.
Term
Fixed-point Iteration for Systems of Equations
Definition
For some starting vector x0, xk+1=g(xk), convergent if ρ(G(x))<1 where G(x) is the Jacobian of g evaluated at x.  If G(x)=0 then convergence rate is at least quadratic.
Term
Jacobian Matrix
Definition
G(x)ij=∂gi(x)/∂xj
Term
Newton's Method for Systems of Equations
Definition

At each iteration:

Solve Jf(xk)sk=-f(xk)

xk+1=xk+sk

Term
Broyden's Method
Definition

B0=initial Jacobian approx (I)

for k = 0, 1, 2, ...

    Solve Bksk=-f(xk) for sk

    xk+1=xk+sk

    yk=f(xk+1)-f(xk)

    Bk+1=Bk+((yk-Bksk)skT)/(skTsk)

Term
Joe DeGol
Definition
Sexy Genius
Term
Damped Newton Method
Definition
Newton step is computed as usual, but new iterate is taken to be xk+1=xkksk.  Far from solution, step is often too large.  Monitor ||f(x)|| and ensure that it decreases sufficiently each iteration.  When close, αk=1 to maintain usual convergence.
Term
Newton's Method with Trust Region
Definition
Maintain an estimate of the radius of a trust region within which the Taylor Series approximation is sufficiently accurate. Adjust size of trust region as necessary to constrain step size.  Large enough to permit full Newton steps when close.
Term
Golden Section Search
Definition
If f is unimodal on [a,b], x1,x2 in [a,b] with x1<x2.  Evaluate f(x1), f(x2) and discard either [a,x1) or (x2,b].  Want τ2 = (1-τ).
Term
Successive Parabolic Interpolation
Definition
Initially, the function f to be minimized is evaluated at three points and a quadratic polynomial is fit to the three resulting function values.  The minimum of the parabola is the new approximate minimum, and oldest of last three points is dropped.  Repeat.
Term
Newton's Method for One Dimensional Optimization
Definition

xk+1=x- f '(xk)/f ''(xk)

f(x+h)≈f(x)+f '(x)h+.5f ''(x)h2

f '(x+h) ≈ f'(x) + f''(x)h = 0

h = -f '(x) / f ''(x)

Term
Safeguarded One Dimensional Optimization
Definition
Use interval bracketing to maintain a safe method whenever the fast method generates an iterate that lies outside the interval
Term
Steepest Descent Method
Definition

x0=initial guess

for k = 0, 1, 2, ...

    sk = -grad(f(xk))

    Choose αk to minimize f(xk+ αksk)

    xk+1=xk+ αksk

Term
Newton's Method for Unconstrained Optimization
Definition

x0=initial guess

for k = 0, 1, 2, ...

    Solve Hf(xk)sk=-grad(f(xk))

    xk+1=xk+sk

Term
Quasi-Newton Methods
Definition

Newton's method requires substantial amount of work per iteration (especially with dense Hessian).  Reduce overhead or improve reliability.

xk+1=xk - αkBk-1grad(f(xk)) where αk is line search parameter and Bk is approximation to Hessian

Term
Conjugate Gradient Method for Unconstrained Optimization
Definition

for k = 0, 1, 2, ...

    Choose αk to minimize f(xk+ αksk)

    xk+1=xk+ αksk

    gk+1=grad(f(xk+1))

    sk+1= - gk+1 + ||gk+1||/||gk|| sk

Supporting users have an ad free experience!