You have 3 free guides left 😟
Unlock your guides
You have 3 free guides left 😟
Unlock your guides

Line search methods are crucial for unconstrained optimization. They find the best direction and to move towards the minimum. These methods balance between computational efficiency and convergence speed, making them practical for many real-world problems.

Steepest descent and are two key line search approaches. , like BFGS, offer a middle ground. They use clever approximations to avoid costly calculations while still converging quickly in most cases.

Search Direction and Step Size

Fundamentals of Line Search Methods

  • Search direction determines the path of optimization in multidimensional space
  • Step size controls the magnitude of movement along the search direction
  • Line search methods iterate between choosing a direction and determining an appropriate step size
  • Optimization process continues until convergence criteria are met (gradient norm below threshold)

Steepest Descent and Newton's Method

  • Steepest descent method uses negative gradient as search direction
    • Computationally inexpensive but can be slow to converge
    • Formula: dk=f(xk)d_k = -\nabla f(x_k)
    • Performs well for functions with circular contours
  • Newton's method utilizes both gradient and Hessian information
    • Faster convergence rate, especially near the optimum
    • Search direction given by: dk=Hk1f(xk)d_k = -H_k^{-1} \nabla f(x_k)
    • Requires computation and inversion of
  • Both methods update the current point using: xk+1=xk+αkdkx_{k+1} = x_k + \alpha_k d_k
    • αk\alpha_k represents the step size
    • Determined through line search procedures (exact or inexact)

Quasi-Newton Methods

Principles and Advantages

  • Quasi-Newton methods approximate the Hessian matrix or its inverse
  • Balance between computational efficiency and convergence speed
  • Avoid explicit calculation and inversion of the Hessian matrix
  • Update formula maintains positive definiteness of approximation
  • rate in many practical applications

BFGS Method Implementation

  • Broyden-Fletcher-Goldfarb-Shanno (BFGS) algorithm widely used in practice
  • Approximates the inverse Hessian matrix directly
  • BFGS update formula: Bk+1=Bk+ykykTykTskBkskskTBkskTBkskB_{k+1} = B_k + \frac{y_k y_k^T}{y_k^T s_k} - \frac{B_k s_k s_k^T B_k}{s_k^T B_k s_k}
    • BkB_k represents the current approximation of the inverse Hessian
    • sk=xk+1xks_k = x_{k+1} - x_k denotes the step taken
    • yk=f(xk+1)f(xk)y_k = \nabla f(x_{k+1}) - \nabla f(x_k) measures the gradient change
  • Limited-memory BFGS (L-BFGS) variant for large-scale problems
    • Stores only a fixed number of vector pairs
    • Reduces memory requirements for high-dimensional optimization tasks

Line Search Conditions

Armijo Rule and Sufficient Decrease

  • ensures sufficient decrease in objective function value
  • Condition: f(xk+αkdk)f(xk)+c1αkf(xk)Tdkf(x_k + \alpha_k d_k) \leq f(x_k) + c_1 \alpha_k \nabla f(x_k)^T d_k
    • c1c_1 typically set to a small value (0.0001 to 0.1)
  • Prevents excessively large steps that may increase function value
  • Often combined with backtracking to find suitable step size

Wolfe Conditions for Step Size Selection

  • comprise two inequalities for step size selection
  • (Armijo rule)
  • Curvature condition: f(xk+αkdk)Tdkc2f(xk)Tdk\nabla f(x_k + \alpha_k d_k)^T d_k \geq c_2 \nabla f(x_k)^T d_k
    • c2c_2 typically chosen between 0.1 and 0.9
  • Strong Wolfe conditions use absolute value in curvature condition
  • Ensure reasonable progress and avoid excessively small steps

Backtracking Line Search Algorithm

  • Iterative procedure to find step size satisfying Armijo rule
  • Start with initial step size (often α=1\alpha = 1)
  • Reduce step size by a factor ρ\rho (typically 0.5) if Armijo condition not met
  • Continue reduction until suitable step size found
  • Pseudocode for backtracking:
    while f(x + α*d) > f(x) + c*α*∇f(x)^T*d:
        α = ρ * α
    
  • Efficient method for inexact line search in practice
© 2024 Fiveable Inc. All rights reserved.
AP® and SAT® are trademarks registered by the College Board, which is not affiliated with, and does not endorse this website.


© 2024 Fiveable Inc. All rights reserved.
AP® and SAT® are trademarks registered by the College Board, which is not affiliated with, and does not endorse this website.

© 2024 Fiveable Inc. All rights reserved.
AP® and SAT® are trademarks registered by the College Board, which is not affiliated with, and does not endorse this website.
Glossary
Glossary