You have 3 free guides left 😟
Unlock your guides
You have 3 free guides left 😟
Unlock your guides

5.1 The chain rule for functions of several variables

3 min readaugust 6, 2024

The chain rule for functions of several variables builds on what you already know about derivatives. It's like leveling up your calculus skills, allowing you to handle more complex situations involving multiple variables.

This rule is super useful when dealing with . It helps you break down tricky problems into manageable pieces, making it easier to find derivatives and understand how different variables affect each other.

Derivatives of Composite Functions

The Chain Rule and Composite Functions

Top images from around the web for The Chain Rule and Composite Functions
Top images from around the web for The Chain Rule and Composite Functions
  • The chain rule extends the concept of the single-variable chain rule to functions of several variables
    • Enables finding the derivative of a composite function by multiplying the of the outer function with the of the inner function
  • A composite function is a function that depends on another function
    • For example, if f(x,y)=x2+y2f(x, y) = x^2 + y^2 and g(t)=(t,t3)g(t) = (t, t^3), then f(g(t))=t2+t6f(g(t)) = t^2 + t^6 is a composite function
  • To apply the chain rule, express the composite function as a composition of an outer function and an inner function
    • The outer function takes the output of the inner function as its input
    • The inner function is a function of the original independent variables

Partial Derivatives and the Total Derivative

  • Partial derivatives measure the rate of change of a function with respect to one variable while holding other variables constant
    • For a function f(x,y)f(x, y), the partial derivative with respect to xx is denoted as fx\frac{\partial f}{\partial x} and the partial derivative with respect to yy is denoted as fy\frac{\partial f}{\partial y}
  • The is the generalization of the single-variable derivative to functions of several variables
    • It measures the rate of change of a function in any direction
    • For a function f(x,y)f(x, y), the total derivative is given by df=fxdx+fydydf = \frac{\partial f}{\partial x}dx + \frac{\partial f}{\partial y}dy
  • When applying the chain rule, the total derivative of the composite function is the product of the partial derivatives of the outer function and the total derivatives of the inner functions
    • For example, if f(x,y)=x2+y2f(x, y) = x^2 + y^2 and x=g(t)=tx = g(t) = t and y=h(t)=t3y = h(t) = t^3, then dfdt=fxdxdt+fydydt=2xdxdt+2ydydt=2t(1)+2t3(3t2)=2t+6t5\frac{df}{dt} = \frac{\partial f}{\partial x}\frac{dx}{dt} + \frac{\partial f}{\partial y}\frac{dy}{dt} = 2x\frac{dx}{dt} + 2y\frac{dy}{dt} = 2t(1) + 2t^3(3t^2) = 2t + 6t^5

Multivariable Calculus Fundamentals

Multivariable Functions and Differentiability

  • Multivariable functions are functions that depend on multiple variables
    • For example, f(x,y)=x2+y2f(x, y) = x^2 + y^2 is a function of two variables, xx and yy
  • A multivariable function is differentiable at a point if its partial derivatives exist and are continuous at that point
    • Differentiability ensures that the function is smooth and well-behaved near the point
  • Differentiability allows for the approximation of a function using a linear function (its total derivative) near a point
    • This linear approximation is the basis for many applications, such as optimization and error estimation

Gradient Vector and Jacobian Matrix

  • The of a multivariable function f(x1,x2,,xn)f(x_1, x_2, \ldots, x_n) is a vector of its partial derivatives
    • It points in the direction of the greatest rate of increase of the function
    • The gradient vector is denoted as f=(fx1,fx2,,fxn)\nabla f = \left(\frac{\partial f}{\partial x_1}, \frac{\partial f}{\partial x_2}, \ldots, \frac{\partial f}{\partial x_n}\right)
  • The is a matrix of partial derivatives for a vector-valued function f(x1,x2,,xn)=(f1(x1,x2,,xn),f2(x1,x2,,xn),,fm(x1,x2,,xn))\mathbf{f}(x_1, x_2, \ldots, x_n) = (f_1(x_1, x_2, \ldots, x_n), f_2(x_1, x_2, \ldots, x_n), \ldots, f_m(x_1, x_2, \ldots, x_n))
    • It generalizes the gradient vector to vector-valued functions
    • The Jacobian matrix is denoted as Jf=(f1x1f1x2f1xnf2x1f2x2f2xnfmx1fmx2fmxn)J_{\mathbf{f}} = \begin{pmatrix} \frac{\partial f_1}{\partial x_1} & \frac{\partial f_1}{\partial x_2} & \cdots & \frac{\partial f_1}{\partial x_n} \\ \frac{\partial f_2}{\partial x_1} & \frac{\partial f_2}{\partial x_2} & \cdots & \frac{\partial f_2}{\partial x_n} \\ \vdots & \vdots & \ddots & \vdots \\ \frac{\partial f_m}{\partial x_1} & \frac{\partial f_m}{\partial x_2} & \cdots & \frac{\partial f_m}{\partial x_n} \end{pmatrix}
  • The gradient vector and Jacobian matrix play crucial roles in optimization, coordinate transformations, and analyzing the behavior of multivariable functions
© 2024 Fiveable Inc. All rights reserved.
AP® and SAT® are trademarks registered by the College Board, which is not affiliated with, and does not endorse this website.


© 2024 Fiveable Inc. All rights reserved.
AP® and SAT® are trademarks registered by the College Board, which is not affiliated with, and does not endorse this website.

© 2024 Fiveable Inc. All rights reserved.
AP® and SAT® are trademarks registered by the College Board, which is not affiliated with, and does not endorse this website.
Glossary
Glossary