Calculus of variations is a powerful mathematical tool for optimizing functionals, which are mappings from function spaces to real numbers. It's crucial in control theory for solving optimization problems and finding optimal trajectories.
This field introduces key concepts like functionals, variations, and the . These tools help engineers and scientists formulate and solve complex problems in mechanics, physics, and control systems, leading to more efficient and effective designs.
Fundamental concepts of calculus of variations
Calculus of variations is a field of mathematical analysis that deals with finding extrema (maxima or minima) of functionals, which are mappings from a set of functions to the real numbers
It has wide-ranging applications in various branches of science and engineering, including control theory, where it is used to formulate and solve optimization problems
Key concepts in calculus of variations include functionals, function spaces, variations, and necessary and for extrema
Functionals and function spaces
Top images from around the web for Functionals and function spaces
calculus of variations - Using Rayleigh-Ritz Method to approximate solutions to extremum problem ... View original
Is this image relevant?
calculus of variations - Consider the following functional which is as follows: - Mathematics ... View original
calculus of variations - Using Rayleigh-Ritz Method to approximate solutions to extremum problem ... View original
Is this image relevant?
calculus of variations - Consider the following functional which is as follows: - Mathematics ... View original
Is this image relevant?
1 of 3
A is a mapping that assigns a real number to each function in a certain function space
Function spaces are sets of functions that share certain properties, such as , , or integrability
Examples of function spaces include the space of continuous functions (C[a,b]), the space of square-integrable functions (L^2[a,b]), and the space of differentiable functions (C^1[a,b])
The choice of function space depends on the specific problem and the desired properties of the solution
Weak and strong variations
Variations are changes or perturbations made to a function to analyze its behavior near an extremum
are infinitesimal changes that satisfy the boundary conditions of the problem and are used to derive for extrema
are finite changes that may not satisfy the boundary conditions and are used to derive sufficient conditions for extrema
The concept of variations is fundamental to the derivation of the Euler-Lagrange equation and other optimality conditions
Necessary and sufficient conditions for extrema
Necessary conditions are criteria that must be satisfied by any extremum of a functional, but they do not guarantee that a given function is an extremum
Sufficient conditions are criteria that, when satisfied, ensure that a given function is an extremum of the functional
The most well-known necessary condition in calculus of variations is the Euler-Lagrange equation, which is derived using the concept of weak variations
Sufficient conditions, such as the Legendre and Jacobi conditions, are used to distinguish between minima, maxima, and saddle points
Euler-Lagrange equation
The Euler-Lagrange equation is a fundamental result in calculus of variations that provides a necessary condition for a function to be an extremum of a functional
It is widely used in various fields, including mechanics, physics, and control theory, to formulate and solve optimization problems
The equation relates the functional to be minimized or maximized to the function and its derivatives
Derivation of Euler-Lagrange equation
The Euler-Lagrange equation is derived by considering a functional J[y]=∫abF(x,y,y′)dx, where y(x) is a function and F is a given function of x, y, and y′ (the derivative of y with respect to x)
By applying the concept of weak variations and requiring that the first variation of the functional vanishes at an extremum, the Euler-Lagrange equation is obtained: ∂y∂F−dxd∂y′∂F=0
This equation must be satisfied by any function y(x) that extremizes the functional J[y]
First and second order conditions
The Euler-Lagrange equation is a first-order necessary condition for an extremum, as it involves only the first derivative of the function
Second-order conditions, such as the Legendre and Jacobi conditions, are used to distinguish between minima, maxima, and saddle points
The states that for a minimum (maximum), the second partial derivative of F with respect to y′ must be non-negative (non-positive)
The involves the analysis of conjugate points and ensures that the extremum is a true minimum or maximum
Generalizations and extensions
The Euler-Lagrange equation can be generalized to handle functionals involving higher-order derivatives, multiple functions, and multiple independent variables
For functionals with higher-order derivatives, the equation becomes ∑k=0n(−1)kdxkdk∂y(k)∂F=0, where y(k) denotes the k-th derivative of y
For functionals with multiple functions y1,…,ym, there will be a separate Euler-Lagrange equation for each function
The Euler-Lagrange equation can also be extended to handle functionals with multiple independent variables, such as those arising in the calculus of variations in several dimensions
Variational problems with constraints
Many optimization problems in control theory and other fields involve constraints on the functions or variables being optimized
Constrained variational problems require specialized techniques to derive necessary and sufficient conditions for extrema
Common types of constraints include holonomic and , isoperimetric constraints, and integral constraints
Holonomic and non-holonomic constraints
are equations that involve only the functions and the independent variables, such as g(x,y1,…,ym)=0
Non-holonomic constraints are equations that involve the derivatives of the functions, such as g(x,y1,…,ym,y1′,…,ym′)=0
Holonomic constraints can be handled using the method of , while non-holonomic constraints require more advanced techniques, such as the method of undetermined multipliers or the Lagrange-d'Alembert principle
Lagrange multipliers and constrained optimization
The method of Lagrange multipliers is a powerful technique for solving optimization problems with equality constraints
The idea is to introduce a new variable (the Lagrange multiplier) for each constraint and to form the L(x,y,λ)=F(x,y)+λg(x,y), where F is the functional to be optimized and g is the constraint equation
Necessary conditions for an extremum are obtained by setting the partial derivatives of the Lagrangian with respect to x, y, and λ equal to zero
The resulting system of equations can be solved to find the extrema of the constrained problem
Isoperimetric problems and applications
are a special class of constrained variational problems in which the constraint involves an integral, such as ∫abg(x,y,y′)dx=c, where c is a constant
A classic example is the problem of finding the curve of a given length that encloses the maximum area (the solution is a circle)
Isoperimetric problems arise in various applications, such as the design of optimal control strategies for systems with limited resources (e.g., fuel or time)
The necessary conditions for an extremum in an isoperimetric problem are obtained by introducing a Lagrange multiplier and forming the modified functional J[y]=∫ab[F(x,y,y′)+λg(x,y,y′)]dx
Hamilton's principle and least action
, also known as the principle of least action, is a fundamental in mechanics and physics
It states that the motion of a system between two fixed points in configuration space is such that the is stationary (usually a minimum)
The principle provides a unifying framework for deriving the equations of motion in classical mechanics and has important implications for the study of conservation laws and symmetries
Formulation of Hamilton's principle
The action integral is defined as S[q]=∫t1t2L(q,q˙,t)dt, where L is the Lagrangian of the system, q represents the generalized coordinates, and q˙ represents the generalized velocities
Hamilton's principle states that the actual path followed by the system is one for which the action integral is stationary with respect to variations of the path that keep the endpoints fixed
Mathematically, this is expressed as δS=0, where δ denotes the variation operator
Applying the techniques of calculus of variations to this condition leads to the Euler-Lagrange equations for the system, which are equivalent to Newton's equations of motion
Principle of least action in mechanics
In mechanics, the Lagrangian is defined as the difference between the kinetic energy T and the potential energy V of the system: L=T−V
The principle of least action states that the motion of a mechanical system between two fixed points in configuration space is such that the action integral ∫t1t2(T−V)dt is minimized
This formulation provides a variational approach to mechanics that is equivalent to Newton's laws but often more convenient for analyzing complex systems
The principle of least action has deep connections to the concepts of energy conservation and the symmetries of the system
Conservation laws and Noether's theorem
is a profound result that links the symmetries of a system to the existence of conservation laws
It states that for every continuous symmetry of the action integral, there exists a corresponding conserved quantity
For example, time translation symmetry leads to the conservation of energy, spatial translation symmetry leads to the conservation of linear momentum, and rotational symmetry leads to the conservation of angular momentum
Noether's theorem provides a powerful tool for identifying conserved quantities in mechanical systems and has far-reaching implications in various branches of physics
Direct methods in calculus of variations
are a class of techniques used to find approximate solutions to variational problems by converting them into finite-dimensional optimization problems
Unlike , which rely on the Euler-Lagrange equation and other necessary conditions, direct methods seek to minimize the functional directly by considering a finite-dimensional subspace of the original function space
Common direct methods include the Ritz method, the Galerkin method, and the
Ritz and Galerkin methods
The Ritz method approximates the solution of a variational problem by a linear combination of basis functions, such as polynomials or trigonometric functions
The coefficients of the linear combination are determined by minimizing the functional with respect to these coefficients, leading to a system of algebraic equations
The Galerkin method is similar to the Ritz method but differs in the way the coefficients are determined: instead of minimizing the functional, the Galerkin method requires that the residual of the Euler-Lagrange equation is orthogonal to the basis functions
Both methods provide a systematic way to obtain approximate solutions to variational problems and have been widely used in various fields of science and engineering
Finite element approximations
The finite element method (FEM) is a powerful numerical technique for solving variational problems by discretizing the domain into a set of simple subdomains, called finite elements
The solution is approximated by a piecewise polynomial function over the finite elements, and the coefficients are determined by minimizing the functional or by requiring that the residual is orthogonal to a set of test functions
FEM is particularly well-suited for problems with complex geometries and has been extensively used in structural mechanics, fluid dynamics, and electromagnetic field analysis
The method provides a systematic way to obtain accurate approximate solutions and has a solid mathematical foundation based on the theory of Sobolev spaces and weak formulations
Convergence and error analysis
The accuracy of direct methods depends on the choice of the approximating functions and the size of the finite-dimensional subspace
Convergence analysis studies the behavior of the approximate solutions as the dimension of the subspace increases, with the goal of showing that the approximate solutions converge to the exact solution in a suitable norm
Error analysis provides bounds on the difference between the approximate and exact solutions, often in terms of the size of the finite elements or the degree of the approximating polynomials
A priori error estimates give bounds on the error before the approximate solution is computed, while a posteriori error estimates use information from the computed solution to assess the error and guide adaptive refinement strategies
Applications in control theory
Calculus of variations has numerous applications in control theory, where it is used to formulate and solve optimal control problems
seeks to determine the control inputs that minimize a given performance index (cost functional) while satisfying the system dynamics and any additional constraints
The performance index often includes terms that penalize the deviation from a desired state, the control effort, or the final time
Optimal control problems and formulations
An optimal control problem consists of a dynamical system (often described by ordinary or partial differential equations), a cost functional to be minimized, and possibly some constraints on the states and controls
The cost functional typically takes the form J[x,u]=∫t0tfL(x,u,t)dt+ϕ(x(tf),tf), where x represents the state variables, u represents the control inputs, L is the running cost, and ϕ is the terminal cost
The system dynamics are described by differential equations of the form x˙=f(x,u,t), where f is a given function
The goal is to find the optimal control u∗(t) that minimizes the cost functional while satisfying the system dynamics and any additional constraints
Pontryagin's maximum principle
is a fundamental result in optimal control theory that provides necessary conditions for the optimal control
It introduces the concept of the H(x,u,λ,t)=L(x,u,t)+λTf(x,u,t), where λ is a vector of adjoint variables (costate variables)
The maximum principle states that for the optimal control u∗(t), the Hamiltonian function attains its maximum with respect to u at each time t, i.e., H(x∗,u∗,λ∗,t)=maxuH(x∗,u,λ∗,t)
The adjoint variables satisfy a system of differential equations known as the adjoint equations, and the optimal state and control trajectories can be found by solving a two-point boundary value problem
Dynamic programming and Hamilton-Jacobi-Bellman equation
Dynamic programming is a powerful approach to solving optimal control problems based on the principle of optimality, which states that an optimal policy has the property that whatever the initial state and initial decision are, the remaining decisions must constitute an optimal policy with regard to the state resulting from the first decision
The Hamilton-Jacobi-Bellman (HJB) equation is a partial differential equation that characterizes the optimal cost-to-go function V(x,t), which represents the minimum cost incurred from time t onwards starting from state x
The HJB equation takes the form −∂t∂V=minu(L(x,u,t)+∂x∂Vf(x,u,t)), with boundary condition V(x,tf)=ϕ(x,tf)
Solving the HJB equation yields the optimal cost-to-go function, from which the optimal control can be derived as u∗(x,t)=argminu(L(x,u,t)+∂x∂Vf(x,u,t))
Numerical methods for variational problems
Numerical methods play a crucial role in solving variational problems that arise in control theory and other fields, as analytical solutions are often difficult or impossible to obtain
These methods discretize the problem in time and/or space, leading to finite-dimensional optimization problems that can be solved using various computational algorithms
Key aspects of numerical methods include the choice of discretization scheme, the handling of boundary conditions, and the efficient implementation of the algorithms
Discretization techniques and algorithms
convert the continuous variational problem into a finite-dimensional optimization problem by approximating the functions and integrals involved
Common discretization methods include finite difference methods, which approximate derivatives using difference quotients, and finite element methods, which approximate the solution using piecewise polynomial functions
The resulting discrete optimization problem can be solved using various algorithms, such as gradient-based methods (e.g., steepest descent, conjugate gradient), Newton's method, or interior-point methods
The choice of the algorithm depends on the structure of the problem, the desired accuracy, and the computational resources available
Shooting methods and boundary value problems
Shooting methods are a class of numerical techniques for solving boundary value problems, which often arise in the context of optimal control
The idea is to convert the boundary value problem into an initial value problem by guessing the initial values of the unknown variables and then adjusting them iteratively to satisfy the boundary conditions
Simple shooting methods use a single initial guess and propagate the solution forward in time, while multiple shooting methods divide the time interval into subintervals and use separate initial guesses for each subinterval
Shooting methods can be combined with optimization algorithms to solve the resulting nonlinear system of equations efficiently
Computational challenges and solutions
Numerical methods for variational problems often face computational challenges, such as high dimensionality, stiffness, and ill-conditioning
High-dimensional problems arise when the state space or the control space is large, leading to a large number of variables
Key Terms to Review (37)
Action Integral: The action integral is a fundamental concept in calculus of variations, defined as the integral of a Lagrangian function over time. This integral represents the total 'action' of a system, and is used to determine the path that a system will take by minimizing or extremizing this action. The principle of least action states that the actual path taken by a system between two states is the one that makes the action integral stationary, leading to equations of motion derived from this principle.
Brachistochrone problem: The brachistochrone problem is a famous problem in the calculus of variations that seeks the curve of fastest descent between two points, not necessarily directly vertical. This problem highlights the principle that the path taken by an object under the influence of gravity can be optimized, showing that a cycloid is the solution for the quickest descent, rather than a straight line.
Continuity: Continuity refers to the property of a function or a system where small changes in input result in small changes in output, ensuring that the function is unbroken and can be traced without jumps or interruptions. This concept is crucial when analyzing how variations affect outcomes and leads to smoother transitions in behavior, which is especially relevant when looking at how systems behave under changes.
David Hilbert: David Hilbert was a German mathematician who made significant contributions to various fields, particularly in the foundations of mathematics, algebra, and mathematical logic. He is best known for formulating the Hilbert problems, a list of 23 unsolved problems that guided much of 20th-century mathematics, influencing areas such as calculus of variations and optimization.
Differentiability: Differentiability refers to the property of a function that allows it to be differentiated, meaning that a derivative can be computed at a given point. This concept implies that the function has a well-defined tangent line at that point, which leads to various applications in optimization and modeling. When working with functions in calculus, understanding differentiability is essential because it determines how the function behaves locally and influences methods like the calculus of variations, where one seeks to find functions that optimize certain criteria.
Direct methods: Direct methods are techniques used in the calculus of variations that focus on finding solutions to optimization problems by directly assessing the properties of functional spaces and functionals. These methods rely on constructing appropriate test functions and applying variational principles to identify extremal functions without resorting to indirect approaches such as Euler-Lagrange equations.
Dirichlet boundary condition: A Dirichlet boundary condition specifies the values a solution must take on the boundary of the domain. It is a type of boundary condition often used in mathematical problems involving partial differential equations and variational calculus, where the solution is constrained to match given values at specific locations.
Discretization techniques: Discretization techniques are methods used to convert continuous models or equations into discrete counterparts, making them suitable for numerical analysis and computational applications. These techniques are essential for solving problems in various fields by transforming the dynamics of continuous systems into a format that can be handled by digital computers. This process often involves approximating derivatives and integrals with finite differences or summation methods, allowing for easier implementation in simulations and control algorithms.
Euler-Lagrange Equation: The Euler-Lagrange equation is a fundamental equation in the calculus of variations that provides a necessary condition for a function to be an extremum of a functional. This equation arises from the need to find the path or function that minimizes or maximizes a given functional, which is often expressed as an integral involving the function and its derivatives. Understanding this equation is key to solving problems in physics, engineering, and optimization, where the goal is to determine optimal trajectories or configurations.
Extremal: In the context of calculus of variations, an extremal refers to a function or curve that makes a functional reach its maximum or minimum value. Understanding extremals is crucial because they represent the solutions to variational problems, where one seeks to optimize a particular quantity, such as minimizing energy or maximizing distance.
Finite element method: The finite element method (FEM) is a numerical technique used to find approximate solutions to boundary value problems for partial differential equations. It works by breaking down a large problem into smaller, simpler parts called finite elements, which can be analyzed individually and then combined to create an overall solution. This method is widely used in engineering and physics for structural analysis, heat transfer, fluid dynamics, and more.
Functional: In mathematics, a functional is a specific type of mapping from a vector space into its field of scalars, often real or complex numbers. Functionals play a crucial role in various areas such as calculus of variations, where they are used to express quantities that depend on functions rather than just numerical values, allowing for optimization and analysis of functionals in order to find extrema.
Hamilton-Jacobi-Bellman Equation: The Hamilton-Jacobi-Bellman (HJB) equation is a partial differential equation used in optimal control theory that describes the value function of a control problem. It connects the optimal controls to the dynamics of the system and represents a necessary condition for optimality, providing a framework for finding the best possible strategy in dynamic programming problems.
Hamilton's Principle: Hamilton's Principle states that the actual path taken by a mechanical system between two points in time is the one for which the action integral is stationary (usually a minimum). This principle serves as a foundation for deriving the equations of motion in classical mechanics and connects to the calculus of variations, which involves finding optimal solutions among a set of possible functions.
Hamiltonian function: The Hamiltonian function is a fundamental concept in physics and mathematics that represents the total energy of a dynamic system, expressed in terms of its generalized coordinates and momenta. It is a central component in the Hamiltonian formulation of classical mechanics, which reformulates Newtonian mechanics to provide a powerful approach for analyzing the behavior of physical systems over time.
Holonomic constraints: Holonomic constraints are restrictions on a system's configuration that can be expressed as functions of the generalized coordinates and time, allowing the system to be described entirely in terms of position variables. These constraints are often integrable, meaning they can be derived from potential energy functions, which connects them directly to the dynamics of a system. Holonomic constraints play a crucial role in formulating problems in the calculus of variations, where the goal is to find the path or function that minimizes or maximizes a certain quantity while satisfying these constraints.
Indirect methods: Indirect methods refer to techniques used in calculus of variations that focus on solving optimization problems without directly computing the extremal paths. These methods often involve finding a relationship between the objective functional and the constraints, allowing for the derivation of solutions using alternative approaches such as Lagrange multipliers or variational principles.
Isoperimetric problems: Isoperimetric problems involve finding a shape that has the largest possible area for a given perimeter or the smallest perimeter for a given area. These problems are foundational in the field of calculus of variations and relate to optimizing geometric properties while adhering to specific constraints.
Jacobi Condition: The Jacobi Condition is a criterion in the calculus of variations used to determine whether a function can represent an extremum of a functional. This condition ensures that the second variation of the functional is non-negative for all variations, which is crucial for identifying local minima or maxima. Essentially, it helps to differentiate between potential extrema and those that do not meet necessary conditions for optimization.
Lagrange Multipliers: Lagrange multipliers are a mathematical tool used to find the local maxima and minima of a function subject to equality constraints. By introducing additional variables, known as Lagrange multipliers, the method transforms a constrained optimization problem into an unconstrained one, enabling the identification of optimal solutions while respecting the given constraints.
Lagrangian function: The Lagrangian function is a mathematical formulation used in classical mechanics and calculus of variations, defined as the difference between kinetic and potential energy of a system. It plays a crucial role in deriving the equations of motion for systems and optimizing functional outputs by providing a systematic way to analyze how systems evolve over time based on their energy states.
Legendre Condition: The Legendre Condition is a necessary condition for a function to be a local extremum in the calculus of variations. It states that the second derivative of the Lagrangian function with respect to the velocity variable must be non-negative at the optimal solution, indicating that the functional being minimized or maximized has a local minimum or maximum point.
Leonhard Euler: Leonhard Euler was an influential Swiss mathematician and physicist who made significant contributions to various fields, including calculus, graph theory, mechanics, and number theory. He is particularly known for his work in the calculus of variations, where he formulated the Euler-Lagrange equation, a cornerstone concept used to determine the optimal path of a functional.
Linear Functional: A linear functional is a specific type of linear map that transforms elements from a vector space into its underlying field, usually the real or complex numbers. It has the property of linearity, meaning it satisfies both additivity and homogeneity, allowing for the evaluation of vectors in a way that is compatible with the operations of addition and scalar multiplication. This concept is crucial in variational calculus as it helps define functionals that map functions to real numbers, facilitating optimization problems.
Minimum Surface Area Problem: The minimum surface area problem is a concept in calculus of variations that involves finding the shape or surface with the least area while enclosing a specific volume. This problem typically leads to the Euler-Lagrange equation, which is used to determine the optimal shape that minimizes surface area, illustrating the balance between geometric constraints and physical properties.
Necessary Conditions: Necessary conditions are criteria that must be satisfied for a certain outcome or theorem to hold true. In the realm of optimization and calculus, particularly when determining optimal solutions, necessary conditions outline the minimum requirements that must be met for a function to achieve an extremum, such as a minimum or maximum. Understanding these conditions helps in evaluating various problems, leading to the development of methods and principles aimed at finding optimal solutions.
Neumann Boundary Condition: The Neumann boundary condition specifies the derivative of a function on the boundary of a domain, often representing a physical quantity such as heat flux or pressure gradient. This type of condition is essential in problems where the behavior of the solution at the boundary is determined by its gradient rather than its value, making it a fundamental concept in variational calculus and partial differential equations.
Noether's Theorem: Noether's Theorem is a fundamental principle in theoretical physics and mathematics that establishes a deep connection between symmetries and conservation laws. It states that for every continuous symmetry of the action of a physical system, there exists a corresponding conserved quantity, meaning that if the action remains invariant under certain transformations, some physical property will not change over time. This theorem has significant implications for understanding the principles governing dynamic systems.
Non-holonomic constraints: Non-holonomic constraints are restrictions on a system's motion that cannot be integrated into positional coordinates, meaning they depend on both position and velocity. This concept is crucial when analyzing systems that have limitations on their movement, such as wheeled vehicles or robotic arms, making them more complex than holonomic systems, which can be described solely by their position.
Nonlinear functional: A nonlinear functional is a mapping from a function space to the real numbers that does not satisfy the properties of additivity and homogeneity. This means that for a nonlinear functional, the output does not behave in a predictable linear manner when inputs are combined or scaled. Understanding nonlinear functionals is crucial as they often arise in optimization problems, especially in the context of variational principles where the goal is to find functions that minimize or maximize these functionals.
Optimal Control Theory: Optimal control theory is a mathematical framework used to determine the best possible control inputs to achieve a desired outcome in dynamic systems. This theory focuses on optimizing a performance index, which quantifies the efficiency and effectiveness of the control strategy over time, while also taking into account constraints imposed by the system dynamics. The approach often involves sophisticated techniques like the calculus of variations to derive optimal solutions.
Pontryagin's Maximum Principle: Pontryagin's Maximum Principle is a mathematical framework used in optimal control theory to find the best possible control inputs that will minimize or maximize a certain performance index over a given time period. It connects the ideas of performance indices and calculus of variations by providing necessary conditions for optimality in dynamic systems, allowing the determination of optimal trajectories and control strategies through Hamiltonian functions.
Shape optimization: Shape optimization refers to the process of finding the best geometric configuration of a structure or domain to achieve a desired performance criterion, often in terms of minimizing cost or maximizing efficiency. This technique utilizes mathematical methods and calculus of variations to identify optimal shapes that satisfy specific constraints and objectives, such as reducing drag in aerodynamics or enhancing load-bearing capacity in structures.
Strong variations: Strong variations refer to a specific type of perturbation in the calculus of variations where the variations of a function are sufficiently large and can affect the extremum properties of the functional being considered. These variations help in establishing conditions under which a functional achieves its extrema, contributing significantly to the analysis and understanding of functionals in optimization problems.
Sufficient Conditions: Sufficient conditions refer to a set of criteria or requirements that, if met, ensure a certain outcome or conclusion is true. In various contexts, these conditions provide assurance that a particular statement holds, establishing a necessary link between the cause and effect. Understanding sufficient conditions is crucial for analyzing situations where multiple factors contribute to an outcome, allowing for clearer decision-making and logical reasoning.
Variational Principle: The variational principle is a fundamental concept in physics and mathematics that states that the path taken by a system between two states is the one for which a particular quantity, often an integral called the action, is stationary (usually minimized or maximized). This principle is key in deriving equations of motion and connecting various fields like classical mechanics, quantum mechanics, and optimal control.
Weak variations: Weak variations refer to a type of perturbation used in the calculus of variations, where the variations are considered in a weaker sense than traditional variations. This concept allows for the analysis of functionals that may not be differentiable or may not possess classical derivatives, enabling a broader set of functions to be included in optimization problems.