A Chebyshev-Based State Representation for Linear Quadratic Optimal Control

1993 ◽  
Vol 115 (1) ◽  
pp. 1-6 ◽  
Author(s):  
M. L. Nagurka ◽  
S.-K. Wang

A computationally attractive method for determining the optimal control of unconstrained linear dynamic systems with quadratic performance indices is presented. In the proposed method, the difference between each state variable and its initial condition is represented by a finite-term shifted Chebyshev series. The representation leads to a system of linear algebraic equations as the necessary condition of optimality. Simulation studies demonstrate computational advantages relative to a standard Riccati-based method, a transition matrix method, and a previous Fourier-based method.

1991 ◽  
Vol 113 (2) ◽  
pp. 206-215 ◽  
Author(s):  
V. Yen ◽  
M. Nagurka

A method for determining the optimal control of unconstrained and linearly constrained linear dynamic systems with quadratic performance indices is presented. The method is based on a modified Fourier series approximation of each state variable that converts the linear quadratic (LQ) problem into a mathematical programming problem. In particular, it is shown that an unconstrained LQ problem can be cast as an unconstrained quadratic programming problem where the necessary condition of optimality is derived as a system of linear algebraic equations. Furthermore, it is shown that a linearly constrained LQ problem can be converted into a general quadratic programming problem. Simulation studies for constrained LQ systems, including a bang-bang control problem, demonstrate that the approach is accurate. The results also indicate that in solving high order unconstrained LQ problems the approach is computationally more efficient and robust than standard methods.


2013 ◽  
Vol 2013 ◽  
pp. 1-9 ◽  
Author(s):  
Guiling Li ◽  
Weihai Zhang

This paper studies the indefinite stochastic linear quadratic (LQ) optimal control problem with an inequality constraint for the terminal state. Firstly, we prove a generalized Karush-Kuhn-Tucker (KKT) theorem under hybrid constraints. Secondly, a new type of generalized Riccati equations is obtained, based on which a necessary condition (it is also a sufficient condition under stronger assumptions) for the existence of an optimal linear state feedback control is given by means of KKT theorem. Finally, we design a dynamic programming algorithm to solve the constrained indefinite stochastic LQ issue.


2020 ◽  
Vol 26 ◽  
pp. 41
Author(s):  
Tianxiao Wang

This article is concerned with linear quadratic optimal control problems of mean-field stochastic differential equations (MF-SDE) with deterministic coefficients. To treat the time inconsistency of the optimal control problems, linear closed-loop equilibrium strategies are introduced and characterized by variational approach. Our developed methodology drops the delicate convergence procedures in Yong [Trans. Amer. Math. Soc. 369 (2017) 5467–5523]. When the MF-SDE reduces to SDE, our Riccati system coincides with the analogue in Yong [Trans. Amer. Math. Soc. 369 (2017) 5467–5523]. However, these two systems are in general different from each other due to the conditional mean-field terms in the MF-SDE. Eventually, the comparisons with pre-committed optimal strategies, open-loop equilibrium strategies are given in details.


Author(s):  
Andrea Pesare ◽  
Michele Palladino ◽  
Maurizio Falcone

AbstractIn this paper, we will deal with a linear quadratic optimal control problem with unknown dynamics. As a modeling assumption, we will suppose that the knowledge that an agent has on the current system is represented by a probability distribution $$\pi $$ π on the space of matrices. Furthermore, we will assume that such a probability measure is opportunely updated to take into account the increased experience that the agent obtains while exploring the environment, approximating with increasing accuracy the underlying dynamics. Under these assumptions, we will show that the optimal control obtained by solving the “average” linear quadratic optimal control problem with respect to a certain $$\pi $$ π converges to the optimal control driven related to the linear quadratic optimal control problem governed by the actual, underlying dynamics. This approach is closely related to model-based reinforcement learning algorithms where prior and posterior probability distributions describing the knowledge on the uncertain system are recursively updated. In the last section, we will show a numerical test that confirms the theoretical results.


Axioms ◽  
2021 ◽  
Vol 10 (3) ◽  
pp. 137
Author(s):  
Vladimir Turetsky

Two inverse ill-posed problems are considered. The first problem is an input restoration of a linear system. The second one is a restoration of time-dependent coefficients of a linear ordinary differential equation. Both problems are reformulated as auxiliary optimal control problems with regularizing cost functional. For the coefficients restoration problem, two control models are proposed. In the first model, the control coefficients are approximated by the output and the estimates of its derivatives. This model yields an approximating linear-quadratic optimal control problem having a known explicit solution. The derivatives are also obtained as auxiliary linear-quadratic tracking controls. The second control model is accurate and leads to a bilinear-quadratic optimal control problem. The latter is tackled in two ways: by an iterative procedure and by a feedback linearization. Simulation results show that a bilinear model provides more accurate coefficients estimates.


Author(s):  
Nacira Agram ◽  
Bernt Øksendal

The classical maximum principle for optimal stochastic control states that if a control [Formula: see text] is optimal, then the corresponding Hamiltonian has a maximum at [Formula: see text]. The first proofs for this result assumed that the control did not enter the diffusion coefficient. Moreover, it was assumed that there were no jumps in the system. Subsequently, it was discovered by Shige Peng (still assuming no jumps) that one could also allow the diffusion coefficient to depend on the control, provided that the corresponding adjoint backward stochastic differential equation (BSDE) for the first-order derivative was extended to include an extra BSDE for the second-order derivatives. In this paper, we present an alternative approach based on Hida–Malliavin calculus and white noise theory. This enables us to handle the general case with jumps, allowing both the diffusion coefficient and the jump coefficient to depend on the control, and we do not need the extra BSDE with second-order derivatives. The result is illustrated by an example of a constrained linear-quadratic optimal control.


Sign in / Sign up

Export Citation Format

Share Document