The maximum principle for positional controls and the problems of optimal system synthesis (Q1361961)
From MaRDI portal
| This is the item page for this Wikibase entity, intended for internal use and editing purposes. Please use this page instead for the normal view: The maximum principle for positional controls and the problems of optimal system synthesis |
scientific article; zbMATH DE number 1042173
| Language | Label | Description | Also known as |
|---|---|---|---|
| English | The maximum principle for positional controls and the problems of optimal system synthesis |
scientific article; zbMATH DE number 1042173 |
Statements
The maximum principle for positional controls and the problems of optimal system synthesis (English)
0 references
17 May 1998
0 references
The authors study the problem \[ \Phi \bigl(x(t_1) \bigr)\to \min, \quad \dot x= f(x,u,t) \] \[ u(x,t)\in V, \quad (x,t)\in\mathbb{R}^n \times T, \] where the optimal control is sought in the class of positional controls \(u=u(x,t)\). An analogue of Pontryagin's maximum principle directly in the class of positional controls is derived. The relations with the classical maximum principle and the dynamic programming method are considered. Examples are presented in which the Bellman function is not continuously differentiable so that there is no formal justification for using the classical Bellman equation whereas the positional maximum principle obtained in the paper successfully solves the problem of optimal system design.
0 references
necessary conditions
0 references
optimal system synthesis
0 references
positional controls
0 references
Pontryagin's maximum principle
0 references
dynamic programming
0 references
Bellman function
0 references