Next Article in Journal
Refinements on Some Classes of Complex Function Spaces
Next Article in Special Issue
A Survey on Knowledge Graph Embeddings for Link Prediction
Previous Article in Journal
Engineering Graphics for Thermal Assessment: 3D Thermal Data Visualisation Based on Infrared Thermography, GIS and 3D Point Cloud Processing Software
Previous Article in Special Issue
Stochastic Diffusion Model for Analysis of Dynamics and Forecasting Events in News Feeds
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Research of Trajectory Optimization Approaches in Synthesized Optimal Control

by
Askhat Diveev
and
Elizaveta Shmalko
*,†
Federal Research Center “Computer Science and Control”, Russian Academy of Sciences, 119333 Moscow, Russia
*
Author to whom correspondence should be addressed.
These authors contributed equally to this work.
Symmetry 2021, 13(2), 336; https://doi.org/10.3390/sym13020336
Submission received: 19 January 2021 / Revised: 7 February 2021 / Accepted: 9 February 2021 / Published: 18 February 2021
(This article belongs to the Special Issue 2020 Big Data and Artificial Intelligence Conference)

Abstract

:
This article presents a study devoted to the emerging method of synthesized optimal control. This is a new type of control based on changing the position of a stable equilibrium point. The object stabilization system forces the object to move towards the equilibrium point, and by changing its position over time, it is possible to bring the object to the desired terminal state with the optimal value of the quality criterion. The implementation of such control requires the construction of two control contours. The first contour ensures the stability of the control object relative to some point in the state space. Methods of symbolic regression are applied for numerical synthesis of a stabilization system. The second contour provides optimal control of the stable equilibrium point position. The present paper provides a study of various approaches to find the optimal location of equilibrium points. A new problem statement with the search of function for optimal location of the equilibrium points in the second stage of the synthesized optimal control approach is formulated. Symbolic regression methods of solving the stated problem are discussed. In the presented numerical example, a piece-wise linear function is applied to approximate the location of equilibrium points.

1. Introduction

Russian neurophysiologist, Nobel laureate Ivan Pavlov in his Lectures on the work of the cerebral hemisphere [1] said: “Every material system can exist as an entity only so long as its internal forces of attraction, cohesion, etc., are equilibrated with the external forces influencing it. This applies in equal measure to such a simple object as a stone and to the most complex chemical substance, and it also holds good for the organism. As a definite material system complete in itself, the organism can exist only so long as it is in equilibrium with the environment”. Pavlov assumed that all complex physiological organisms aimed to balance external influences. In this regard, new modern computational methods can be considered to have emerged to counterbalance new complex computational problems.
The present work is devoted to the development and application of the synthesized optimal control method [2]. It was developed according to Pavlov’s paradigm for balancing the challenges of new complex control problems. It is a new approach to optimal control based on the control of the equilibrium point of a stabilized object. The approach is in two stages. Initially, a stabilization system is introduced into the control object, so the object obtains a special property—a special stable equilibrium point appears. It is known that the object possess good properties for control when its mathematical model in the phase space has a stable equilibrium point. The equilibrium point can be changed after some time, but the object, being stabilized, maintains equilibrium at each moment of time. Then, the object can be controlled by optimally changing the position of the stable equilibrium point.
The key stage of the considered approach is the solution of the control synthesis problem for constructing the stabilization system. The challenge is to find a mathematical expression for the control function. In the general case, this control function must have a special property: when it is inserted into the differential equations of the mathematical model of the control object, then a special stable equilibrium point appears in the space of solutions of these differential equations. All solutions of differential equations from a certain region of initial conditions will tend to this singular point of attraction.
Most of the known methods of control synthesis are associated with specifics of the model of the control object [3,4,5,6]; for example, analytical methods [7,8,9,10], as well as the use of various controllers [11,12,13]. However, there are now general numerical approaches to solve the synthesis problem based on application of symbolic regression methods [14,15,16,17,18,19]. These can be applied to find a solution without reference to specific model equations.
These methods belong to the class of machine learning control [20,21,22]. They allow automatically by means of computer searching for the desired mathematical expression in a coded form. When it comes to a machine search for mathematical expressions, it is usually assumed that the researcher firstly determines a structure of the mathematical expression accurate to parameters, and then the computer looks for the optimal values of these parameters in accordance with some given criterion. Popular neural networks [23,24,25,26,27] are also used for these purposes. In this respect, neural networks are also functions of a given structure, in which it is necessary to find the optimal values of a large number of parameters. Symbolic regression methods, in contrast to neural networks, allow to search for the optimal structure of the desired function together with the optimal parameters of this function. In this context, this approach is more unambiguous and universal. Moreover, the received function is described in a way which is understandable to a researcher, not a black box as in the case with neural networks. It can be said that symbolic regression is a generalization of neural networks. Therefore, all analytical problems, where solutions must be obtained in the form of functions, can now be solved automatically by computers. Nonlinear and differential equations, integrals, inverse functions and other problems can now be solved by numerical symbolic regression methods.
Thus, symbolic regression methods allow a mathematical expression describing the stabilization system that ensures the presence of an equilibrium point in the state space to be found. According to the approach, the synthesized function includes a set of parameters that affect the position of the equilibrium point. Optimal control is realized by changing the position of the equilibrium point. In previous works on synthesized optimal control [2,28], the theoretical foundations of the method were presented and comparative estimates with other known approaches to solving the optimal control problem were given. The emphasis was placed on the substantiation of an approach to control based on the equilibrium point and on the numerical solution of the synthesis problem based on symbolic regression methods. One of the main features of the synthesized approach is in practical feasibility of the received numerical solutions of the optimal control problem for complex systems. In applied devices, the control system requires additional adjustments because there are always some differences between real control systems and analytical solutions made using models of the control objects. Thus, to implement analytical solutions into practice, it is necessary to construct additional stabilization systems, which modify the object model as a whole, significantly complicating it. As a result, analytical trajectories are no longer optimal. The synthesized solution can be implemented to the object without additional construction of a stabilization system.
In the present paper, the research of the synthesized optimal control approach is aimed at studying ways to control equilibrium points. Particularly, this can be just a set of points. In the general case, the vector of parameters that effect the positions of equilibrium points can be considered as some function. In this work, the problem statement of optimal movement according to a set of equilibrium points is presented (Section 2), methods for solving the stated problem are proposed (Section 3 and Section 4) and a computational example (Section 5) is provided to demonstrate how the approach performs the task with piece-wise linear functions defining the positions of the equilibrium points.

2. Problem of Optimal Movement on Trajectory Determined by a Set of Points

The mathematical model of a control object is given in the form of the ordinary differential equation system
x ˙ = f ( x , u ) ,
where x R n , u U R m , U is a compact set, m n .
Given initial conditions
x ( 0 ) = x 0 .
The ordered set of points is given
Y = ( y * , 1 , , y * , N ) ,
where y R l , l n .
The formulas of transformation from state space vector x to the vector y are determined
y = w ( x ) ,
where
w ( x ) : R n R l .
The control is to be found in the form
u = v ( t ) U
to deliver minimum to the functional
J = t f + a 1 k = 1 N min t y * , k w ( x ( t ) ) min u U ,
where t f is a finishing time,
t f = t , if t < t + and y * , N w ( x ( t ) ) ε t + , otherwise ,
a 1 is a weight coefficient, t + and ε are given positive numbers.
To solve this problem (1)–(8), the control function can be found as
u = h ( g ( k , t ) w ( x ( t ) ) ) U ,
where
g ( k , t ) = y * , k , if ( t t k 1 < Δ ) ( y * , k w ( x ( t ) ) > ε 1 ) k k + 1 , t k t , otherwise ,
k = 1 , , N 1 , t 0 = 0 , g ( N , t ) = y * , N , Δ and ε 1 are given small positive numbers.

3. Synthesized Optimal Control

For solution of the problem (1)–(8), the synthesized optimal control approach was used. According to the approach, firstly, it is necessary to solve the control synthesis problem and to provide existence of a stable equilibrium point in the phase space of differential Equation (1). For this purpose, the control function with the following properties must be found
u = g ( x * x ) ,
where x * is a constant vector, x * = [ x 1 * x n * ] T . The function is such that if it is inserted into right side of differential Equations (1), there is an equilibrium point in the state space
x ˜ ( x ˜ * ) R n
f ( x ˜ ( x ˜ * ) , g ( x ˜ x * ) ) = 0 ,
det ( λ E A ) = λ n + b n 1 λ n 1 + + b 1 λ + b 0 = j = 1 n ( λ λ j ) = 0 ,
λ j = α j + i β j , j = 1 , , n ,
where
α j < 0 , j = 1 , n ,
i = 1 ,
A = f ( x ˜ ( x * ) , g ( x * x ˜ ( x * ) ) ) x ,
E is n × n unit matrix.
In the result of the synthesis problem solution, a new mathematical model of control object is obtained
x ˙ = f ( x , g ( x * x ) ) .
In this model (18), there is a new control vector x * . This vector x * has the same dimension n as the main state space vector x . There is a domain of values
x * X * R n
where the system (18) always has a stable equilibrium point x ˜ ( x * ) in the phase space. Therefore, the system (18) is controlled by changing the position of the equilibrium point.
Now, in the second stage of the synthesized optimal control approach, it is necessary to find a control function in the form
x * = v ( t )
for optimization of criterion (7).
The found control can be implemented directly in the on-board computer of the control object. Here, it is not necessary to design a stabilization system for movement of the control object in an optimal trajectory.
When solving the control synthesis problem, the given trajectory (3) is not taken into account; therefore, in the second stage, the optimal control problem can be solved for any given trajectory.

4. Numerical Methods

4.1. Methods for the Control Synthesis Problem

The control synthesis problem is the most difficult with regard to realization of the synthesized optimal control. In control theory, for the synthesis problem, different analytical methods are used. These are backstepping integrator [8], design of aggregated controller [9], synthesis on the base of Lyapunov functions [10] and others. All analytical methods are not universal as they are connected with the mathematical model of the control object. Technical methods have the same disadvantages [11]. To insert a controller in the control object, it is necessary to know which control vector component affects which state vector component. Transformation of the mathematical model to the linear equation system can be made correctly only for stable system in neighborhood of a stable equilibrium point.
In the synthesized optimal control approach, to solve the control synthesis problem, machine learning control by symbolic regression is applied. These methods code the mathematical expression on the base of elementary functions sets and search the optimal code of mathematical expression by a special genetic algorithm.
Symbolic regression learning of control is a universal numerical tool for the synthesis problem and it is independent of the mathematical model of the control object. However, the application of symbolic regression methods is very difficult since the search for the function must be carried out in a space where there is no metric. Furthermore, there are only a few research groups in the world dealing with this issue. The majority uses genetic programming [16,17,20,29,30,31]. It codes a mathematical expression in the form of a computing tree and presents this code in the form of an array in computer memory. GP has a computational drawback that consists of the fact that different mathematical expressions have codes of different lengths. There are many other symbolic regression methods that have fixed lengths of code, such as Cartesian genetic programming [32], network operator [33] and complete binary analytic programming [34]. In the application example of the present paper, the method of the network operator was applied. It is based on the principle of small variations of the basic solution [35]. The main idea of the principle is that the search for a function begins not with a random possible solution, but with some given possible solution, which is called the basic solution. The basic solution can also be random if there is no preliminary data on the desired function. However, if the developer has some experience or intuition, then he can put them in the description of the basic solution. The application of this principle helps to significantly narrow the search space and overcome the mentioned difficulty of search space complexity, which allows for the synthesis problems to be solved in a reasonable time.
To solve the control synthesis problem, it is necessary for the mathematical model of the control object (1) to give a set of initial conditions
X 0 = { x 0 , 1 , , x 0 , K } R n ,
and one terminal condition
x * R n .
The found optimal control function (11) has to satisfy the optimal value of the criterion
J s = max { t f , i : i = 1 , , K } min ,
where t f , i is the time it takes to achieve the terminal condition (22) from the initial condition x 0 , i , i = 1 , , K ,
t f , i = t , if t < t + and x * x ( t , x 0 , i ) ε 0 t + , otherwise ,
x ( t , x 0 , i ) is a partial solution of the differential equation system (1) from the initial condition x 0 , i , i { 1 , , K } .
When solving the control synthesis problem (1), (21)–(24) every control function (11) is inserted into the right part of the mathematical model (18) in the form of a special symbolic regression code. For calculation of the criterion (23), a special algorithm calculates a value of control by the code of control function (11).

4.2. Methods for the Optimal Control Problem

The formal mathematical statement for the optimal control problem in the second stage consists of that for the mathematical model of the control object (18). It is necessary to find control in the form (20) in order to provide the optimal value of the quality criterion (7).
For numerical solution of the optimal control problem, the axis of time was divided into some intervals. In each interval, the control function (20) is approximated as a function of time defined with accuracy to values of some parameters. This allows for the transformation of the optimal control problem from infinite dimension optimization to a nonlinear programming problem in the space with finite dimension. The dimension of the search space is equal to a number of parameters in each interval multiplied by the number of intervals. For example, if the control function is approximated in each interval as a piece-wise linear function, then a number of parameters is equal to the number of intervals plus one. The piece-wise linear control function (20) has the following form
x i * = q k + ( i 1 ) d + ( q k + ( i 1 ) d + 1 q k + ( i 1 ) d ) t k Δ t Δ t ,
where q = [ q 1 q d n ] T is the vector of searched parameters, k is the interval number, k Δ t t < ( k + 1 ) Δ t , Δ t is a time interval, d is a number of interval boundaries, k = 1 , , d 1 , i = 1 , , n .
To solve the optimization problem, various numerical methods of optimization can be used. However, many studies have shown [36,37,38,39] that the most effective methods for solving the problem are evolutionary algorithms, in so as far as, generally, the objective functional could be not unimodal.

5. Computational Example

As an example, consider the problem of optimal control for one mobile robot following the given trajectories.
The mathematical model of a robot has the following form
x ˙ 1 = 0.5 ( u 1 + u 2 ) cos ( x 3 ) , x ˙ 2 = 0.5 ( u 1 + u 2 ) sin ( x 3 ) , x ˙ 3 = 0.5 ( u 1 u 2 ) ,
where x = [ x 1 x 2 x 3 ] T is a vector of state space, u = [ u 1 u 2 ] T is a control vector.
Restrictions on control are given
10 u i 10 , i = 1 , 2 .
Initial conditions are set
x i ( 0 ) = 0 , i = 1 , 2 , 3 .
The program trajectory is given as an ordered set of N = 40 points
Y = ( y * , 1 = [ y 1 * , 1 x 2 * , 1 y 3 * , 1 ] T , , y * , 40 = [ y 1 * , 40 y 2 * , 40 y 3 * , 40 ] T ) .
In this example, spaces R l and R n coincide
y = y 1 y 2 y 3 = w ( x ) = 1 0 0 0 1 0 0 0 1 x 1 x 2 x 3 .
In the first stage, the synthesis problem (1), (21)–(24) was solved. For the solution, the network operator method was used. As a result, the following control function was received
u i = 10 , if u ˜ i > 10 10 , if u ˜ i < 10 u ˜ i , otherwise , i = 1 , 2 ,
where
u ˜ 1 = ρ 16 ( A ) + sgn ( x 3 * x 3 ) + ρ 19 ( r 3 ( x 3 * x 3 ) ) + D 3 + ( ρ 23 ( A ) + D ) 1 ,
u ˜ 2 = ρ 23 ( sin ( x 1 * x 1 ) + ρ 19 ( r 2 ( x 2 * x 2 ) ) + tanh ( r 1 ( x 1 * x 1 ) ) + r 2 ( x 2 * x 2 ) + r 3 ( x 3 * x 3 ) ) +
u ˜ 1 + sgn ( x 1 * x 1 ) + ρ 16 ( A ) + arctan ( B ) + sin ( E ) ,
A = x 1 * x 1 + tanh ( tanh ( r 1 ( x 1 * x 1 ) ) + r 2 ( x 2 * x 2 ) sgn ( x 1 * x 1 ) + r 3 ( x 3 * x 3 ) ) +
sin ( x 1 * x 1 ) + ρ 19 ( r 2 ( x 2 * x + 2 ) ) + tanh ( r 1 ( x 1 * x 1 ) ) + r 2 ( x 2 * x 2 ) sgn ( x 1 * x 1 ) + r 3 ( x 3 * x 3 ) ,
B = ρ 19 ( r 2 ( x 2 * x 2 ) sgn ( x 1 * x 1 ) ) + tanh ( r 1 ( x 1 * x 1 ) ) + r 2 ( x 2 * x 2 ) sgn ( x 1 * x 1 ) + r 3 ( x 3 * x 3 ) ) +
arctan ( r 1 ( x 1 * x 1 ) ) + ρ 4 ( sin ( x 1 * x 1 ) + sgn ( A ) + x 1 * x 1 3 + A 3 +
ρ 4 ( r 1 ( x 1 * x 1 ) ) + tanh ( z 7 ) + r 2 ( x 2 * x 2 ) sgn ( x 1 * x 1 ) + r 3 ( x 3 * x 3 ) + sgn ( z 12 ) + x 1 * x 1 3 + A ,
C = ρ 9 ( x 3 * x 3 ) + arctan ( r 1 ) + sgn ( r 1 ( x 1 * x 1 ) ) + ρ 23 ( sin ( x 1 * x 1 ) + B +
ρ 19 ( r 2 ( x 2 * x 2 ) sgn ( x 1 * x 1 ) ) + tanh ( r 1 ( x 1 * x 1 ) ) + r 2 ( x 2 * x 2 ) sgn ( x 1 * x 1 ) + r 3 ( x 3 * x 3 ) ) ,
D = sin ( r 3 ( x 3 * x 3 ) ) + sin ( x 1 * x 1 ) + ρ 19 ( r 2 ( x 2 * x 2 ) sgn ( x 1 * x 1 ) + tanh ( r 1 ( x 1 * x 1 ) ) +
r 2 ( x 2 * x 2 ) sgn ( x 1 * x 1 ) + r 3 ( x 3 * x 3 ) + x 1 * x 1 3 + A 3 + tanh ( C ) ,
E = ρ 16 ( A ) + sgn ( x 3 * x 3 ) + ρ 19 ( r 3 ( x 3 * x 3 ) ) + D 3 + ( ρ 23 ( A ) + D ) 1 ρ 16 ( z ) = z , if | z | < 1 sgn ( z ) , otherwise , ρ 19 ( z ) = sgn ( z ) exp ( | z | ) , ρ 23 ( z ) = z z 3 , ρ 4 ( z ) = sgn ( z ) | z | , ρ 9 ( z ) = 1 , if z > 0 0 , otherwise , r 1 = 14.72876 r 2 = 2.02710 r 3 = 4.02222 z 9 = r 3 x 3 z 10 = ρ 8 ( z 7 ) + z 8 + z 9 z 11 = ρ 12 ( x 1 ) + ρ 19 ( z 8 ) + z 10 z 12 = x 1 + ρ 8 ( z 10 ) + z 11 z 13 = ρ 15 ( x 1 ) + z 12 z 14 = ρ 4 ( z 7 ) + z 10 + ρ 10 ( z 12 ) + z 13 z 15 = ρ 13 ( z 7 ) + ρ 4 ( z 11 ) + ρ 10 ( z 12 ) + ρ 15 ( z 13 ) + z 14 z 16 = ρ 9 ( x 3 ) + ρ 13 ( r 1 ) + ρ 10 ( z 7 ) + ρ 23 ( z 11 ) + z 15 z 17 = ρ 12 ( z 9 ) + z 11 + ρ 14 ( z 13 ) + ρ 8 ( z 16 ) z 18 = z 17 z 19 = z 18 , z 20 = ρ 23 ( z 12 ) + z 19 z 21 = ρ 10 ( x 3 ) + ρ 19 ( z 9 ) + ρ 15 ( z 18 ) + ρ 5 ( z 20 ) z 22 = ρ 16 ( z 12 ) + z 21 u 1 = z 22 u 2 = ρ 10 ( x 1 ) + ρ 23 ( z 11 ) + ρ 16 ( z 12 + ρ 13 ( z 15 ) + ρ 12 ( z 22 ) + u 1
In the second stage, the optimal control problem (7), (18), (20) was solved. The control function was searched in the form (25).
In the first experiment, the goal trajectory describes function sinus on the plane { x 1 ; x 2 }
y i * , 1 = 0 , i = 1 , 2 , 3
y 1 * , j = y 1 * , j 1 + Δ 1 , j = 2 , , 40 ,
y 2 * , j = 10 sin y 1 * , j π 10 , j = 2 , , 40 ,
y 3 * , j = arctan y 2 * , j y 2 * , j 1 Δ 1 , j = 2 , , 40 .
where Δ 1 = 0.5 .
In the search, the following values of parameters were used: Δ t = 0.28 , t + = 5.6 . In total, there were 20 intervals and d = 21 parameters for each component. To find vector a with 63 components, q = [ q 1 q 63 ] T . The parameters had the following constraints
2 q 1 + 3 k 22 12 q 2 + 3 k 12 0.5 π q 3 + 3 k 0.5 π , k = 1 , , 21 .
To solve the problem particle swarm optimization [40] (PSO) algorithm was applied.
As a result, the following solution was found:
q = [ 18.074 16.838 18.12 11.388 18.379 11.768 11.513 19.312 6.4712 20.686 12.367 16.460 14.280 21.869 21.999 21.974 21.204 21.994 21.992 22.000 22.000 10.438 4.645 3.847 8.718 2.988 5.609 5.462 3.956 4.437 3.960 0.784 0.185 7.510 10.515 9.344 9.5199.0666.529 2.842 2.623 1.674 0.487 1.236 1.547 0.310 1.0902 0.514 0.015 0.005 0.492 0.801 0.651 1.535 0.881 1.099 0.505 0.272 1.379 1.567 0.894 1.404 1.547 ] T
The value of functional (7) is J = 6.4004 . In Figure 1 projections of trajectories on the plane { x 1 ; x 2 } are presented. In the Figure 1 blue line is the given trajectory, black line is the trajectory of mobile robot with found optimal control.
In Figure 2, Figure 3 and Figure 4 plots of optimal trajectories x i ( t ) , i = 1 , 2 , 3 (black lines) and optimal controls x i * ( t ) , i = 1 , 2 , 3 (red lines) are presented.
As seen from the plots, optimal trajectories and their goal control trajectory of the position of equilibrium points do not coincide. According to synthesized control, the vector x * ( t ) in control function (25) determines the position of a stable equilibrium point in the current moment, t. However, according to the criterion (7) the control object should strive to reach ordered points on the trajectory as quickly as possible, so the control point x * should be located behind the point on the trajectory. The control object must not stop at the trajectory point, and it must go through it or near it. The object moves more slowly near the equilibrium point than far from it. Therefore, this is the reason why the equilibrium points trajectory coincides with the trajectory of the object movement.
In the second experiment, the trajectory in the form of a square was considered. The points were defined by the following Equations (39) and
y 1 * , i = y 1 * , i 1 + Δ 2 , i = 2 , , 11 ,
y 2 * , i = 0.5 y 1 * , i , i = 2 , , 11 ,
y 3 * , i = arctan y 2 * , i y 2 * , i 1 Δ 2 , i = 2 , , 11 ,
where Δ 2 = 0.7 ,
y 1 * , i = y 1 * , i 1 + Δ 3 , i = 12 , , 21 ,
y 2 * , i = 2 y 1 * , i + 17.5 , i = 12 , , 21 ,
y 3 * , i = arctan y 2 * , i y 2 * , i 1 Δ 3 , i = 12 , , 21 ,
where Δ 3 = 0.35 ,
y 1 * , i = y 1 * , i 1 + Δ 4 , i = 22 , , 31 ,
y 2 * , i = 0.5 y 1 * , i 8.75 , i = 22 , , 31 ,
y 3 * , i = arctan y 2 * , i y 2 * , i 1 Δ 4 , i = 22 , , 31 ,
where Δ 4 = 0.7 ,
y 1 * , i = y 1 * , i 1 + Δ 5 , i = 32 , , 41 ,
y 2 * , i = 2 y 1 * , i , i = 32 , , 41 ,
y 3 * , i = arctan y 2 * , i y 2 * , i 1 Δ 5 , i = 32 , , 41 ,
where Δ 5 = 0.35 ,
The optimal control problem had the following parameters t + = 4.8 , Δ t = 2.4 . The vector of parameters included 63 components. The constraints on the parameters were
2 q 1 + 3 k 15 12 q 2 + 3 k 8 0.5 π q 3 + 3 k 0.5 π , k = 1 , , 21 .
Particle swarm optimization [40] (PSO) algorithm was used for the solution.
The optimal values of parameters were
q = [ 11.713 12.049 11.075 2.216 7.729 9.378 6.723 12.249 14.714 14.395 12.629 9.562 0.576 1.997 1.998 0.180 0.832 1.976 1.860 1.963 1.963 2.887 2.593 5.988 1.803 0.593 4.089 6.419 1.526 1.047 2.163 2.196 9.116 6.031 5.418 7.228 2.923 6.839 2.463 0.255 3.917 0.131 0.403 1.461 1.200 0.338 1.471 0.053 1.431 0.213 0.591 0.096 1.451 0.7988 0.934 0.451 0.348 1.0291 1.512 0.566 0.125 1.424 1.025 ] T .
The value of the functional (7) was 5.0544 . Projections on the horizontal plane { x 1 ; x 2 } of given (blue lines) and found (black lines) trajectories are presented in Figure 5.
Plots of optimal trajectories and controls are presented in Figure 6, Figure 7 and Figure 8.
As can be seen from the graphs, the obtained control based on the synthesized optimal control allows for the problem to be solved with a given accuracy. In Figure 6, Figure 7 and Figure 8, the effect of non-coincidence of the trajectories of the equilibrium points and the trajectories of the object’s movement is also observed, which is quite understandable, as in the first example, in view of the presence of a performance criterion in the functional.

6. Conclusions

The optimal control problem of movement along a given trajectory determined in the form of a points set is considered. The quality criterion consists of trajectory travel time and accuracy. To solve the problem, the synthesized optimal control was used. The main advantage of the solution achieved using this approach is that it can be realized on the on-board computer without additional construction of a stabilization system. According to synthesized optimal control, firstly the control synthesis problem is solved and then a control is performed by changing equilibrium point location. The position of the equilibrium point is defined by some parametric vector, which corresponds to the state vector. In this work, for the first time—in contrast to other works devoted to the synthesized approach—an optimal function of time is used to control the equilibrium point position. In the experiment, two different trajectories for a mobile robot consisting of forty points were considered. For the control synthesis problem, a machine learning control method of the network operator was applied. Computational experiments have shown that the found optimal trajectories of the parametric vector do not coincide with the points of the trajectory, but the control object moves along the given trajectory fast enough and with a high accuracy.
Thus, we have demonstrated that the synthesized optimal control method is a mathematical approach to solve the optimal control problem in terms of the quality functional. The approach is quite universal as it is not tied to specific properties of the control object model or a type of trajectory, nor does it require manual selection of control channels or selection and adjustment of controllers. The development of the control system occurs automatically using modern computational machine learning methods.

Author Contributions

Conceptualization, A.D. and E.S.; methodology, A.D.; software, A.D.; validation, E.S.; formal analysis, A.D.; investigation, E.S.; data curation, E.S.; writing–original draft preparation, A.D. and E.S.; writing–review and editing E.S. All authors have read and agreed to the published version of the manuscript.

Funding

This research is supported by the Ministry of Science and Higher Education of the Russian Federation, project No. 075-15-2020-799.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Pavlov, I. Experimental Psychology and Other Essays; Philosophical Library: New York, NY, USA, 1957. [Google Scholar]
  2. Diveev, A.; Shmalko, E.; Serebrenny, V.; Zentay, P. Fundamentals of Synthesized Optimal Control. Mathematics 2021, 9, 21. [Google Scholar] [CrossRef]
  3. Liu, P.; Yu, H.; Cang, S. Trajectory Synthesis and Optimization of an Underactuated Microrobotic System with Dynamic Constraints and Couplings. Int. J. Control. Autom. Syst. 2018, 16, 2373–2383. [Google Scholar] [CrossRef]
  4. Matschek, J.; Bäthge, T.; Faulwasser, T.; Findeisen, R. Nonlinear predictive control for trajectory tracking and path following: An introduction and perspective. In Handbook of Model Predictive Control; Birkhäuser: Cham, Switzerland, 2019; pp. 169–198. [Google Scholar]
  5. Chen, H.; Mitra, S. Synthesis and verification of motor-transmission shift controller for electric vehicles. In Proceedings of the 2014 ACM/IEEE International Conference on Cyber-Physical Systems, ICCPS 2014, Art. No. 6843708, Berlin, Germany, 14–17 April 2014; pp. 25–35. [Google Scholar]
  6. Merkulov, V.I. Algorithm synthesis of the trajectory control of a mobile bistatic system of command radio control. Radiotekhnika 1998, 4, 99–101. [Google Scholar]
  7. Aguilar, C.O.; Krener, A.J. Numerical Solutions to the Bellman Equation of Optimal Control. J. Optim. Theory Appl. 2014, 160, 527–552. [Google Scholar] [CrossRef]
  8. Luo, Y.; Zhao, S.; Zhang, H. A New Robust Adaptive Neural Network Backstepping Control for Single Machine In finite Power System With TCSC. IEEE/CAA J. Autom. Sin. 2020, 7, 48–56. [Google Scholar]
  9. Kolesnikov, A.A.; Kolesnikov, A.A.; Kuz’menko, A.A. Backstepping and ADAR Method in the Problems of Synthesis of the Nonlinear Control Systems. Mekhatronika Avtom. Upr. 2016, 17, 435–445. (In Russian) [Google Scholar] [CrossRef] [Green Version]
  10. Nersesov, S.G.; Haddad, W.M. On the Stability and Control of Nonlinear Dynamical Systems via Vector Lyapunov Functions. IEEE Trans. Autom. Contr. 2006, 51, 203–215. [Google Scholar] [CrossRef]
  11. Åström, K.; Hägglund, T. PID Controllers: Theory, Design and Tuning; Instrument Society of America: Research Triangle Park, NC, USA, 1995. [Google Scholar]
  12. Yang, J.; Lu, W.; Liu, W. PID Controller Based on the Artificial Neural Network. In Advances in Neural Networks, Lecture Notes in Computer Science; Yin, F.L., Wang, J., Guo, C., Eds.; Springer: Berlin/Heidelberg, Germany, 2004; Volume 3174. [Google Scholar]
  13. Wang, L.-X. Stable Adaptive Fuzzy Control of Nonlinear Systems. IEEE Trans. Fuzzy Syst. 1993, 1, 146–155. [Google Scholar] [CrossRef]
  14. Freitas, A. Data Mining and Knowledge Discovery with Evolutionary Algorightms; Springer: Berlin/Heidelberg, Germany, 2002. [Google Scholar]
  15. Winkler, S. Evolutionary System Identification—Modern Concepts and Practical Applications. In Reihe C—Technik und Naturwissenschaften; Trauner: Linz, Austria, 2008. [Google Scholar]
  16. Alibekov, E.; Kubalık, J.; Babushka, R. Symbolic method for deriving policy in reinforcement learning. In Proceedings of the 55th IEEE Conference on Decision and Control (CDC), Las Vegas, NV, USA, 12–14 December 2016; pp. 2789–2795. [Google Scholar]
  17. Derner, E.; Kubalík, J.; Babushka, R. Reinforcement Learning with Symbolic Input–Output Models. In Proceedings of the 2018 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Madrid, Spain, 1–5 October 2018; pp. 3004–3009. [Google Scholar]
  18. Diveev, A.I.; Shmalko, E.Y. Automatic Synthesis of Control for Multi-Agent Systems with Dynamic Constraints. IFAC-PapersOnLine 2015, 48, 384–389. [Google Scholar] [CrossRef]
  19. Diveev, A.; Shmalko, E. Machine-made Synthesis of Stabilization System by Modified Cartesian Genetic Programming. IEEE Trans. Cybern. 2020. in print. [Google Scholar]
  20. Duriez, T.; Brunton, S.L.; Noack, B.R. Machine Learning Control—Taming Nonlinear Dynamics and Turbulence; Springer: Cham, Switzerland, 2017. [Google Scholar]
  21. Diveev, A.; Konstantinov, S.; Shmalko, E.; Dong, G. Machine Learning Control Based on Approximation of Optimal Trajectories. Mathematics 2021, 9, 265. [Google Scholar] [CrossRef]
  22. Upadhyay, D.; Schaal, C. Optimizing the driving trajectories for guided ultrasonic wave excitation using iterative learning control. Mech. Syst. Signal Process. 2020, 144, 106876. [Google Scholar] [CrossRef]
  23. Persio, L.D.; Garbelli, M. Deep Learning and Mean-Field Games: A Stochastic Optimal Control Perspective. Symmetry 2021, 13, 14. [Google Scholar] [CrossRef]
  24. He, W.; Huang, Y.; Fu, Z.; Lin, Y. ICONet: A Lightweight Network with Greater Environmental Adaptivity. Symmetry 2020, 12, 2119. [Google Scholar] [CrossRef]
  25. Sutton, R.S.; Barto, A.G. Reinforcement Learning: An Introduction; MIT Press: Cambridge, MA, USA, 1998. [Google Scholar]
  26. Martens, J. Deep learning via hessian-free optimization. In Proceedings of the 27th International Conference on Machine Learning, Haifa, Isreal, 21–24 June 2010. [Google Scholar]
  27. Witten, I.H.; Frank, E. Data Mining: Practical Machine Learning Tools and Techniques, 2nd ed.; Morgan Kaufman—Elsevier: Amsterdam, The Netherlands, 2005. [Google Scholar]
  28. Diveev, A.; Shmalko, E. Optimal Feedback Control through Numerical Synthesis of Stabilization System. In Proceedings of the 7th International Conference on Control, Decision and Information Technologies, CoDIT 2020, Prague, Czech Republic, 29 June–2 July 2020; pp. 112–117. [Google Scholar]
  29. Koza, J.R. Genetic Programming: On the Programming of Computers by Means of Natural Selection; MIT Press: Cambridge, MA, USA; London, UK, 1992. [Google Scholar]
  30. Yang, G.; Jeong, Y.; Min, K.; Lee, J.-W.; Lee, B. Applying Genetic Programming with Similar Bug Fix Information to Automatic Fault Repair. Symmetry 2018, 10, 92. [Google Scholar] [CrossRef] [Green Version]
  31. Dracopoulos, D.C.; Kent, S. Genetic programming for prediction and control. Neural Comput. Appl. 1997, 6, 214–228. [Google Scholar] [CrossRef] [Green Version]
  32. Miller, J.F. Cartesian Genetic Programming; Springer: Berlin/Heidelberg, Germany, 2011; 342p, ISBN 978-3-642-17310-3. [Google Scholar]
  33. Diveev, A.I.; Shmalko, E.Y. Self-adjusting control for multi robot team by the network operator method. In Proceedings of the 2015 European Control Conference, Linz, Austria, 15–17 July 2015; pp. 709–714. [Google Scholar]
  34. Diveev, A.I.; Shmalko, E.Y. Complete binary variational analytic programming for synthesis of control at dynamic constraints. In Proceedings of the ITM Web of Conferences, Moscow, Russia, 14–15 February 2017; Volume 10. [Google Scholar]
  35. Diveev, A. Small Variations of Basic Solution Method for Non-numerical Optimization. IFAC-PapersOnLine 2015, 48, 28–33. [Google Scholar] [CrossRef]
  36. Andrzej, O.; Stanislaw, K. Evolutionary Algorithms for Global Optimization. In Global Optimization. Nonconvex Optimization and Its Applications; Pintér, J.D., Ed.; Springer: Boston, MA, USA, 2006; Volume 85. [Google Scholar]
  37. Diveev, A.I.; Konstantinov, S.V. Study of the practical convergence of evolutionary algorithms for the optimal program control of a wheeled robot. J. Comput. Syst. Sci. Int. 2018, 57, 561–580. [Google Scholar] [CrossRef]
  38. Michalewicz, Z. Genetic Algorithms, Numerical Optimization, and Constraints. In Proceedings of the Sixth International Conference on Genetic Algorithms, Pittsburgh, PA, USA, 15–19 July 1995; pp. 151–158. [Google Scholar]
  39. Homaifar, A.; Qi, C.; Lai, S. Constrained optimization via genetic algorithms. Simulation 1994, 62, 242–254. [Google Scholar] [CrossRef]
  40. Kennedy, J.; Eberhart, R. Particle swarm optimization. In Proceedings of the IEEE International Conference on Neural Networks IV, Perth, Australia, 27 November–1 December 1995; pp. 1942–1948. [Google Scholar]
Figure 1. Projections of the given and found trajectories on the horizontal plane.
Figure 1. Projections of the given and found trajectories on the horizontal plane.
Symmetry 13 00336 g001
Figure 2. Optimal trajectory x 1 ( t ) (black) and optimal control x 1 * ( t ) (red).
Figure 2. Optimal trajectory x 1 ( t ) (black) and optimal control x 1 * ( t ) (red).
Symmetry 13 00336 g002
Figure 3. Optimal trajectory x 2 ( t ) (black) and optimal control x 2 * ( t ) (red).
Figure 3. Optimal trajectory x 2 ( t ) (black) and optimal control x 2 * ( t ) (red).
Symmetry 13 00336 g003
Figure 4. Optimal trajectory x 3 ( t ) (black) and optimal control x 3 * ( t ) (red).
Figure 4. Optimal trajectory x 3 ( t ) (black) and optimal control x 3 * ( t ) (red).
Symmetry 13 00336 g004
Figure 5. Projections of trajectories on the horizontal plane.
Figure 5. Projections of trajectories on the horizontal plane.
Symmetry 13 00336 g005
Figure 6. Optimal trajectory x 1 ( t ) (black) and optimal control x 1 * ( t ) (red) for movement on square.
Figure 6. Optimal trajectory x 1 ( t ) (black) and optimal control x 1 * ( t ) (red) for movement on square.
Symmetry 13 00336 g006
Figure 7. Optimal trajectory x 2 ( t ) (black) and optimal control x 2 * ( t ) (red) for movement on square.
Figure 7. Optimal trajectory x 2 ( t ) (black) and optimal control x 2 * ( t ) (red) for movement on square.
Symmetry 13 00336 g007
Figure 8. Optimal trajectory x 3 ( t ) (black) and optimal control x 3 * ( t ) (red) for movement on square.
Figure 8. Optimal trajectory x 3 ( t ) (black) and optimal control x 3 * ( t ) (red) for movement on square.
Symmetry 13 00336 g008
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Diveev, A.; Shmalko, E. Research of Trajectory Optimization Approaches in Synthesized Optimal Control. Symmetry 2021, 13, 336. https://doi.org/10.3390/sym13020336

AMA Style

Diveev A, Shmalko E. Research of Trajectory Optimization Approaches in Synthesized Optimal Control. Symmetry. 2021; 13(2):336. https://doi.org/10.3390/sym13020336

Chicago/Turabian Style

Diveev, Askhat, and Elizaveta Shmalko. 2021. "Research of Trajectory Optimization Approaches in Synthesized Optimal Control" Symmetry 13, no. 2: 336. https://doi.org/10.3390/sym13020336

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop