Abstract | The automatic construction of controllers would be ideal in situations where traditional control theory and algorithms fail, as it is the case with certain dynamical systems. Genetic programming, a field under the "umbrella" of evolutionary computation, is capable of creating computer programs given a high level description of a problem. The evolution of such computer programs is driven by their fitness. The fitness is defined by an objective function, which measures how well a particular program performs for the specific problem that tries to solve. Any controller can be described in terms of a computer program and thus, at least in theory, genetic programming offers an ideal candidate for the automatic construction of controllers. This paper considers the application of genetic programming on two different problems: the aircraft autolanding problem and the bioreactor control problem, both of which have been suggested in the literature as challenging benchmarks in the quest for building automatic controllers. The results presented here show that successful control laws in analytic form are derived for both cases. |
---|