Cybernetics for Systems Engineers – DRAFT F Cybernetics for Systems Engineers (DRAFT) 1 Contents 1 Introduction .................................................................................................................................. 12 2 Cybernetics: Basics ........................................................................................................................ 15 3 4 5 2.1 What is Cybernetics?............................................................................................................. 15 2.2 Information ........................................................................................................................... 15 2.3 Transformation ..................................................................................................................... 17 2.4 Algebraic Description ............................................................................................................ 19 2.5 The Determinate Machine with Input .................................................................................. 21 2.6 Coupling Systems .................................................................................................................. 22 2.7 Feedback ............................................................................................................................... 23 2.8 Immediate Effect and Reducibility ........................................................................................ 24 2.9 Stability ................................................................................................................................. 24 2.10 Equilibrium in Part and Whole .............................................................................................. 25 2.11 Isomorphism and Homomorphism ....................................................................................... 25 Cybernetics: Variety ...................................................................................................................... 27 3.1 Introduction .......................................................................................................................... 27 3.2 Constraint.............................................................................................................................. 27 3.3 Incessant Transmission ......................................................................................................... 31 3.4 Entropy .................................................................................................................................. 32 Cybernetics: Regulation ................................................................................................................ 35 4.1 Requisite Variety ................................................................................................................... 35 4.2 Error Controlled Regulation .................................................................................................. 37 Practical Finite State Machine Control ......................................................................................... 39 5.1 Introduction .......................................................................................................................... 39 5.2 The Relay and Logic Functions .............................................................................................. 39 5.2.1 Example : Capacitor Discharge Unit .............................................................................. 41 5.2.2 Reversing a DC Motor ................................................................................................... 44 5.2.3 Alarm Circuit ................................................................................................................. 45 5.3 6 1 Conclusions ........................................................................................................................... 46 Background Calculus ..................................................................................................................... 47 Cybernetics for Systems Engineers – DRAFT F 6.1 Basic Ideas ............................................................................................................................. 47 6.1.1 Introduction .................................................................................................................. 47 6.1.2 Areas Bounded by Straight Lines .................................................................................. 48 6.1.3 Coordinate Geometry ................................................................................................... 51 6.2 Areas ..................................................................................................................................... 52 6.2.1 Curved Boundary .......................................................................................................... 54 6.2.2 Mathematical Induction................................................................................................ 56 6.2.3 Higher Order Polynomials ............................................................................................. 57 6.2.4 The Circle....................................................................................................................... 59 6.2.5 Concluding Comment.................................................................................................... 61 6.3 Some Applications................................................................................................................. 62 6.3.1 Introduction .................................................................................................................. 62 6.3.2 Uniform Acceleration .................................................................................................... 62 6.3.3 Volumes of revolution................................................................................................... 65 6.4 Centres of Gravity ................................................................................................................. 67 6.5 Second Moment of Area ....................................................................................................... 69 6.5.1 6.6 Engineer’s Bending Theory ................................................................................................... 71 6.7 Ship Stability.......................................................................................................................... 74 6.8 Fractional Powers.................................................................................................................. 76 6.9 The Usual Starting Point........................................................................................................ 79 6.9.1 6.10 Introduction .................................................................................................................. 79 Applications........................................................................................................................... 81 6.10.1 Newton’s Method ......................................................................................................... 81 6.10.2 Maxima and Minima ..................................................................................................... 83 6.10.3 Body under Gravity ....................................................................................................... 84 6.10.4 Maximum Area.............................................................................................................. 85 6.10.5 Concluding Comments .................................................................................................. 85 6.11 Higher Order Derivatives ...................................................................................................... 86 6.11.1 Introduction .................................................................................................................. 86 6.11.2 Radius of Curvature Approximation ............................................................................. 86 6.12 More Exotic Functions .......................................................................................................... 88 6.12.2 6.13 2 Moment of inertia of a Disc .......................................................................................... 70 The Integral of the Reciprocal Function ........................................................................ 89 Searching for a Solution ........................................................................................................ 90 Cybernetics for Systems Engineers – DRAFT F 6.13.1 Finding the Magic Number. .......................................................................................... 92 6.13.2 Cooling of a Bowl of Soup ............................................................................................. 93 6.13.3 Capacitor Discharge ...................................................................................................... 94 6.13.4 Concluding Comments .................................................................................................. 95 6.14 Onwards and Upwards.......................................................................................................... 95 6.15 Trigonometrical Functions .................................................................................................... 96 6.15.1 Introduction .................................................................................................................. 96 6.15.2 Angles ............................................................................................................................ 96 6.15.3 The Full Set of Trigonometric Functions ....................................................................... 98 6.15.4 Relationhips Between Trigonometric Functions ......................................................... 100 6.15.5 Special Cases ............................................................................................................... 101 6.15.6 Functions of Sums and Differences of Angles ............................................................. 102 6.15.7 Derivatives of Trigonometric Functions ...................................................................... 103 6.16 Binomial Theorem ............................................................................................................... 105 6.17 A Few Tricks ........................................................................................................................ 106 6.18 Differential Equations ......................................................................................................... 107 6.18.1 Introduction ................................................................................................................ 107 6.18.2 First Order Equation .................................................................................................... 107 6.18.3 Integration by Parts..................................................................................................... 112 6.18.4 More Exotic Forms ...................................................................................................... 115 6.19 The Lanchester Equation of Combat................................................................................... 117 6.20 Impulse Response ............................................................................................................... 120 6.20.1 Second, and Higher Order ........................................................................................... 122 6.20.2 Resonant Circuit .......................................................................................................... 125 6.20.3 Pin Jointed Strut .......................................................................................................... 125 6.20.4 Mass – Spring – Damper ............................................................................................. 127 6.20.5 Higher Order Systems ................................................................................................. 130 6.21 7 Conclusions ......................................................................................................................... 130 More Advanced Calculus............................................................................................................. 132 7.1 Introduction ........................................................................................................................ 132 7.2 Arc Length and Surface areas of Revolution ....................................................................... 132 7.2.1 7.3 The Catenary ....................................................................................................................... 134 7.3.1 3 Why Bother? ............................................................................................................... 134 Comment..................................................................................................................... 138 Cybernetics for Systems Engineers – DRAFT F 7.4 About Suspension Bridges .................................................................................................. 139 7.5 Arc Length of an Ellipse ....................................................................................................... 141 7.5.1 Towards a better solution ........................................................................................... 144 7.5.2 About Errors ................................................................................................................ 147 7.6 Convergence ....................................................................................................................... 147 7.7 Concluding Comment.......................................................................................................... 149 8 The Calculus of Variations ........................................................................................................... 151 9 8.1.1 Application to Maximum Area .................................................................................... 152 8.1.2 Comment..................................................................................................................... 155 Cybernetics and Automatic Control ............................................................................................ 157 9.1 Introduction ........................................................................................................................ 157 9.2 Implications for System Specification ................................................................................. 157 9.3 On Feedback ....................................................................................................................... 158 9.4 About Intelligence ............................................................................................................... 159 9.5 Automatic Control............................................................................................................... 161 10 Stability ................................................................................................................................... 162 10.1 Introduction ........................................................................................................................ 162 10.2 Additional Background ........................................................................................................ 162 10.3 Matrix Algebra .................................................................................................................... 164 10.3.1 Definition and Basic Operations ................................................................................. 164 10.3.2 Additional Operations ................................................................................................. 165 10.3.3 The Identity Matrix and the Matrix Inverse ................................................................ 166 10.3.4 Checking for Linear Independence - Determinants .................................................... 167 10.4 10.4.1 ‘Sign’ of a Matrix ......................................................................................................... 170 10.4.2 Eigenvalues and Eigenvectors ..................................................................................... 171 10.4.3 Calculation of Eigenvalues .......................................................................................... 173 10.5 11 Stability – a Different Perspective....................................................................................... 174 Examples ................................................................................................................................. 177 11.1 Introduction ........................................................................................................................ 177 11.2 Vibrating String ................................................................................................................... 177 11.3 Directional Stability of a Road Vehicle ................................................................................ 179 11.4 Hunting of a Railway Wheelset ........................................................................................... 182 12 4 State Space Representation ................................................................................................ 169 Linearisation ............................................................................................................................ 187 Cybernetics for Systems Engineers – DRAFT F 12.1 Introduction ........................................................................................................................ 187 12.2 Stability of a Circular Orbit .................................................................................................. 187 12.3 Rigid Body Motion............................................................................................................... 188 12.3.1 Nomenclature ............................................................................................................. 189 12.3.2 Translation Equation ................................................................................................... 190 12.3.3 Rotation Equations...................................................................................................... 192 12.4 12.4.1 Cartesian Rocket – Steady Flight ................................................................................. 194 12.4.2 Cartesian Rocket – High Angle of Attack..................................................................... 195 12.4.3 Spinning Projectile ...................................................................................................... 196 12.4.4 Spacecraft ................................................................................................................... 197 12.4.5 Aircraft ........................................................................................................................ 198 12.4.6 Wider Application ....................................................................................................... 198 12.5 Partial Derivatives ............................................................................................................... 199 12.5.1 The Symmetrical Rocket Re-visited............................................................................. 200 12.5.2 Stability of a Spinning Shell ......................................................................................... 201 12.6 A Few More Tricks............................................................................................................... 202 12.6.1 Spinning Shell Conditions for Stability ........................................................................ 205 12.6.2 The Aeroplane ............................................................................................................. 206 12.6.3 The Brennan Monorail ................................................................................................ 209 12.6.4 Equation of Motion ..................................................................................................... 210 12.6.5 Approximate Factorisation.......................................................................................... 211 12.6.6 Brennan’s Solution ...................................................................................................... 211 12.6.7 Cornering..................................................................................................................... 212 12.6.8 Side Loads ................................................................................................................... 213 12.7 13 Concluding Comments ........................................................................................................ 213 Modifying the System Behaviour ............................................................................................ 214 13.1 Introduction ........................................................................................................................ 214 13.2 The Laplace Transform ........................................................................................................ 215 13.3 Transforms of Common Functions...................................................................................... 216 13.3.1 Derivatives and Integrals............................................................................................. 216 13.3.2 Exponential Forms ...................................................................................................... 217 13.3.3 Polynomials ................................................................................................................. 218 13.4 5 Perturbation forms ............................................................................................................. 194 Impluse Function and the Z Transform .............................................................................. 218 Cybernetics for Systems Engineers – DRAFT F 13.5 14 Laplace Transform Solution of Non-Homogenous Equations ............................................. 219 Impulse Response Methods – Evan’s Root Locus ................................................................... 222 14.1 Introduction ........................................................................................................................ 222 14.2 Example: Missile Autopilot ................................................................................................. 223 14.3 General Transfer Function .................................................................................................. 226 14.4 Rules for Constructing the Root Locus ................................................................................ 227 14.4.1 Real Axis ...................................................................................................................... 227 14.4.2 Asymptotes ................................................................................................................. 228 14.4.3 Exit and Arrival Directions ........................................................................................... 229 14.4.4 Break Away Points....................................................................................................... 229 14.4.5 Value of Gain ............................................................................................................... 230 14.4.6 Comment on Rules ...................................................................................................... 230 14.5 Autopilot Revisited.............................................................................................................. 231 14.5.1 Numerical Example ..................................................................................................... 231 14.5.2 Initial Selection of Servo Pole ..................................................................................... 233 14.5.3 Adding Compensation ................................................................................................. 235 14.5.4 Concluding Comments ................................................................................................ 237 14.6 Gyro Monorail ..................................................................................................................... 238 14.6.1 Representative Parameters ........................................................................................ 238 14.6.2 Root Locus ................................................................................................................... 239 14.7 Comments on Root Locus Examples ................................................................................... 242 14.8 The Line of Sight Tracker ..................................................................................................... 242 14.9 Tracking and the Final Value Theorem ............................................................................... 244 14.10 The Butterworth Pole Pattern............................................................................................. 245 14.11 Status of Root Locus Methods ............................................................................................ 246 15 6 Representation of Compensators ........................................................................................... 247 15.1 Introduction ........................................................................................................................ 247 15.2 Modelling of Continuous Systems ...................................................................................... 247 15.3 Sampled Data Systems ........................................................................................................ 250 15.4 Zero Order Hold .................................................................................................................. 251 15.5 Sampled Data Stability ........................................................................................................ 251 15.6 Illustrative Example............................................................................................................. 252 15.7 Sampled Data Design .......................................................................................................... 256 15.8 Concluding Comments ........................................................................................................ 257 Cybernetics for Systems Engineers – DRAFT F 16 Frequency Domain Description............................................................................................... 258 16.1 Introduction ........................................................................................................................ 258 16.2 Sinusoidal Response............................................................................................................ 258 16.3 Stability ............................................................................................................................... 260 16.4 Stability Margins ................................................................................................................. 261 16.5 Gain and Phase Margin ....................................................................................................... 263 16.6 Closed Loop Response from the Harmonic Locus............................................................... 263 16.7 The Monorail Example ........................................................................................................ 265 16.8 Sampled Data Systems ........................................................................................................ 268 16.9 Common z transforms......................................................................................................... 269 16.9.1 Sine and Cosine ........................................................................................................... 269 16.10 Frequency Warping ............................................................................................................. 271 16.11 Decibels ............................................................................................................................... 271 16.12 Signal Shaping in the Frequency Domain – Bode Plots....................................................... 272 16.12.1 Second Order Transfer Functions and Resonance .................................................. 274 16.13 Gain and Phase Margins from Bode Plots........................................................................... 276 16.14 Phase Unwrapping .............................................................................................................. 277 16.15 The Nichol’s Chart ............................................................................................................... 277 16.16 A Word of Warning ............................................................................................................. 278 16.17 Concluding Comment.......................................................................................................... 278 17 Time Varying Systems ............................................................................................................. 280 17.1 Introduction ........................................................................................................................ 280 17.2 Missile Homing .................................................................................................................... 281 17.3 Significance of Sight Line Spin ............................................................................................. 283 17.4 Pursuit Solution ................................................................................................................... 283 17.5 Seeking a More Satisfactory Navigation Law ...................................................................... 285 17.5.1 17.6 Blind Range ......................................................................................................................... 287 17.7 Sensitivity ............................................................................................................................ 289 17.7.1 Block Diagram Adjoint................................................................................................. 291 17.8 Example: Homing Loop with First Order Lag ...................................................................... 292 17.9 Concluding Comment.......................................................................................................... 301 18 18.1 7 Lateral Acceleration History ........................................................................................ 286 Rapid Non-Linearities .............................................................................................................. 302 Introduction ........................................................................................................................ 302 Cybernetics for Systems Engineers – DRAFT F 18.2 Describing the Distorted Signal ........................................................................................... 303 18.3 Fourier Series Curve Fit ....................................................................................................... 303 18.4 The Describing Function...................................................................................................... 305 18.5 Application of Describing Function to the Harmonic Locus ................................................ 307 18.6 Bang-Bang Control .............................................................................................................. 307 18.7 Schilovski Monorail ............................................................................................................. 309 18.8 More General Servo Considerations ................................................................................... 310 18.9 The Phase Plane .................................................................................................................. 312 18.9.1 Dominant Modes ........................................................................................................ 312 18.9.2 Representation of a Second Order Plant .................................................................... 313 18.9.3 Unforced Plant ............................................................................................................ 314 18.10 Introducing Control ............................................................................................................. 316 18.11 Hierarchical System Considerations.................................................................................... 319 18.12 Scanning System ................................................................................................................. 320 18.12.1 Outer Loop .............................................................................................................. 323 18.13 Concluding Comments ........................................................................................................ 325 19 The Representation of Noise .................................................................................................. 326 19.1 Introduction ........................................................................................................................ 326 19.2 Correlation .......................................................................................................................... 326 19.2.1 The Same, Or Different? ............................................................................................. 326 19.2.2 Pattern Matching ........................................................................................................ 328 19.3 Auto Correlation ................................................................................................................. 331 19.4 Characterising Random Processes ...................................................................................... 333 19.4.1 White Noise................................................................................................................. 334 19.4.2 System Response to White Noise ............................................................................... 335 19.4.3 Implications for Time-Varying Systems....................................................................... 337 19.5 19.5.1 The Fourier Transform ................................................................................................ 338 19.5.2 Application to Noise .................................................................................................... 339 19.6 20 8 Frequency Domain .............................................................................................................. 338 Application to a Tracking System ........................................................................................ 341 Line of Sight Missile Guidance ................................................................................................ 343 20.1 Introduction ........................................................................................................................ 343 20.2 Line of Sight Variants .......................................................................................................... 343 20.3 Analysis ............................................................................................................................... 344 Cybernetics for Systems Engineers – DRAFT F 20.4 21 Multiple Outputs ..................................................................................................................... 347 21.1 Introduction ........................................................................................................................ 347 21.2 Nested Loops....................................................................................................................... 347 21.3 Pole Placement ................................................................................................................... 350 21.4 Algebraic Controllability...................................................................................................... 351 21.4.1 Monorail ...................................................................................................................... 353 21.4.2 Road Vehicle................................................................................................................ 353 21.5 Comment on Pole-placement ............................................................................................. 355 21.6 The Luenberger Observer ................................................................................................... 357 21.7 The Separation Theorem .................................................................................................... 358 21.8 Reduced Order Observers ................................................................................................... 359 21.9 Concluding Comments ........................................................................................................ 360 22 Respecting System Limits – Optimal Control .......................................................................... 362 22.1 Introduction ........................................................................................................................ 362 22.2 Optimal Control................................................................................................................... 362 22.3 Optimisation Background ................................................................................................... 364 22.3.1 Principle of Optimality ................................................................................................ 364 22.3.2 Application to the Control Problem ............................................................................ 366 22.4 Gyro Monorail ..................................................................................................................... 368 22.5 The Optimal Observer ......................................................................................................... 370 22.5.1 Covariance Equation ................................................................................................... 371 22.5.2 The Observer ............................................................................................................... 373 22.5.3 Tracker Example .......................................................................................................... 374 22.5.4 Alpha beta Filter .......................................................................................................... 375 22.6 23 9 Comments ........................................................................................................................... 346 Comment ............................................................................................................................ 375 Multiple Inputs ........................................................................................................................ 377 23.1 Introduction ........................................................................................................................ 377 23.2 Eigenstructure Assignment ................................................................................................. 378 23.3 Using Full State Feedback ................................................................................................... 378 23.3.1 Basic Method .............................................................................................................. 378 23.3.2 Fitting the Eigenvectors .............................................................................................. 379 23.3.3 Calculation of Gains .................................................................................................... 381 23.3.4 Comment on Stability ................................................................................................. 382 Cybernetics for Systems Engineers – DRAFT F 23.4 Loop Methods ..................................................................................................................... 383 23.4.1 General Considerations............................................................................................... 383 23.4.2 Nyquist Criterion ......................................................................................................... 384 23.5 Sequential Loop Closing ...................................................................................................... 385 23.6 Diagonalisation Methods .................................................................................................... 386 23.6.1 ‘Obvious’ Approach ..................................................................................................... 386 23.6.2 Commutative Compensators ...................................................................................... 386 23.7 24 Comment ............................................................................................................................ 386 Closed Loop Methods ............................................................................................................. 388 24.1 Introduction ........................................................................................................................ 388 24.2 The H-infinity Norm ............................................................................................................ 389 24.3 MIMO Signals ...................................................................................................................... 390 24.4 Closed Loop Characterisation ............................................................................................. 392 24.4.1 Tracker Example .......................................................................................................... 393 24.4.2 Comment..................................................................................................................... 393 24.4.3 Input Output Relationships ......................................................................................... 394 24.4.4 Weighted Sensitivity ................................................................................................... 395 24.5 Stacked Requirements ........................................................................................................ 396 24.5.1 Robustness .................................................................................................................. 397 24.5.2 Phase in MIMO Systems ............................................................................................. 397 24.6 25 Concluding Comment.......................................................................................................... 398 Catastrophe and Chaos ........................................................................................................... 400 25.1 Introduction ........................................................................................................................ 400 25.2 Catastrophe......................................................................................................................... 400 25.3 The ‘Fold’............................................................................................................................. 401 25.4 The ‘Cusp’ ............................................................................................................................ 402 25.5 Comment ............................................................................................................................ 403 25.6 Chaos ................................................................................................................................... 404 25.6.1 Adding Some Dynamics............................................................................................... 404 25.6.2 Attractors .................................................................................................................... 406 25.7 The Lorenz Equation ........................................................................................................... 407 25.7.1 Explicit Solution ........................................................................................................... 409 25.8 The Logistics Equation......................................................................................................... 410 25.9 Bluffer’s Guide .................................................................................................................... 412 10 Cybernetics for Systems Engineers – DRAFT F 25.9.1 Basin of Attraction ...................................................................................................... 412 25.9.2 Bifurcation................................................................................................................... 412 25.9.3 Cobweb Plot ................................................................................................................ 412 25.9.4 Feigenbaum Number .................................................................................................. 412 25.9.5 Liapunov Exponent...................................................................................................... 412 25.9.6 Poincaré Section.......................................................................................................... 413 25.9.7 Symmetry Breaking ..................................................................................................... 413 25.10 Concluding Comment.......................................................................................................... 413 26 Artificial Intelligence and Cybernetics .................................................................................... 414 26.1 Introduction ........................................................................................................................ 414 26.2 Intelligent Knowledge Based Systems ................................................................................ 414 26.3 Fuzzy Logic .......................................................................................................................... 415 26.4 Neural Nets ......................................................................................................................... 417 26.5 Concluding Comment.......................................................................................................... 418 27 Conclusions ............................................................................................................................. 419 27.1 Introduction ........................................................................................................................ 419 27.2 Hierarchy ............................................................................................................................. 419 27.3 Intelligence .......................................................................................................................... 420 27.4 Seeking Null......................................................................................................................... 420 27.5 And Finally... ........................................................................................................................ 421 11 Cybernetics for Systems Engineers – DRAFT F 1 Introduction It is only fair to the reader to start with sufficient information to decide whether to proceed. The word ‘system’ as used today is most frequently meant in the sense implied by ‘systems analyst’, a specialist concerned with mapping technically naive user’s requirements into computer code. Before the pre-eminence of software engineering, it had another meaning, as defined by the likes of Norbert Wiener, William Ross Ashby, Harry Nyquist, and Claude Shannon to describe a subject which was intended for the highly numerate, in order to gain a scientific understanding of system behaviour. Without passing further comment on the works which expound the modern innumerate approach, I will merely comment that this is not such a work. In order to discriminate between the two usages, I shall refer to the former as ‘systems engineering’ and the latter I shall call ‘cybernetics’ from the Greek word for ‘helmsman’. To supplement the management/organizational characterization of the system, we require a means of characterizing components of system elements which do not exist, but are postulated as potential means of achieving improved system performance. This requires the methods of cybernetics. The methods which work, and have found practical application over the years belong to a body of knowledge called ‘control theory’. In recent years, for various reasons, this has become perceived as an arcane specialisation, which is unfortunate because it constitutes a universal means of system description and analysis, which should form part of every engineer’s toolkit, regardless of specialisation. The book is unusual in appealing to the broad knowledge and experience of the practitioner, rather than trying to present the subject as self-contained. Like all systems approaches, it is useless within itself; it’s value becomes apparent in its application to real problems. With that in mind, the text presents a number of practical problems to illustrate the methods used. The approach is eclectic, taking examples from all branches of engineering, to illustrate the universal application of cybernetics to all self-correcting systems, regardless of their actual physical implementation. As far as possible, the concepts are explained in plain English, so hopefully a reasonable understanding of the subject matter may be gained from the text alone. The early chapters present the fundamental ideas more as a philosophical discussion, than pragmatically useful methods. However, the reader is advised to get to grips with the Principle of Requisite Entropy, on which all valid systems thinking, mathematical or not, is based, yet judging by the decrees of some so-called ‘experts’, is not as widely understood as it should be. This is followed by a brief presentation on the practical implementation of finite state controllers, leading to the conclusion that a problem formulation which enables hardware behaviour to be taken into account is to be preferred to general software development methods. This theme of the 12 Cybernetics for Systems Engineers – DRAFT F constraint imposed by the real world forms the basis of the remainder of the book, but requires significantly greater mathematical knowledge. We then proceed with the background mathematics needed to understand the remainder of the book. This is no more than would be covered in the first year of a typical engineering course. It is included because many line engineers may not have had call to use their mathematics in support of their largely administrative and managerial duties, and a reminder is in order. The mathematically adept may safely skip these chapters. The presentation differs from the more usual approach in that we start with integral calculus, which is immediately applicable to engineering problems, and differentiation is introduced as its inverse. This is expected to appeal to the natural engineer, who has little inclination to discuss the number of angels which can dance on a pinhead. The main body of the book then considers the natural behaviour of dynamic systems, covering the methods used to uncover the causes of instability, and the identification of critical parameters. Many systems issues are resolved cheaply and simply by changes to the plant parameters, without introducing control at all. We emphasise the need to understand the dynamic behaviour of the plant before considering introducing artificial means of improving its behaviour. All the clues we need to design our closed loop system arise from this understanding of the basic plant dynamics. After the basic stability has been considered, we cover the body of knowledge which has become known as ‘classical control’. These methods relate the open loop plant behaviour to that of the closed loop, so that the effect of compensators introduced within the loop may be predicted. Central to this presentation are the Laplace and Z transforms, which enable the behaviour of complex networks of dynamic elements to be predicted readily. These methods are concerned dominantly with stability, although tracking behaviour is also considered. It becomes evident that stability alone is insufficient, we need to introduce constraints of noise and saturation to decide the characteristics of the compensation we may employ. To this end we investigate the effects of non-linearities on stability using describing functions and phaseplane methods. Before proceeding to characterise noise, we take a detour into the world of terminal controllers, which naturally introduces the method of adjoint simulation. The ideas introduced here are needed to understand the effects of noise impulses applied to a linear system using a time domain description. The concept of white noise is introduced, and from it the means of predicting the accuracy of tracking from the noise sources within the loop. All the ‘classical’ approaches are intended for single input, single output systems. With the reduced cost of sensors in recent years, most closed loop systems have multiple outputs, and we revert back to the state-space descriptions used to describe our basic plant dynamics. This introduces pole placement and the idea of an observer, as a special type of compensator which models the plant dynamics. The separation theorem is invoked to show that the dynamics of the controller do not affect those of the observer, so the two can be designed in isolation. These methods work directly with the closed loop plant. The state space methods are extended to consider controllers whose states are limited in range (typically by saturation), and observers whose accuracy is limited by noise (the fixed gain Kalman 13 Cybernetics for Systems Engineers – DRAFT F filter). The fundamental limitation of these approaches, that the separation theorem is unlikely to apply in practice, is identified. Moving on from muti-output systems, we proceed to consider systems which are multi-input as well as multi-output. In view of the cost of servos, such systems are unlikely to be common. Usually we deal with multiple de-coupled single-input/single-output systems. However, these are recognised as potentially difficult to design, and considerable intellectual effort has been invested in them over the past thirty years. Most of the recent work lies beyond the scope of the current book, which is not intended for post graduate studies into control. Some techniques which the author has found practically useful are presented. However, the most important question to be resolved before embarking on a MIMO controller; is why is the plant so horribly cross-coupled in the first place? The dominant method, that of H∞ ‘robust’ control, or the ideas behind it, are presented briefly, but the absence of examples which were not contrived for the sole purpose of illustrating the method, is rather a handicap when presenting it to an audience of pragmatic engineers. The remainder of the book reverts back to the philosophical discussion of the opening chapters, except that the exposition is more speculative. These sections may be skipped unless the ideas presented are of particular interest to the reader. Well, that completes the introduction, it should be evident whether or not the book is for you. If not, I thank you for your time. If so, do I hear the satisfying ring of a cash register? 14 Cybernetics for Systems Engineers – DRAFT F 2 Cybernetics: Basics 2.1 What is Cybernetics? Cybernetics is a theory of machines, but does not deal with the gears or circuit boards of physical machines, but in how machines behave independently of their physical implementation. In essence, it is the study of machines in the abstract. As it deals with behaviour, rather than implementation, the ‘what’ of the system function, rather than the ‘how’, it would be reasonable to use it as the basis for system specification, particularly as the resulting definitions of behaviour are concise, compact and unambiguous. The approach deals with ‘black boxes’, that is we deal with how entities behave, or are required to behave, not how they are implemented in any specific instance. We find that even the most complex behaviours can be characterized with extremely simple system models. Indeed, it is shown that much of the complexity of ‘complex’ systems serves to simplify behaviour. This is in contrast to the ‘white box’ approach characteristic of more mainstream systems engineering, where internal structure and component functions must be known, in order for the system to be specified. It follows that the methods of cybernetics, in containing no pre-conceptions as to how behaviours are achieved ,furnishes the techniques most appropriate for system specification, yet paradoxically, their strong emphasis on mathematical formulation, appears to have relegated them to the ‘implementation’ stages of system development, in the minds of many who call themselves ‘systems engineers’. 2.2 Information In the author’s opinion, the best introduction to cybernetics was written in the mid-Twentieth Century by William Ross Ashby, and is entitled, not surprisingly; ‘An Introduction to Cybernetics’. This book has been out of print for many years, but it is well worth seeking it out. What follows summarises his ideas, but with nowhere near the same quality of prose. For consistency with modern systems engineering practice I have attempted to introduce the subject using UML (Unified Modelling Language) diagrams, as these are becoming the de-facto standard, as a means of capturing basic system function. Unfortunately, it soon came apparent that cybernetic ideas cannot be expressed using them, as they are too application-oriented for the level of abstraction required. We are told that with the internet, we have access to more information than has ever been possible in the past. Well, it all depends on how you define ‘information’. We have the basis of information when we can change the state of the environment, for example tying a knot in a handkerchief to remind us to do something. Whenever we alter our surroundings, we may ascribe meaning to the changes we introduce. We code our abstract thoughts as configurations of matter in the physical world. The action of forces of nature on the configuration of matter may introduce changes, so that it no longer conveys the same idea, e.g. an animal may accidentally trample the stack of stones. 15 Cybernetics for Systems Engineers – DRAFT F No matter how we encode it any physical medium can become corrupted, there will always be noise present. We usually distinguish a valid configuration from the corrupted ones by having a limited number of valid configurations, which have a low probability of occurring by chance. The commonest form of encoding is through speech. Of all the potential sounds and sequences of sounds the voice is capable of making, only a tiny part of the potential repertoire is used to convey messages. Claude Shannon proved that there exists a coding scheme which will convey an error-free message regardless of the noise present in the communication channel. It is a reasonable guess that spoken language isn’t too far off this ideal encoding for the noise sources in existence when language first emerged. Whether it remains ideal in the industrial age is a moot point. The simplest form of encoding is to ensure the signal level is higher than the noise level. Ultimately, the noise limits the difference between adjacent signal amplitudes, introducing a quantisation error, even when the signal is not digital in nature. More typically, a set of symbols would be defined as a sequence of signal levels in time (e.g. Morse code symbols), or in space (printed characters). We know that the set of permissible sequences is limited for sequences conveying information. So a sequence of letters taken from the Latin alphabet in which the letter X is as common as the letter E, is not likely to convey English language messages. Similarly we could check to see whether the combinations of letters formed English words. We could repeat this process by checking the frequency of words, but ultimately all we are doing is checking that the message is valid English. We are not able, through this process, to decide whether the message is true, or indeed conveys any meaning. It is in the information-theoretic sense described above, that we claim there is an information explosion. The reality is, searching for useful information is like picking up a weak signal on a radio receiver – turning up the volume just increases the noise. The internet has given us a spectacular increase in uninformed opinion, misinformation and utter nonsense, with valid information becoming ever more elusive. The needle is the same size as ever, but the haystack has increased by orders of magnitude. Information theory infers that information is being conveyed if the statistical distribution of the symbols in a message is different from a uniform distribution in which all symbols are equally likely. The distribution implies the conveyance of information, without specifying its exact form. For this reason the information is characterised as a function of the probabilities of each symbol occurring, called the entropy of the message. The entropy is defined in such a way that if a message is passed through a process the output entropy is equal to the sum of the process entropy and that of the original message. As an example, in order to improve the match between a message and a communications channel (which must have capacity to transmit every symbol with equal probability), the message may pass through some form of coder. At the other end, a decoder, matched to the coder, so by implication has same entropy, re-creates the original message. This matching of entropies illustrates a fundamental principle of cybernetics, called the principle of requisite entropy. 16 Cybernetics for Systems Engineers – DRAFT F A system which remains in the same state indefinitely is not particularly interesting, so cybernetics concerns itself with the evolution of system states with respect to time, or some other independent variable. The most important results can be derived using what is called a finite state machine formulation. The machine has a number of states which we can give English language labels, or symbols (e.g; letters, pictograms, or numbers ). These are precisely the same as a set of symbols we might use to code information. We assume that the transitions between states take place instantaneously. We describe our system as a set of symbols, and we can, over a long period of time determine the probability of the machine being in each of its states at any particular time. Thus we can characterize the machine’s evolution as an entropy, in much the same way as we can define the entropy of a messages made up of symbols taken from a set. However, the statistical behaviour of a finite state machine introduces excessive complication, which will obscure the concepts I am trying to introduce. The most important ideas can be understood, assuming the machine to be deterministic, rather than statistical (stochastic) in nature. Admittedly, systems which actually transition neatly between states in this manner are rarely, if ever, to be found in nature. Our primary system consideration is whether the states can in fact be reached and maintained. With a digital computer this issue is hidden in the design of the circuits, in more general systems, we cannot be certain that this will be the case. 2.3 Transformation Imagine yourself in an environmentally-controlled greenhouse, and you notice the following: The temperature falls The carbon dioxide concentration reduces The humidity falls You conclude that somebody has opened the door. 17 Cybernetics for Systems Engineers – DRAFT F The UML state chart for this behaviour shows ‘Open door’ as the condition which causes the transitions from the states before the door was opened to some (unspecified) time later. The implication is that the disturbance caused by leaving the door open exceeds the capability of the environmental control system to maintain regulation, so that a new (unsatisfactory) equilibrium condition results. Alternatively we may consider ‘Open door’ to be an operator, and we may denote it with an arrow . We take one of the effects, say the first, and write it as a compact expression: AB Where A is shorthand for ‘warm’ is the ‘operand’, the arrow implies an influence, B is shorthand for ‘cool’, is what Ross Ashby calls the ‘transition‘. The total set of transitions consists of a vector of operands which become a vector of transitions. The set of transitions is called a ‘transformation‘. The UML state chart could represent this by defining a pair of compound states which include the individual states: The UML description necessarily contains labels associating the diagrams with specific entities, which is fine for describing specific systems, but of little help in seeking universal principles which apply to all systems. For this reason, the transition arrow label will be omitted from the state diagrams and states will be labelled by letters, without reference to any specific system. This example is an ‘open’ transformation, because the transitions and operands are from different sets, if the transform has the same set of symbols in the operand and transition sets, it is called a closed transformation. All the above transforms are single valued. Other types are possible, for example: AB or D BA CB or C DD Is not single valued. In principle, the system may start and end on any state, so we have omitted the start and end node from the state chart. The alternative is to present 16 practically identical diagrams, which seems an inefficient means of presenting a relatively simple idea. 18 Cybernetics for Systems Engineers – DRAFT F Another important type of transformation is ‘one-to-one‘. This is single valued, so that each operand yields a unique transition, but in addition, each transition indicates a unique operand. A transformation which is single valued, but not one-to-one is called ‘many to one’. To complete the set of possible transforms, we have the identity which transforms the operand to itself. The transitions in UML state diagrams usually correspond to a defined condition, rather than the natural evolution of the system over a single fixed time interval. For this reason they do not represent an efficient means of expressing system behaviour, beyond the execution of computer code. Activity diagrams, sequence diagrams and collaboration charts, likewise are doubtlessly invaluable for defining how code should be organised and executed, but outside the realm of software engineering UML appears somewhat over-sold. 2.4 Algebraic Description 19 Cybernetics for Systems Engineers – DRAFT F The arguments of this and subsequent sections, are more compactly illustrated by using a matrix representation of the transitions. The first row is the start state, whilst the second row is the final state. The presence of transformations which are not single valued raises the difficulty that more than one matrix is needed. When this occurs, it is evident that more information is needed to determine the system evolution. For one to one, or many to one transitions, we have sufficient information in our transform definition to determine how the system behaves when the transform is repeated. We refer to Tn as the nth power of T, i.e. the result of applying T n times. More generally, we have the product, or composition of two different transforms operating on the same set of operands: Consider now a more complicated transformation: 20 Cybernetics for Systems Engineers – DRAFT F By starting at any state and following the arrows, the repeated transformation ends up cycling around a set of states indefinitely. If any state transitions back to itself, the transformation will stop once it is reached. The separate regions of the kinematic graph are known as ‘basins‘. By implication, the system either converges on to its final state or cycles around a set of states in a basin. Unfortunately, this benign behaviour is only a consequence of the finite state machine assumption. In practice it is very likely that there will be states which the machine can reach which are not in the finite state machine model of the system. The belief that a finite state machine will be adequate outside of a controlled environment is a dangerous delusion. As an example; consider a strategy system which has a gambit for every anticipated move the opponent may make. There is nothing to stop the opponent from making a move which was not anticipated, and for which no counter was anticipated. The implicit assumption of the finite state machine, that the set of disturbances to which it is subjected is closed, limits its operation to extremely benign controlled environments. A system which is capable of surviving a real environment, with its open set of disturbances, presents a more challenging problem. 2.5 The Determinate Machine with Input A determinate machine is any process which behaves as a closed single-valued transformation. By ‘single value’ we do not necessarily imply a single number, but an operand in a transformation. Each operand in the previous section was a label covering sets of attributes; a particular combination of data and/or behaviour which we called a ‘state‘. Since each state is assigned a unique label, the ‘single-valued ‘requirement implies that the machine’s transition from one state to another is strictly one-to-one. We have defined our deterministic cybernetic machine by the transforms which characterise its behaviour. The set of operands are states, which themselves may be transforms. Thus we can define a transform U whose operands are transforms T1, T2 and T3. Applying U has the effect, for example: T1T2 T2T2 T3T1 In addition to our requirement for single valued transformations, this discussion requires all transformations to be closed. The three transformations in U could have distinct sets of operands, in which case, U is merely a ‘big’ transform, no different in nature from its component transforms. The more interesting case is when all three component transforms have the same set of operands. For example, we assume all three component have a common set of operands; A,B,C,D: 21 Cybernetics for Systems Engineers – DRAFT F The three transformations may be considered three ways in which the same machine can be made to behave. That it is the same machine is implicit in the common operands (states). If we refer to ‘change’ in the context of such a machine, it is change in behaviour which is of interest, rather than the transformations associated with each individual behaviour. Those with programming experience will recognise the idea of a subroutine call in assembler, or a function /procedure call in higher level languages. The subscript 1,2 or 3 applied to the individual transformations indicates which behaviour is selected at any time, and will be called a ’parameter’. A real machine, which can be described as a set of closed single-valued transformations will be called a machine with input. The parameter specifying which behaviour is selected is called the ‘input’. 2.6 Coupling Systems Suppose a machine P is to be connected to another, R. This is possible if the parameters of R are functions of the states of P. We define the two machines by their sets of transformations. R will have three, P only one. 22 Cybernetics for Systems Engineers – DRAFT F We construct a relation, which is a single valued transformation relating the state of P to the parameter values of R. We assume also that the two machines transforms are synchronised to a common timescale. An example of the coupling relationship could be: State of P i j k Value of parameter 2 3 2 The two machines now form a new machine of completely determined behaviour. The set of states is equal to the number of combinations of the 3 states of P and 4 states of R, and each individual state is a vector quantity, e.g. (a,i). Taking these compound states, a kinematic graph may be drawn which has behaviour which is different from both P and R. If machine P had more than one transform, it could accept input from R, so that state of each machine would determine the transform of the other. Again, the result is a new machine with its 12 states, and completely new behaviour. Its behaviour is determined by considering its time evolution from a start state, rather than directly from the coupling scheme. The coupling is achieved by specifying what function the input parameters of one machine are to be of the other machines states. Evidently, the parts may be coupled in different ways to form a whole. In other words knowledge of the behaviour of the parts of a machine is inadequate to predict the behaviour of the whole. This is an important result of cybernetics, and explains the emphasis on system behaviour. 2.7 Feedback In the previous section P and R were joined such that P affected R but R did not affect P. In this relationship, P is said to dominate R: P R In order to avoid confusion with transformations, boxes have been drawn around the machine identifiers to indicate that we are talking about relationships between machines. The more interesting case arises when R also affects P. P 23 R Cybernetics for Systems Engineers – DRAFT F This circularity of action is frequently referred to as ‘feedback’. However, as will be shown , feedback is a particular kind of coupling between machines which has the effect of driving both to a state of equilibrium. 2.8 Immediate Effect and Reducibility We say a variable has an immediate effect on a transformation, if it affects the result during a single update interval. In practice, this is how transformations are characterised, and the workings of the machine deduced. A variable u which has no immediate effect on variable y, may change the value of x, which does, so that after two update steps, u appears to affect y. Other variables may influence y through longer sequences, and these will be referred to as ‘ultimate effects’. Once the set of all immediate effects for all transforms has been derived, ultimate effects may be determined. If a variable has no ultimate effect on another, the second is said to be independent of the first. In a system consisting of a pair of machines with feedback, the parts of each may have an immediate effect on the other. The feedback may be present for ultimate effects, but as far as immediate effects are concerned, the action may be one way, in which case we deduce that the system is less richly connected internally. It is possible for the complexity of connection to be reduced until the two machines have no immediate effect on each other. This permits the behaviour of each machine to be determined independently of the other, and hence the system is termed ‘reducible’. 2.9 Stability As long as we are dealing with finite state machines, the worst behaviour we can expect is a cycling through states in a basin. Ross Ashby overcame this difficulty by considering that for an organism interacting with its environment, one of those states would be ‘death’, which effectively stops further cycling through states. The problem with finite state machines when we consider self correction, is there is no representation of the process of change itself. We can trigger events to change states, and we can include a process called ‘adaptation’ amongst the available transforms, which must be triggered by events. If the only state which triggers adaptation is ‘death’, the adaptation will be too late. In order to cover the process of adaptation, Ross Ashby cheated by introducing an additional state, triggered when the essential variables (those associated with survival) exceed certain bounds. This is just a quantisation of the ‘death’ state, implying that the quantities on which adaptation depends must be quantised to at least two levels. Instability may be defined as the condition where the essential variables are out of bounds for a sufficiently long period for the organism to die. In a mechanical system we have upper bounds on all variables describing the system, where beams break, gear teeth strip, tyres slip, wings stall, gyros topple, measurements saturate – there are myriad and usually fairly obvious ways of defining catastrophic failure of artificial systems. Disaster is normally easily recognisable; it is preventing it from happening which is the difficult task. This is a more general definition of instability, where the system deviates to some point of no return. It is a property of the entire system and not any particular component. 24 Cybernetics for Systems Engineers – DRAFT F 2.10 Equilibrium in Part and Whole A property of coupled systems is that equilibrium of the whole is only possible if all the parts are in equilibrium in the conditions applied by the other parts. Effectively, each part has a veto on the equilibrium condition. Suppose a system is composed of two parts, A and B, which are coupled together. The state of the whole is a vector made up of the state of A and the state of B. Equilibrium requires this vector to remain constant in time, which is only possible if the components remain constant in time, i.e. both A and B must be in equilibrium under the inputs from each other. It follows that no state (of the whole) can be a state of equilibrium unless it is acceptable to every one of the component parts. 2.11 Isomorphism and Homomorphism Isomorphic means ‘similar in pattern’. A map and the countryside it represents are isomorphic. The streamlines around a faired shape form a pattern which is the same as the electric field in a nonconducting medium, about a metal object of the same shape. In studying systems we require models which are isomorphic with the real system. In general, this is impossible, the best we can do is specify the aspects of system behaviour which interest us and construct a model which is an adequate isomorphism of that aspect. We would not use the dynamics model used to design a motor car suspension to determine the layout of the production line to produce the components. Although the same physical objects are represented, our cybernetic requirements differ. More frequently we are concerned with a reduced set of system attributes. What we usually use is a ‘homomorphism’ of the machine of interest, which is a machine having the behaviours of interest. This is achieved when the one-to-one mapping of the states of the isomorphism is replaced with a many-to-one mapping. Consider the machine: M↓ a b c d e i b a b c j a b b c k a b b e d l b e e c c a a The bottom right box contains states e and d, whilst all other entries are a,b,c. We can imagine circumstances where the differences between states a,b,c are irrelevant to our current model of the 25 Cybernetics for Systems Engineers – DRAFT F system. Likewise, d and e could be treated as the same: So we can define a many to one mapping of the original five states to a machine of only two states: N↓ g h α g g β g h The parameters i and j are replaced with α, k and l with β, and the states g and h correspond to separate groupings of the original states. Machine N is equivalent to a simplified version of M. Or, if two machines are so related that a many to one transformation can be found that, applied to one of the machines, gives a machine which is isomorphic with the other, then the simpler of the two is a homomorphism of the more complex. The important point is that although the homomorphism is only a partial description it is complete in itself. What we are doing is associating states, which at a simpler level of representation are indistinguishable. The homomorphism is not simply an approximation, and hence simply wrong, it is correct in its representation of the features of interest. The autopilot designer of a guided weapon may insist that every detail of construction, and every disturbance, be represented as fully as possible. The performance modeller only requires at most a second order lag representation of the airframe response, whilst the battle modeller is happy with coverage diagrams, hit probabilities and flight times. The representation at each level is derived from lower level analysis, but is complete in itself. With modern levels of processing, the tendency is to use the autopilot designer’s model at the battle level, for fear that omission of all the extraneous detail may risk ignoring the potential for ‘emergent’ properties. If the entities concerned had been studied at the detailed level, this fear would be unfounded. Note that our ‘high level’ system description can only arise by simplifying more detailed descriptions. To work from the ‘top level’ down requires the spontaneous creation of information, which is a direct contravention of the Second Law of Thermodynamics. Either we must abandon the idea of having no pre-conceptions, or we must characterise our system from the bottom up, before we have sufficient information to commence the ‘top down’ requirements analysis prescribed by modern systems engineering practice. 26 Cybernetics for Systems Engineers – DRAFT F 3 Cybernetics: Variety 3.1 Introduction In this chapter we revisit the idea of entropy introduced earlier, but extend the ideas to consider the implications of an organism (represented as a deterministic machine) with its environment. Ideas of the previous section considered the functioning of a deterministic machine, which once defined by configuring its inputs followed a fixed behaviour. When considering regulation and control, however, we need to consider the broader set of what it might do, when ‘might’ is given an exact definition. We used the term ‘dynamic range’ to characterise the finite selection of potential signals, more generally we refer to a set of possibilities which the signal may adopt at any sample. For example, if the sequence is: c,b,c,a,c,c,a,b,c,b,b,a it contains twelve elements but only three distinct elements. In this discrete signal case, we say that the signal has a variety of 3. We can measure variety either as the number of options, or as the logarithm to base 2 of the variety, in which case it is measured in bits. 3.2 Constraint An important concept in cybernetics is that of constraint. It is a relation between two sets, and occurs when the variety that exists under one condition is less than exists under another. In an allmale gentleman’s club (an anachronism in the 21st Century) the variety in the sexes represented has dropped from the 1 bit of the human race to 0, indicating that a constraint is present. Laws of Nature impose constraints on potential observations; we do not expect a planet to spontaneously skip to the opposite side of the Solar System within a few seconds. The World is therefore extremely rich in constraints. Indeed, the whole concept of a deterministic machine, comes from recognising that sequences of behaviours show a particular form of constraint. Were the machine output to show no constraint, the observer would conclude that the device was something akin to a roulette wheel, and intrinsically unpredictable. The effect of a machine on variety can be illustrated with the example transformation: A→B B→C C→C The operand may take on three values, whilst the transform has only two, the variety drops from 3 to 2 on applying the machine to a sequence of operands. Evidently, no loss of variety implies a oneto-one transformation. The machine, by applying a constraint, reduces variety. The variety in a machine cannot increase, although it usually decreases, but what happens if we have a machine with input? 27 Cybernetics for Systems Engineers – DRAFT F Since behaviour depends both on the state as well as the input parameter, we shall consider a set of machines, all subjected to the same input but differing regarding their start state. If we observe the variety of states over the set of machines, we shall see it fall to some minimum. When the variety has reached its minimum the input parameter is changed. The change will change the machine’s graph from one form to another, as (for a machine having states A …F: From: A ↓ C ↓ D←B E←F to: A→B ↑ D→ E C ↑ F Under the first input, all machines started at A,B or D would go to D, and those started at C,E or F would go to E. The variety after some time would fall to 2 states. When the input is changed, all machines would then end up at B. The second input causes a further fall in variety. Changing the parameter value to a set of identical machines cannot increase the set’s variety. The continued application of inputs also reduces the dependency of the final state on the initial conditions. This is what Ross-Ashby refers to as the law of experience. Our initial considerations of variety were concerned with efficient coding of messages to make maximum use of channel capacity and/or storage, in this chapter we will consider encoding as the effect the machine has on a message of known variety. ‘Encoding’ as used here refers to the transformation of a message to render it compatible with the medium. The message could use sound waves in the air, electrical impulses in a wire, positions of the arms of a semaphore, the smoke of a fire, reflections of the Sun off a mirror, - the list is practically endless. The physical means varies, yet something is preserved which does not depend on the material processes involved. The operation of the coding is to transform the messages Mi to their corresponding codes C. It is a transformation process, and may therefore be represented as a deterministic machine. However, we note that all messages undergo a one-to-one transformation, so that the machine preserves variety. Indeed, if the coding were multi-stage, each stage must preserve variety if the original message is to be recoverable. Assume we have a machine with input that can be in some state Si, where the number of states is finite. It also has a set of parameters that can take, at each moment, one of the values Pj each of which determines a transformation in the S’s. For simplicity, suppose the machine M can take four states: A,B,C and D; and the parameters can adopt three values: Q,R and S. M↓ A 28 B C D Q C C A B R A C B B S B D C D Cybernetics for Systems Engineers – DRAFT F The complete behaviour of the machine is presented in the above table. If the machine starts in state B, and the input follows the sequence R Q R S S Q R R Q S R, the output will follow B C A A B D B C B C C B. There is no difficulty, given the machine and its initial state, in deducing the output sequence from the input sequence. Having coded the sequence the next issue is whether it can be decoded. The shortest message is a sequence of two values, of which there are nine possibilities: QQ, QR, QS, RQ, RR, RS, SQ, SR, SS Starting with the machine in state A, we have as output: CA, CB, CC, AC, AA, AB, BC, BC, BB The sequence BC repeats, so that there are only 8 possibilities at the output. Starting at B: CA, CB, CC, CA, … Here CA repeats. With C as the starting point: AC, AA, AB, BC, BC, … Again BC repeats. Starting with D: BC, BC,… The output sequence in all cases has lower variety than the input sequence, it follows that the inverse to the encoder does not exist. It is easy to see that if, in each column of the table, every state had been different then every transition would have indicated a unique value of the parameter; so we would be able to decode any sequence of states emitted by the machine. The converse is also true; if we can decode any sequence of states each transition must determine a unique value of the parameter, and thus the states in the columns must all be different. This is the requirement for a perfect coder. Given that we have designed the coder according to the above principles, how do we design a machine to perform the decoding? i.e. how do we recover the original sequence? We note that it is the transition which yields information about the original input, and not the actual value of the coder’s output. We want a machine which accepts transitions as input, but this involves 29 Cybernetics for Systems Engineers – DRAFT F two states which do not exist at the same sample interval. Consider the machine with input (D )represented by the above table: If the start state is r, the input sequence Q S S R Q S becomes: r q s s r q s, which is the same as the input delayed by a single sample interval. A number of such machines will delay the input for as long as is required. The new coder has the vices of the previous one corrected, with only single value transformations: The inverter is the transformation: The input sequence to the decoder is the vector (state of delayer, state of coder) By introducing a delayer, which enables transitions to be detected, provided the coder does not lose variety, an automatic inverter can always be built. We could imagine the input to the coder to originate from a deterministic machine. The variety of the input cannot exceed 3 regardless of the structure of the machine. The coder itself has 4 states, hence a variety of 4, so the variety at the output from the coder is 12. Without the delay, the variety of the output stream is only 4, hence it is impossible to reconstruct the input signal. With the delay, the variety of the input to the decoder is 12, and a potential solution exists. 30 Cybernetics for Systems Engineers – DRAFT F More generally, the output from a sequence of machines cannot exceed the product of the varieties of each, or expressed in logarithmic terms the output variety cannot exceed the sum of the two varieties. We note however, that there are 16 potential pairs of four letters, but not all are selected (AC, for example). Knowing the quantity of variety of the two processes is insufficient to predict how they will actually change. Another way of considering the problem is to think of the coder as a communication channel sending the input message; it cannot transmit a variety greater than 4 at each sample time, but can potentially transmit a variety of 16 in two sample times. Expressed logarithmically, the capacity of the coder, treating it as a channel is log24 = 2 bits per sample. In general: C log 2 N Where N is the number of distinct values the channel can take. 3.3 Incessant Transmission Whilst the finite state deterministic machine is useful for illustrating basic principles, they settle down in a relatively short time and are not representative of the more general interchange of information between machines or between an organism and its environment. We must consider a more comprehensive form of machine and transformation – the non-determinate. Instead of transforming to a particular new state, we may go to some one of the possible states, the selection of the particular state being made by a process that gives each state a constant probability. Such a transformation, and the set of trajectories it produces, is called ‘stochastic’. We can express a deterministic single value transformation by its individual transitions: A→B B→A C→C Another way of representing it is as a matrix of probabilities: ↓ A B C A 0 1 0 B 1 0 0 C 0 0 1 Since the transformation is deterministic, the probabilities have values 0 or 1. The more general case deals with transition probabilities: 31 Cybernetics for Systems Engineers – DRAFT F ↓ A A 0 B 0.9 B C 0.9 0.1 0 0 C 0.1 0.1 0.9 The protocol to construct the table of transition probabilities is the same as for the deterministic case - the number of times each transition takes place is recorded and each tally is divided by the total number of transitions from each state. The resulting sequence of states is known as a Markov chain. The transition matrix probabilities show that some form of constraint is acting, because all outcomes are not equally likely. An unconstrained stochastic process is one for which all transitions are equally likely, so that all elements of the transition matrix are equal. 3.4 Entropy The discussion so far has used the word ‘variety’ to describe the diversity of choice of a symbol set, where the symbol set corresponds to a signal over a continuous domain, we used the term ‘dynamic range’. But the variety is insufficient to describe the propagation of information. Consider the communication with a prisoner in a cell, an accomplice does not have direct access, but may provide him with a cup of coffee via the warden. Now the coffee presents many opportunities for coded communication, shape of cup, black or white, hot or cold, with or without sugar. So to prevent the possibility of a message being concealed behind this apparently charitable act, if it is black, the warden will add milk, similarly will add sugar and will not hand it to the prisoner until it is cold. Furthermore, he will transfer the contents of the proffered cup to a standard mug from the canteen, and will tell the prisoner that he added the milk and sugar. The state of the coffee is always the same one of the possible options, but because the variability is cut to one, no information can be exchanged. In our coding example, a decoder could not be constructed considering the output signals individually, information about the input stream required consecutive samples to be considered. The extra information added to the input stream by the coder was the internal working of the coder, decoding required this information to be removed from the output sequence to reveal the input sequence. More importantly, whatever information was conveyed in the original input was conveyed with a different symbol set, which only needed at least the same variety as the input set. If the symbol set had lower variety, it would be impossible to avoid repeating symbols in the columns of the coder’s transition matrix, so a perfect coder would be impossible. This is an example of the Law of Requisite Variety, which will be covered in detail in later sections. 32 Cybernetics for Systems Engineers – DRAFT F Information then does not depend on the symbols used, nor is it specific to the physical medium used to convey it. It is evident that more can be conveyed per unit time if the variety of the symbol set is large. A symbol set with only one entry cannot convey information. This is the fundamental reason for the logarithmic measure of variety; using it, the single symbol had zero variety. Also the minimum amount of information which can be conveyed is when the differences between any two signals is not obscured by noise. Since we are always, in one form or another comparing two entities; noise and signal, it is reasonable to employ a logarithm to base 2 and deal with information in bits. Variety characterises the potential of a symbol set to convey information, yet actual messages made up of the symbols are unlikely to use all of them. A message consisting of the same symbol repeated indefinitely has no variety at all, yet a message which uses a different symbol from the set at each sample has the same variety as the symbol set itself. Information is a property of the diversity of symbols in a sequence. If we are dealing with a Markov chain, we can characterise the symbol set by the probability of each symbol occurring. Suppose we have two distinct messages a,b, with probabilities of occurrence pa, and pb. The options are mutually exclusive as only one symbol may be selected at a time. Also: p a pb 1 If we consider a channel which has N potential levels and partition it into two intervals say 0 to n and n to N, and draw samples from a uniform random sequence of width N, whenever the sequence is above n, we add it to b, if below, we add it to a. In this way we ensure the two messages are distinct: pa n N n , pb N N Note: we could have partitioned N any way we choose provided the total intervals corresponding to each message are proportional to the probabilities. Since each message uses a reduced set of the potential symbols it cannot have the complete variety of the symbol set. The variety of a is: log 2 n log 2 pa log 2 N The variety conveyed by both messages is, therefore: V pa log 2 pa pb log 2 pb pa pb log 2 N Or: V pa log 2 pa pb log 2 pb log 2 N If the channel has 2 symbols each drawn with equal probability: V 0.5 log 2 2 0.5 log 2 2 log 2 2 0 33 Cybernetics for Systems Engineers – DRAFT F If one symbol is repeated indefinitely, this formulation would put N=1, and again the variety would be zero. The variety which we have calculated is that of the message plus that of the channel, much as we did in the coder example. The variety of the channel is log2N, regardless of the value of N, so the variety of the set of messages which can be conveyed is: V p a log 2 p a pb log 2 pb The above reasoning may be extended to any size symbol set, also the negative variety obtained from taking logarithms of fractions introduces an inconvenient minus sign. Finally, it is normal to use the symbol H to denote the variety of a stochastic signal: H pi log 2 pi In this form, it is called ‘entropy’, which has the merit of fewer syllables than ‘stochastic variety’, but tends to be confused with the entropy of thermodynamics. There is a relationship between the two entropies but it lies well outside the scope of the current discussion. In the interests of clarity, the deterministic counterpart of entropy i.e. ‘variety’ will be used as far as possible. 34 Cybernetics for Systems Engineers – DRAFT F 4 Cybernetics: Regulation 4.1 Requisite Variety Regulation is the process of ensuring that the states of a system remain within permissible bounds, as the organism interacts with its environment. I.e, states such as ‘death’ are to be avoided. The action of the regulator could be considered the resolution of a conflict between the environment and the organism which we hope the regulator wins. With that in mind, consider a game in which one player, R (regulator) is trying to reach an outcome ‘a’, and the other player, D (disturbance) is trying to prevent it. D moves first, playing gambit 1,2 or 3, and then R responds by playing either of α, β or γ. The outcomes for each pair of gambits is summarised in the following table: α β γ 1 b a c 2 a c b 3 b a c In this particular case, R can always win, as each row contains an option whose outcome will be ‘a’. If R acts according to the transformation: 1→β 2→α 3→γ The outcome will always be ‘a’. A less favourable game is presented in the following: α β 35 γ δ 1 b d a a 2 a d a d 3 d a a a 4 d b a b 5 d a b d Cybernetics for Systems Engineers – DRAFT F Evidently, if ‘a’ is the target, R can always win. If ‘b’ is the target, R can only win if D plays 1,4 or 5. If the target is ‘c’, R’s position is hopeless. In order to restrict the consideration to the non-trivial, we shall consider only cases where R must make a move in response to D, i.e. there will be no repeated entries in the game matrix columns, enabling R simply to repeat the previous move. Thus each move of D must be countered by a move by R. As an example of such a game consider the following table: α β γ 1 f f k 2 k e f 3 m k a 4 b b b 5 c q c 6 h h m 7 j d d 8 a p j 9 l n h R’s set of responses to each of D’s moves converts the game into a transformation. This is singlevalued because R may only make one move at a time. An example might be: 1→γ 2→α 3→β : 9→α This transform uniquely specifies the outcome: 1→k 2→k 3→k : 9→l 36 Cybernetics for Systems Engineers – DRAFT F We see that the variety of the set of outcomes cannot be less than D’s variety minus R’s variety If R always plays the same move, the outcome will have variety equal to the variety in D’s moves. If R uses two moves, the outcomes can be reduced to half the potential outcomes, if three moves, to one third, and so on. Therefore, if the variety is to be reduced to a specified level, R’s variety must be increased to at least the appropriate minimum. Expressing the varieties logarithmically, we have for the variety of the outcome Vo: Vo VD VR Where the subscripts denote D and R. This is the Law of Requisite Variety. The variety of a regulator must be at least equal to that of the disturbance if the variety of the essential variables is to be kept at zero. This can be expressed, for stochastic processes, in terms of entropies. H E H D H D R H R Where HD(R) is the entropy in R, given D. If the relationship between R and D is deterministic, HD(R)=0, and the expression is the same as above, but in terms of entropy. More typically, D and R play simultaneously, and the strategy consists of selecting the probabilities of a random play which on average achieves R’s requirements.. The entropy of the outcome cannot be less than the difference in entropy of the disturbance and the regulation. In this more general case, we could refer to the Law of Requisite Entropy. Except in the most benign of environments, an organism will be subjected to disturbances which will threaten, if it fails to perform some regulatory function, to drive the essential variables outside their proper range of values. If regulation is to be successful, it must be related to the disturbance via a relationship akin to the games described above, which summarise the outcome of the interaction of the regulator and the disturbance. Perfect regulation matches the organism’s entropy to that of the environment, resulting in zero entropy. The less perfect the regulation, the more entropy the system will exhibit. There is a widespread myth that complexity of behaviour increases with system complexity ( i.e. component count). The principle of requisite entropy shows this to be complete nonsense. We do not add servos and guidance system to an unguided rocket to render its behaviour less predictable, we do so to make it more predictable, to the extent that it hits the target. 4.2 Error Controlled Regulation Designing a controller to deal with a closed set of disturbances is simply a matter of listing the disturbances and devising strategies to deal with them, as and when they were detected. Life would indeed be simple if the environment presented us with a closed set of disturbances. 37 Cybernetics for Systems Engineers – DRAFT F Real environments present us with open sets of disturbances, so our organism must adapt to changing circumstances or die. The Law of Requisite Entropy provides us with the clue as to how this can be achieved, and indeed how it has been achieved in Nature. The organism may not have sufficient a priori knowledge of the environment to construct a strategy from first principles, but it will know whether its current strategy works or not. By acting on the degree of success or failure of a trial strategy, the organism may survive environmental change, perhaps not as an individual, but as a species, depending on how effective the process of formulating new strategies is. Adaptation to an open environment requires the ability to detect when the system entropy is reduced near zero. What is needed is a null-seeking system. Feedback is the adjustment of the organism’s entropy based on the difference between the desired system entropy and its actual value. Ross Ashby demonstrated that adaptation by random selection would work if there were few changes within in given timescale. The ability to handle ever greater numbers of changes within a fixed timescale forms the basis of an eclectic measure of intelligence. 38 Cybernetics for Systems Engineers – DRAFT F 5 Practical Finite State Machine Control 5.1 Introduction The earliest automata devised by human beings were essentially finite state machines implemented in clockwork. A toy music box, steam engine valve gear, fairground organs, battleship gun loaders, capstan lathes, the Jacquard loom, to mention but a few, are cases where human effort was replaced by a mechanical system for performing highly repetitive, although sometimes very complex, tasks. Over the years, cams, gears and mechanical links have been replaced by electromagnetic relays and finally by digital processors. The convenience and flexibility of the digital processor, over the limitations of earlier technologies has led to a significant increase in the amount of this type of automation. From the principle of requisite variety, such systems can only operate in a controlled environment, where the set of disturbances is more or less closed. Such environments are commonplace – they are called factories. The matrix description of the finite state machine used to explain the fundamentals of cybernetics in the previous three chapters is too abstract to be of practical value in this context. As we shall see, the UML/high level language (such as C) of the software engineer is not the preferred approach either. To understand why this is so, we need to know a little about a typical industrial programmable logic controller (PLC). However, the method applies to all finite state machine controllers, regardless of their physical implementation, otherwise it would have no place in this book. 5.2 The Relay and Logic Functions A relay is a device which deals with all-or nothing signals. In its most abstract form it is a device in a low power consumption circuit (electrical, compressed air hydraulic, or just as a software entity), which when activated, switches the flow in a high power circuit on and off. An electrical relay consists of a coil of wire on an iron core which attracts an armature when the power is on, causing contacts to separate or meet. How it is implemented is of no concern. The relay, whether implemented as software or hardware, is the fundamental building block of this type of system. The convention for representing relays is presented in Figure 5-1. A pair of normally open contacts is represented as a pair of spaced vertical lines, similar to the conventional representation of a capacitor. A pair of normally closed contacts is the same but with a diagonal line drawn through it. The controlling ‘coil’ is drawn as a circle. The contacts are associated with a specific coil by giving them the same letter identifier. For example, when coil A is activated the normally open contacts A close and the normally closed contacts A open. Using this representation, we can implement the three logic functions, AND, OR and NOT, from which all logic functions can be derived. 39 Cybernetics for Systems Engineers – DRAFT F Figure 5-1: Basic Symbol Set The NOT function is a set of normally closed contacts, so that when the coil is energised by the input, the circuit opens. AND consists of wiring the normally open contacts from each input in series, and OR is obtained by connecting them in parallel. Thinking in terms of opening and closing contacts is a very natural way to think of circuit function, but by identifying the serial and parallel paths through the control circuit it is a simple matter to map them into logic functions which once would have been actually implemented using logic integrated circuits. Figure 5-2 : Logic Symbol Equivalents 40 Cybernetics for Systems Engineers – DRAFT F 5.2.1 Example : Capacitor Discharge Unit A model railway is required to operate the points automatically on detecting an approaching train. The points motor consists of a solenoid which is usually energised manually by applying a currently transiently via a switch. This provides an impulse which causes the points to move. The part of the problem which concerns us is when the train is approaching from a branch, with the points normally set for free movement along the straight. A detector of some kind (e.g. a reed switch in the track which detects a magnet under the train) is triggered by the approaching train. Evidently, if the points do not change, there will be an accident. Figure 5-3: Model Railway Example Suppose we have a capacitor with sufficient charge to operate the solenoid which changes the points. Detecting the approaching train could then close a pair of contacts causing the capacitor to discharge through the solenoid. That would be fine, apart from the fact that once the train passes the detector the contacts will open again, and there might not be enough energy in the partial discharge to change the points. What we need is a means of keeping the contacts closed until the capacitor is fully discharged. 41 Cybernetics for Systems Engineers – DRAFT F The simplest means of achieving this is to place normally open contacts in parallel with the switch. One the relay is triggered it will close its own circuit and remain on even though the initial trigger is removed. This is called a latch circuit. The implicit assumption in UML state diagrams is that this latching action is always present. This would be the case for digital computers which invariably have latched outputs, but a systems approach cannot pre-judge the hardware solution. The latch causes the solenoid to apply its full impulse, ensuring the points change. The train traverses the junction, requiring the points to be returned to their initial state. This is achieved with a second detector and latching relay which closes the points. That seems OK except for two potential problems – how did the capacitors get charged in the first place? Also, leaving the system with both capacitors discharged and the points set straight does not help the next train coming from the branch. 42 Cybernetics for Systems Engineers – DRAFT F Evidently, when we latch one relay, we must unlatch the other, so there will be a normally closed contact pair driven by the other relay in series with the contacts causing the latching. This returns the system to its original state when the train passes the second detector. The resulting circuit is called an SR – latch (for set-reset). The same circuit expressed as logic functions is as shown in Figure 5-4. This may look neater, but we would have to write out the truth table to convince ourselves that it would work, whilst the relay-based representation is more readily understood. Figure 5-4: Implementation of SR-Latch as Logic Functions We see that the SR latch looks like two rungs of a ladder between the power line and the earth. More complex systems have many rungs running concurrently, hence this system characterisation has become known as ‘ladder logic’. The final issue is that of charging the capacitors Each relay requires a ‘normally closed’ set of contacts to connect its capacitor to the power supply. In practice, the capacitor cannot be charged 43 Cybernetics for Systems Engineers – DRAFT F directly from the power supply, as the charging current would initially be very high, sufficient to damage whatever is used to implement the ‘normally closed’ contacts. A ballast resistor would be included in this path to limit the charging current, but this has been omitted for clarity. We haven’t finished yet. Suppose a train on the straight track passes the detector. It will cause the capacitor to discharge and latch the relay in the discharge position. Worse still, it will inhibit the latching of the relay on the ‘branch’ side. In order to avoid this, further sensors are needed to detect the presence of the train in the straight section, which inhibit the operation of the discharge unit for the time it takes to traverse the junction. The resulting system is by no means robust, as it relies on its own correct operation to determine its state; it is control by decree. It is preferable to sense the position of the points explicitly, so that the system more closely resembles a feedback controller rather than a set of open loop controllers in close formation. 5.2.2 Reversing a DC Motor Operating a DC motor seems a similar problem to the capacitor discharge unit problem with forward and reverse push buttons operating an SR latch. It is evident from the ladder diagram that there is potential for a direct short circuit across the power supply. No switching process is instantaneous and there remains a possibility of a transient as the power supply is briefly shorted out. This might be sufficient time to weld relay contacts or blow power transistors, but also may introduce a spurious spike into other control lines of the system, causing spurious switching and unpredictable behaviour. In practice we should expect the power supply to be switched out during the polarity reversal. In fact, if the motor is coupled to a high inertia load, reversing the polarity of the applied voltage may give rise to very high currents until it has spun down. Delays may be included to give the motor time to slow down before the voltage is reversed. 44 Cybernetics for Systems Engineers – DRAFT F 5.2.3 Alarm Circuit It would appear reasonable to design an alarm circuit with a number of activating switches in parallel, all connected to the warning device (e.g. a siren). This is not as sensible as it first appears because it relies on the connection between the switches and the siren remaining intact. However, the most likely mode of failure is expected to be an open circuit. A break in the wire could lead to one of the alarm triggers failing, or if the wire connecting the common wire to the siren is broken, none of the triggers will work. It is important to ensure that as far as possible the system should be ‘fail safe’, and relying on the integrity of the cabling is bad practice. A better design would connect the trigger contacts as normally closed in series providing power to a relay whose normally closed contacts are in series 45 Cybernetics for Systems Engineers – DRAFT F with the alarm. If any of the triggers breaks the circuit, or if there is an accidental break, the relay will switch off and the contacts will close, setting off the siren. 5.3 Conclusions The examples presented illustrate just how inadequate a system description based on logic alone can be. The types of fault which arise from failure to take real component behaviour into account are usually show-stoppers, and not mere ‘implementation details’ as is too often claimed. The logic representation used in UML assumes an ideal behaviour of the hardware which is not consistent with the real world. Divergences from the ideal behaviour often give rise to system critical issues which are not readily identifiable from the state charts. The desired logic is seldom a source of problems in real systems. Difficulties typically arise because the mathematical logic is only an approximation to the real world, and a formulation is needed which can take account of this discrepancy between ideal and actual behaviour. Even for finite state machine controllers, it is necessary to have a formulation which enables the effects of real world hardware to be taken into account, and mitigating measures taken. For this reason all commercial programmable logic controllers (PLCs) are usually programmed using ladder logic, with software to convert the pictorial system description into code. Evidently we must take real hardware behaviour into account, even when our system is nominally a finite state machine. How much less can we afford to ignore the intrinsic behaviour of less forgiving systems. 46 Cybernetics for Systems Engineers – DRAFT F 6 Background Calculus 6.1 Basic Ideas 6.1.1 Introduction Dealing with complex environments with open sets of disturbances is a more daunting prospect than controlling systems which are amenable to simple logic, and we need to extend our mathematical tool set accordingly. The engineering temperament is characterised by a pragmatic view of the world. Motivation to learn is predicated on an awareness of the potential usefulness of the material. This introduces the difficulty that the most useful techniques tend to have obscure beginnings, which may test the patience of the natural engineer. At the same time, the engineer is expected to have a certain curiosity as to why the world is the way it is. For example, rather than view the bizarre contraptions which emerged during the infancy of a new technology (e.g. early attempts at aircraft or motor cars) merely as objects of mirth, the engineer will consider why the particular solutions were adopted, and what their merits were. Figure 6-1: We Calculate Area by Tesselating the Plane with Squares At the elementary level, we are motivated largely by the handling of money to learn arithmetic. Also we can see a potential use for algebra in calculation of areas and volumes. These enable us to estimate quantities of paint, numbers of floor tiles or whether an object will float. So it is from the practical question of finding areas that we begin our discussion here. 47 Cybernetics for Systems Engineers – DRAFT F 6.1.2 Areas Bounded by Straight Lines I expect practically every reader to know all the results in this section, but there is method in my madness, so I hope you will not feel too insulted in being presented with elementary results which you might consider beneath your dignity. In my experience there is no shortage of individuals who can use results, but I am frequently shocked to discover how rarely the underlying principles are actually understood. Being able to manipulate algebra does not in itself constitute understanding. My intention is to introduce the idea of extrapolating from a finite set size to an infinite set, right from the beginning. I have included some revision boxes in this section, which are essential background to what follows. When presented with a physical quantity, we seek a unit to measure it in. When it comes to area, we begin with plane surfaces, so we base our unit on a shape which will tessellate (tile) a plane surface. The candidates are triangles, squares or hexagons. Conventionally, we use the square. The reader is welcome to derive the basic ideas from equilateral triangles or regular hexagons – it could be interesting and amusing. We can only talk about curved surfaces once we consider vanishingly small plane surfaces, which we shall do presently. 48 Cybernetics for Systems Engineers – DRAFT F Figure 6-2: If the Edge is not a Multiple of the Unit, the number of squares becomes large, so multiply length by breadth to get the same answer In order to demonstrate the (almost trivial) result that the area of a rectangle is equal to length times breadth, we had to show that the formula may legitimately be applied to real numbers, whilst our basic definition of the unit of area applies only to integers. If we continued counting squares the problem would become ever more tedious as we are forced to take smaller squares. In the extreme, if the sides were irrational multiples of the unit length, there would be an infinite number of vanishingly small squares. However, by multiplying length by breadth we can find the area without such potential difficulties. This process of finding a way to sum up an infinite number of infinitesimal pieces is, in a nut shell, what calculus is all about. 49 Cybernetics for Systems Engineers – DRAFT F Figure 6-3: Once We Can Deal with an Infinte number of Infinitesmal Squares we are no longer limited to Rectangles. The areas of parallelograms and triangles follow from the area of a rectangle, from elementary geometry, using congruent triangles. We can show the same result by what appears to be a laborious method, but it is one which can be extended to less tractable shapes. One result, which is used frequently in the following sections is the fact that: 𝑥0 = 1 Regardless of the value of x. To appreciate this, apply the rule that we add indices when we multiply: 𝑥𝑛 𝑥 0 = 𝑥 𝑛 𝑥 −𝑛 = 𝑥 𝑛 = 1 Regardless of the values of x or n. Also, all angles are in radians. There is something natural in defining the unit of angle as the rotation of a wheel which rolls a distance equal to its radius, as opposed to dividing a circle arbitrarily into 360 divisions. 50 Cybernetics for Systems Engineers – DRAFT F Figure 6-4: Define Angles as the Arc Length Divided by the Radius, or the Distance Rolled Divided by the Radius 6.1.3 Coordinate Geometry I assume that the reader is familiar with the idea that a straight line may be represented by the formula: 𝑦 = 𝑚𝑥 + 𝑐 Where any vertical coordinate y may be calculated from the horizontal coordinate x. Figure 6-5 : Equation of a Straight LIne The line has slope ‘m’ and intercept ‘c’ on the vertical axis. Using this representation, we can calculate the area of a triangle ABC by calculating the areas under the line segments AC and CB, and subtracting the area under AB. 51 Cybernetics for Systems Engineers – DRAFT F Figure 6-6 : Define a Triangle by the Coordinates of its Corners The coordinate geometry approach is extended to other geometric shapes, which would give us the potential to calculate the areas of figures bounded by curved edges. 6.2 Areas The calculation requires us to calculate the area below the curve for the set of curves forming the edges of the figure. Hence the problem reduces to one of calculating the area under the curve. Figure 6-7: AS the Number of Strips Becomes Infinite, this Approximation to the Area becomes Exact The general straight line: 52 Cybernetics for Systems Engineers – DRAFT F 𝑦 = 𝑚𝑥 + 𝑐 Intersects the x axis (y=0) at –c/m. We shall consider the area under the line from this point to x=a. The height of the triangle at this point is: 𝑦 = 𝑚𝑎 + 𝑐 The area under the curve is known for this simple case; it should be: 1 (𝑎 2 𝑐 + 𝑚) (𝑚𝑎 + 𝑐) = 𝑚 (𝑎 2 𝑐 2 + 𝑚) Rather than apply the formula, we divide the figure into rectangular strips, each of width w. Each strip has area: 𝑆 = 𝑤𝑦 An approximation to the area is obtained by adding up the areas of the strips, the more, and narrower the strips, the more accurate will be the estimate of area. Numbering the strips 1 to n, the x coordinate of the strip number i is: 𝑐 𝑥𝑖 = 𝑖𝑤 − 𝑚 The height of this strip is: 𝑦 = 𝑚𝑖𝑤 + 𝑐 − 𝑐 = 𝑚𝑖𝑤 The area of this strip is: 𝑆𝑖 = 𝑚𝑖𝑤 2 The area under the line is found by adding all n strips: 𝐴 = 𝑚𝑤 2 ∑𝑛1 𝑖 (Read the Greek symbol, sigma ∑ as the ‘sum from the lower value to the upper value’, in this case from 1 to n) In other words, we can find the area if we know the sum of the series; 1+2+3+...+n This is a familiar arithmetic series. The sum is found by writing the series backwards and summing with the original sequence: 2 ∑𝑛1 𝑖 = (𝑛 + 1) + (𝑛 − 1 + 2) + (𝑛 − 2 + 3) ⋯ + (𝑛 + 1) Collecting up terms: ∑𝑛1 𝑖 = 53 𝑛(𝑛+1) 2 Cybernetics for Systems Engineers – DRAFT F The area becomes: 𝐴= 𝑛(𝑛+1) 𝑚𝑤 2 2 𝑛(𝑛+1) (𝑎 2𝑛2 = 𝑐 2 + 𝑚) This may be written: 𝐴= 𝑚 (𝑎 2 𝑐 2 𝑚 + ) + 𝑚 (𝑎 2𝑛2 𝑐 2 𝑚 + ) As n becomes larger the second term on the right hand side becomes smaller. As n becomes infinite this term becomes identically zero, and we are left with the correct equation for the area under the straight line. This approach allows us to consider what will happen when the number of strips becomes infinite. If the bounding edge is a curve, rather than the straight line, the method remains the same, we end up with a term in which the number of strips appears in both numerator and denominator, and other terms which become zero as n becomes infinite. 6.2.1 Curved Boundary As an example of a curved boundary, consider a contour governed by the equation: 𝑦 = 𝑥2 Before proceeding, consider the simpler case of a straight line through the origin, bounded by the vertical line x=z. The equation is: 𝑦=𝑥 As before, the width of the strips is: 𝑧 𝑤=𝑛 The height of the ith strip is: 𝑦𝑖 = 𝑖𝑤 The area is: 𝑆𝑖 = 𝑖𝑤 2 = 𝑖 2 𝑧 𝑛2 Summing over the n strips as before yields an area: 1 𝐴 = 2 𝑧2 An even simpler case is the rectangular region under the horizontal line y=1, between the vertical lines x=0 and x=z. The area is, by inspection: 𝐴=𝑧 54 Cybernetics for Systems Engineers – DRAFT F Figure 6-8 : We can Approximate Any Curve by an Enormous Number of Tiny Strips The new curve: 𝑦 = 𝑥2 is called a parabola. Following exactly the same procedure as for the straight line, we find the area of the ith strip as: 𝑖2 𝑆𝑖 = 𝑖 2 𝑤 3 = 𝑛3 𝑧 3 The problem reduces to finding the sum of the squares of integers, then neglecting the terms which become ever smaller as the number of strips increases to infinity. We notice that the sum of n terms, each equal to 1 is equal to n, and the sum of the integers up to n is given by n(n+1)/2. These contain powers of n which are one greater than the power to which each of the terms is raised (in the first case the terms are raised to the power 0, resulting in n raised to the power of 1, in the second case, the terms are raised to the power 1, resulting in a formula with terms in n2). Whatever the formula for the sum of the squares of integers, it cannot include powers of n greater than 3. If it did the estimate of the area would increase without bounds as the number of strips increased, which is clearly ridiculous, as the actual area is clearly finite. 55 Cybernetics for Systems Engineers – DRAFT F The same may be said of higher order curves, for y=xm we need to sum im, and the area of each strip contains a term 1/nm+1. Consequently the formula for the sum of integers raised to the mth power cannot contain terms higher than m+1. Using this as a clue, we deduce that the formula for the sum of the squares of integers will not contain terms of higher order than n3. We guess the solution is of the form: ∑𝑛0 𝑖 2 = 𝑎𝑛3 + 𝑏𝑛2 + 𝑐𝑛 + 𝑑 Where a,b and c and d are constants. The series is from zero to n, which, since the first term is zero, is exactly the same as running the series from 1 to n, except that if n=0, the series must sum to 0, so d must be zero. If this formula is correct for a series of length n, it will be true for a series of length n+1, and also for one of length n-1. For a single term: 𝑎+𝑏+𝑐 =1 For a sequence of length 2, we have: 8𝑎 + 4𝑏 + 2𝑐 = 5 For a sequence of length 3, we have: 27𝑎 + 9𝑏 + 3𝑐 = 14 These three equations may be solved for a,b and c, yielding an expression for the sum of the squares of integers: 1 1 1 𝑎 = 3,𝑏 = 2,𝑐 = 6 As far as finding the sum of an infinite number of strips is concerned, only ‘a’ is relevant, because the terms in n2 and n tend to zero when divided by n3. 6.2.2 Mathematical Induction We have found a formula, which we know works for n=0 to n=3, but there is no reason to believe it will work for any other value of n. In order to demonstrate that the formula is correct, we show that it works for a sequence of length n, and for one of length n+1. If the difference in these sums is equal to the (n+1)th term of our series, i.e (n+1)2, we can say that for arbitrary n the formula applies also to n+1, so if it holds true for, say, n=1, 2 or 3, it will be true for all n. If the formula is correct; the expression: 1 ((𝑛 3 56 1 1 + 1)3 − 𝑛3 ) + 2 ((𝑛 + 1)2 − 𝑛2 ) + 6 ((𝑛 + 1) − 𝑛) Cybernetics for Systems Engineers – DRAFT F Should be equal to (𝑛 + 1)2 = 𝑛2 + 2𝑛 + 1 Evaluating the expression shows that the two are, in fact equal, so we can conclude that the sum of the squares of the integers is given by: ∑𝑛1 𝑖 2 = 𝑛3 3 + 𝑛2 2 + 𝑛 6 From which it follows that the area under the parabola is: 𝐴= 𝑧3 3 This is one-third the area of the rectangle containing it. 6.2.3 Higher Order Polynomials We could follow the same procedure for curves of the form: 𝑦 = 𝑥𝑝 Where p is a positive integer. If we summarise the results so far, as in the following table, a pattern is becoming apparent. It seems for this small sample at least: 𝐴= 𝑧 𝑝+1 𝑝+1 P y A 0 1 z 1 x z2/2 2 x2 z3/3 We have seen that the process of finding the area under the curve amounts to summing the integers raised to the appropriate powers. The formulae for the sum contains terms of order one greater than the power of the terms of the series, this is the reason why the index in the expression for the area is one higher than that for the formula for the curve. The general formula for the sum takes the form: ∑𝑛0 𝑖 𝑝 = 𝑎𝑛𝑝+1 + 𝑏𝑛𝑝 + 𝑐𝑛𝑝−1 ⋯ Also: ∑𝑛+1 𝑖 𝑝 = 𝑎 (𝑛 + 1)𝑝+1 + 𝑏(𝑛 + 1)𝑝 + 𝑐(𝑛 + 1)𝑝−1 ⋯ 0 57 Cybernetics for Systems Engineers – DRAFT F The difference between these is (n+1)p. Taking the difference term by term, we have: For the term having ‘a’ as coefficient: 𝑎((𝑛 + 1)𝑝+1 − 𝑛𝑝+1 ) The first term will contain n multiplied by itself p+1 times, i.e. np+1, all the other terms will contain lower powers on n, so the term in np+1 cancels in the expansion. By the same reasoning, the term in np cancels in the expression having b as coefficient, so that the remainder of the expressions are polynomials of degree np-1 or lower. The polynomial multiplying ‘a’ contains terms of up to np and is the only one containing terms of this order. We need to know what the second term in the expansion of a polynomial is in order to determine the coefficient of np. In order to proceed, we need to be able to calculate the coefficients of a polynomial when it is expanded term by term. Consider a few low order polynomial expansions: (𝑛 + 1)0 = 𝑛0 + 0 × 1 (𝑛 + 1)1 = 𝑛1 + 1 × 1 (𝑛 + 1)2 = 𝑛2 + 2𝑛 + 1 (𝑛 + 1)3 = 𝑛3 + 3𝑛2 + 3𝑛 + 1 It looks as if the coefficient on nx-1 in the expansion of (n+1)x is just x. Let’s see if this is true for any value of x. Expanding (n+1)x we know that there can only be one term in nx. We wish to know how many terms there will be in nx-1. We may factorise the expansion, assuming the coefficient of the second term in the expansion is equal to the degree of the polynomial: (𝑛 + 1)(𝑛 𝑥−1 + (𝑥 − 1)𝑛 𝑥−2 ⋯ ) = 𝑛 𝑥 + (𝑥 − 1)𝑛 𝑥−1 + 𝑛 𝑥−1 ⋯ The coefficient of nx-1 is therefore equal to x. So if our hypothesis is true for x, it is also true for x-1. As it is true for x=0,1,2 and 3 it must be true for all values of x. So it follows that the actual term in np is (p+1)np. All other coefficients are associated with polynomials in n of order less than p. The (n+1)th term, contains the term np, so that by equating the coefficients of np on both sides of the equation, we have, for a curve defined by xp, the value of the only significant coefficient in the calculation of the area is: 𝑎= 58 1 𝑝+1 Cybernetics for Systems Engineers – DRAFT F So for any curve defined by a polynomial consisting of positive integer powers of x, we have no need to run through the rigmarole of finding the sums of infinite series, we just apply the relationship: 𝑥𝑝 → 𝑧 𝑝+1 𝑝+1 Now graphs may be used to present data other than the shape of a physical boundary. Practically anything can be represented as a graph, and the area under the graph usually has great significance. The discussion was started with the simple application to physical areas to provide an early indication that this most certainly is not ‘useless knowledge’. This process of finding areas under curves by summing an infinite number of pieces (in this case vertical strips) is given the name ‘integration’. The conventional way of representing it is by means of the old English s (for sum). The result is called the ‘integral’ of the argument, and is written: 𝑧 ∫0 𝑥 𝑝 𝑑𝑥 = 𝑧 𝑝+1 𝑝+1 The ‘dx’ is equivalent to the width of the strip, w, as the number of strips becomes infinite. (say’ the integral from zero to z of xp dx’). 6.2.4 The Circle By far the commonest curve we are likely to meet is the circle, which unfortunately cannot be represented as the sum of a number of polynomials. If we restrict ourselves to positive values of x and y, we can represent a quadrant of a circle with the formula: 𝑦 = √𝑟 2 − 𝑥 2 Where r is the circle radius. If we can find the area of this quadrant, the area of the circle follows by quadrupling the result. The range of values x can take lie in the range 0 to r, so dividing this range into n strips, we have the width of each strip: 𝑟 𝑤=𝑛 The height of the ith strip is: 𝑖𝑤 2 𝑟 𝑖 2 𝑛 𝑦𝑖 = 𝑟√1 − ( ) = 𝑟√1 − ( ) The area of the ith strip is: 𝑆𝑖 = 𝑟2 √1 𝑛 𝑖 2 𝑛 −( ) The calculation requires us to evaluate the following sum: 2 1 𝑖 ∑𝑛1 √1 − ( ) 𝑛 59 𝑛 Cybernetics for Systems Engineers – DRAFT F Tricky! We could evaluate it for ever increasing values of n until the error in consecutive iterations becomes small. This is tedious to attempt manually, but should only take a few seconds on a computer. The sum converges very slowly, as it takes the computer 25000 terms to reach a value of 3.1412 for π. There are more efficient algorithms for calculating π, but this is good enough for most purposes. Looking at the geometry, it is evident that vertical strips are unlikely to provide an accurate approximation to the curve as x approaches r. Taking the area below a region between the verticals through x=0 and x=r/√2 , will allow us to calculate the area of the curved sector, by subtracting the area of a square of side r/√2 from the area under the curve. The r/√2 comes from the x coordinate of the intersection of a line at 45 w ̊ ith the circle. The required sum for this region now becomes: ∑𝑛1 1 √1 √2𝑛 1 𝑖 2 2 𝑛 − ( ) This takes 2000 terms to produce a value for π which is correct to four decimal places. Figure 6-9 : 1/8th Is easier to Calculate than 1/4 of the CIrcle 60 Cybernetics for Systems Engineers – DRAFT F I hope the natural curiosity of the reader will motivate him or her to think up more efficient alternatives to those presented here. We can calculate the perimeter of the circle from the ring (annulus) formed by the region between two concentric circles. Taking the radius of the inner circle as r-d, we have for the area: 𝐴 = 𝜋(𝑟 2 − (𝑟 − 𝑑)2 ) = 2𝜋𝑟𝑑 − 𝜋𝑑2 As the ring tends to zero thickness, i.e. d→0, its area tends to the length of the perimeter P multiplied by the width: 𝑃𝑑 = 2𝜋𝑟𝑑 − 𝑑2 If d=0, this yields the result for the circumference of a circle: 𝑃 = 2𝜋𝑟 The area of an ellipse follows a similar procedure. An ellipse is essentially a squashed circle, so that the equation in the positive quadrant is: 𝑥 2 𝑦 = 𝑏√1 − (𝑎) Where a is half the length of the major axis and b is half the length of the minor axis. Dividing the semi major axis into n strips, the area of the ith strip becomes: 2 1 𝑖 𝑆𝑖 = 𝑏𝑎 √1 − ( ) 𝑛 𝑛 𝜋 4 We have just shown that the rather nasty series obtained from the function of i sums to , so the area of the ellipse is given by: 𝐴 = 𝜋𝑎𝑏 We cannot find the perimeter of the ellipse by considering the region between a smaller and a larger ellipse, because in this case the gap between the two curves is not constant. 6.2.5 Concluding Comment We can describe a large number of curves with polynomials, i.e. expressions of the form: 𝑦 = 𝑎𝑛 𝑥 𝑛 + 𝑎𝑛−1 𝑥 𝑛−1 + 𝑎𝑛−2 𝑥 𝑛−2 ⋯ 𝑎0 Where the ai are constants. The boundaries of the area concerned may be defined in this way and the areas under them found by integrating the polynomial term by term. The area of the enclosed region may then be found by adding and subtracting areas as appropriate to the particular boundary. The importance of integration, and calculus in general, is that it deals with things which are changing continuously; the curves are smooth, not discrete steps, and this is how phenomena in nature (at least at the macroscopic scale relevant to most human experience) behaves. It is an extremely efficient means of counting an infinite number of squares. 61 Cybernetics for Systems Engineers – DRAFT F Furthermore, any quantity may be plotted on a graph; the method is not limited to just finding the areas of plane figures bound by curves, it can be used to provide insights into any phenomenon which can be plotted on a graph. As we can define the geometry as algebraic equations, and may then use integration to find areas, there is no need to actually draw the graph. Indeed, the method may be extended to functions of two variables, which require three dimensional surfaces to represent them. The methods apply to functions of more than two variables which cannot be represented as a three dimensional object at all. 6.3 Some Applications 6.3.1 Introduction Some readers, who may have encountered calculus before, may be surprised at the order in which I have chosen to present the subject. Most texts begin with the inverse process of integration which is called differentiation. It is my belief that this, more usual, presentation is the reason calculus is popularly perceived as ‘hard’. I begin with integration because it is immediately applicable to real world problems which actually interest the student. Few things irritate the engineer’s mentality more than wasting vast amounts of time and effort on methods which have no apparent practical application, particularly if they intrude on the traditional engineer’s pursuits of philandering and getting drunk. Having the ability to calculate areas bounded by curves isn’t a bad addition to the engineer’s toolkit, but integration equips us with the ability to do so much more. 6.3.2 Uniform Acceleration Newton’s Second Law tells us that if we apply a force to a body, it will accelerate at a rate proportional to the magnitude of the force. The constant of proportionality is called the inertial mass of the body. This defines mass (not directly measurable) in terms of force (which is measurable) and acceleration (also measurable), rather than some circular definition involving vague and confusing statements about ‘the amount of matter’ in the body. In practice, we compare the gravitational forces (weights) acting on different bodies to determine mass. A standard lump of matter, held in some august institution, forms the basis of this comparison, and is, by international agreement, said to have the unit mass. Since weighing is much simpler than measuring force and acceleration simultaneously, and the consequent measurement is potentially much more accurate, practically all systems of units define force in terms of mass and acceleration. This tendency to view mass as a tightly bound property of the body leads us straight into conceptual difficulties when Special Relativity decrees that the mass of a body increases as it approaches the speed of light. The definition in terms of force and acceleration raises no such conceptual difficulties. However we look at it, a constant force on a body gives rise to a constant acceleration, at least provided the speeds involved are small compared with that of light, which is usually the case. 62 Cybernetics for Systems Engineers – DRAFT F Plotting a constant acceleration a having duration T results in a rectangular plot in the acceleration/time graph. Now the change in speed in a short time δt (say ‘delta tee’) is: 𝛿𝑣 = 𝑎𝛿𝑡 (read this as ‘delta vee equals a delta tee’). The change in speed over the period T is found by summing all these small changes in speed, in other words: 𝑇 𝑣 − 𝑢 = ∫0 𝑎 𝑑𝑡 Where u is the initial speed and v the final speed. Now applying the formula for integration: 𝑧 𝑎 ∫0 𝑥 𝑝 𝑑𝑥 = 𝑎 63 𝑧 𝑝+1 𝑝+1 Cybernetics for Systems Engineers – DRAFT F Figure 6-10 : Accelerating Particle In this case p=0, so the change in speed is given by: 𝑣 − 𝑢 = 𝑎𝑇 Or: 𝑣 = 𝑢 + 𝑎𝑇 The change in position in a short time interval depends on the speed at the time the interval is considered. From the above formula, it is evident that the speed isn’t constant but varies linearly with time. Defining v(t) as the speed at any time between 0 ant T, we have: 𝑣(𝑡) = 𝑢 + 𝑎𝑡 64 Cybernetics for Systems Engineers – DRAFT F (the (t) indicates that the quantity depends on time) This is now a function of time, rather than simply the final value. Plotting v(t) against time yields a straight line. Denoting the distance travelled by s, we have: 𝑡 1 𝑠(𝑡) − 𝑠(0) = ∫0 (𝑢 + 𝑎𝑡) 𝑑𝑡 = 𝑢𝑡 + 2 𝑎𝑡 2 If the upper limit of the integral is the independent variable, the integral becomes a function of the independent variable (in this case time). If it is a constant, the integral evaluates to a number, such as the area under the curve. By starting from the acceleration, and applying the integration formula twice, we have derived expressions for speed and position as functions of time. 6.3.3 Volumes of revolution Figure 6-11: Calculation of the Volume of a Body of Revolution If we take a curve and rotate it about the x-axis we generate a surface of a three dimensional body. The curve is called the generator of the surface. For example, the straight line through the origin: 𝑦 = 𝑎𝑥 will generate a conical surface. Taking thin slices through the body of thickness δx (this is just another way of writing w for the width of strips; it’s the same thing only consistent with accepted calculus jargon), the volume of an individual slice is given by: 𝛿𝑉 = 𝜋𝑦 2 𝛿𝑥 So the volume between the vertical planes passing through x1 and x2 is given by adding all these volumes together: 65 Cybernetics for Systems Engineers – DRAFT F 𝑥 𝑉 = 𝜋 ∫𝑥 2 𝑦 2 𝑑𝑥 1 The straight line gives us the volume of a cone of height h: ℎ 𝜋 𝑉 = 𝜋 ∫0 𝑎2 𝑥 2 𝑑𝑥 = 3 𝑎2 ℎ3 But a×h is equal to the radius of the base (substitute h for x in the equation for the line to see this). The volume of the cone is therefore given by: 1 𝑉 = 3 𝜋𝑟 2 ℎ Or, one-third the base area times the height. Another useful example is the sphere: 𝑦2 = 𝑟2 − 𝑥2 For the hemisphere x=0 to x=r, the volume is: 𝑟 𝑉 = 𝜋 ∫0 (𝑟 2 − 𝑥 2 )𝑑𝑥 = 𝜋 [𝑟𝑥 − 𝑟 𝑥3 ] 3 0 2 3 = 𝜋𝑟 3 Note the use of square brackets to evaluate the integral, before substituting the upper and lower bounds of the integral into it. The upper and lower bounds are usually called the limits of integration. The volume of the sphere is obviously twice this value: 4 3 𝑉 = 𝜋𝑟 3 If we subtract from this the volume of a spherical void of radius t units less than the original sphere, the resulting spherical shell has volume: 4 4 𝑉 = 3 𝜋(𝑟 3 − (𝑟 − 𝑡)3 ) = 3 𝜋(3𝑟 2 𝑡 − 3𝑟𝑡 2 + 𝑡 3 ) If the thickness is small compared with the radius, the second order and higher terms in t may be ignored. Also, as the thickness tends to zero, the volume tends to the surface area S multiplied by the thickness: 𝑆𝑡 = 4𝜋𝑟 2 𝑡 So that as t→0, the surface area becomes: 𝑆 = 4𝜋𝑟 2 Note; this method of finding the surface area only works with shapes like the sphere for which the thickness of the shell is uniform, and thickness is measured perpendicular to the surface. In the case of a cone, the thickness of the shell will be t if the inner cone is displaced along the x axis by an amount: 66 Cybernetics for Systems Engineers – DRAFT F 𝑡 𝑑 = √𝑟 2 + ℎ2 𝑟 This void has volume: 𝜋 𝑉 = 3 𝑎2 (ℎ − 𝑑)3 The volume of the shell is,: 𝜋 𝑆𝑡 ≈ 3 𝑎2 (ℎ3 − ℎ3 + 3ℎ2 𝑑 − 3ℎ𝑑2 + 𝑑3 ) Ignoring high order terms in d: 𝑆𝑡 ≈ 𝜋𝑟 2 𝑑 = 𝜋𝑟√𝑟 2 + ℎ2 𝑡 As t→0, the surface area of the cone, ignoring the base, becomes: 𝑆 = 𝜋𝑟√𝑟 2 + ℎ2 Note the approximation becomes more accurate, the smaller t becomes. When t is zero, the expression is exact. 6.4 Centres of Gravity Figure 6-12 : Definition of Centre of Mass The centre of gravity of a body is the point through which its weight appears to act. For the sake of clarity, we restrict the discussion to solid bodies of uniform density throughout so that mass is directly proportional to volume. The turning moment about the origin of a body of volume V is equal to: 𝑀 = 𝜌𝑉𝑥̅ 67 Cybernetics for Systems Engineers – DRAFT F Where ρ (rho) is the weight per unit volume of the material of the body and 𝑥̅ is the distance of the centre of gravity from the origin. The turning moment of a transverse section of the body at distance x from the origin of thickness δx is given by: 𝛿𝑀 = 𝜌𝐴𝑥𝛿𝑥 Where A is the area of the section. The total turning moment is found by summing over all these sections, so that: 𝐿 𝜌𝑉𝑥̅ = 𝜌 ∫0 𝐴(𝑥)𝑥 𝑑𝑥 L is the position of the point on the body furthest from the origin. The density cancels out, and hence is irrelevant to the calculation. The cross section is a function of the distance from the origin. The centre of gravity is therefore found from: 𝐿 𝑥̅ = ∫0 𝐴(𝑥)𝑥 𝑑𝑥 𝑉 As an example, a cone cross sectional area, is presented in the previous section as: 𝐴(𝑥) = 𝜋𝑎2 𝑥 2 𝜋 The volume was found to be 3 𝑎2 ℎ3 where h is the height of the cone. ℎ ℎ 1 ∫0 𝐴(𝑥)𝑥 𝑑𝑥 = ∫0 𝜋 𝑎2 𝑥 3 𝑑𝑥 = 4 𝜋𝑎2 ℎ4 So the centre of gravity is: 3 𝑥̅ = 4 ℎ Below the apex, or a quarter of the height from the base. For a hemisphere, it has been shown that: 2 𝑉 = 3 𝜋𝑟 3 The cross section is given by: 𝐴(𝑥) = 𝜋(𝑟 2 − 𝑥 2 ) The moment of area is: 𝑟 ∫0 𝜋 (𝑟 2 𝑥 − 𝑥 3 ) 𝑑𝑥 = 𝜋 [ The position of the centre of gravity is, therefore: 3 8 𝑥̅ = 𝑟 68 𝑟2𝑥 2 2 − 𝑟 𝑥4 ] 4 0 𝜋 4 = 𝑟4 Cybernetics for Systems Engineers – DRAFT F 6.5 Second Moment of Area The centre of gravity is found by calculating the first moment of area with respect to a reference point. If the reference point is the centre of gravity, a force applied there will cause the body to move in translation only, and will not cause the body to rotate. In general, an applied force cannot be expected to act through the centre of gravity, and consequently it applies a turning moment to the body. The moment is resisted by an inertial torque proportional to the angular acceleration, analogous to the inertial resistance the body has to translational acceleration by virtue of its mass. Figure 6-13 : Moment of Inertia about an Axis Perpendicular to the Page The inertial moment arises because the angular acceleration of the body as a whole causes the individual particles of the body to accelerate linearly. The amount of linear acceleration is proportional to the distance of the particle from the axis of rotation, and the moment of the force produced by this acceleration is also proportional to the distance from the axis. The contribution of the particle to the inertial torque is equal to the angular acceleration multipled by the particle mass multiplied by the square of the distance of the particle from the axis. The inertial resistance of the whole body to the applied torque is found by summing the contributions of all the particles of the body. This results in a quantity which depends on the mass and shape of the body which characterises its inertial resistance to angular acceleration. This quantity is called the moment of inertia with respect to the axis of rotation. It is also known as the second moment of area. The second moment of area also characterises the ability of a beam cross-section to withstand bending moments, with the fibres furthest from the centre of the section contributing most to the bending stiffness, and those nearest the centre contributing least. Second moments of area also arise in the analysis of ship stability, as the parts of the hull furthest from the centre line contribute most to the righting moment when the ship heels over. 69 Cybernetics for Systems Engineers – DRAFT F All of these applications, and doubtless many more, require us to calculate the product of infinitesimal areas with the square of their distance from the axis of interest, and sum them over the entire body. An example may clarify this concept. Consider a particle of mass m mounted on a rigid bar, length r, of negligible mass, pivoted to rotate freely about its other end, the pivot itself is prevented from translational motion, so maintains its position fixed at all times . The frictional torque in the pivot is considered negligible. The assembly experiences a constant angular acceleration of α (alpha)radians/sec/sec, and we need to find the magnitude of the torque needed to sustain the motion. The instantaneous translational acceleration of the particle is αr, so the inertial force, by Newton’s Second Law is mαr, and the moment of this force about the pivot is mr2α. By analogy with Newton’s Second Law, in which force equals mass times acceleration, we say that the torque equals the moment of inertia times the angular acceleration. The moment of inertia of a single particle offset from the axis of rotation is therefore: 𝐼 = 𝑚𝑟 2 Note that although we are dealing with rotational motion, we derive the torque by considering only translational motion of the particle. If the spinning of the particle introduced any new effects, they would not be predictable from the standard equations of motion for a rigid body. The moment of inertia is to angular motion what mass is to translational motion. 6.5.1 Moment of inertia of a Disc Figure 6-14 : Moment of Inertia of a Uniform Solid Disc 70 Cybernetics for Systems Engineers – DRAFT F We shall consider the moment of inertia of a circular disc, radius R and thickness t, rotating about an axis perpendicular to the plane of the paper. The mass of a ring of radius r and width δr is given by: 𝛿𝑚 = 𝜌2𝜋𝑟𝑡 𝛿𝑟 Where ρ (rho) is the density of the material. Note that the δ (delta) in front of the symbol implies the quantity is ‘small’ in this context, and we can anticipate that the problem reduces to finding out what happens when such quantities approach zero. All the mass of this ring is distance r from the axis of rotation, so it will have moment of inertia: 𝛿𝐼 = 𝜌2𝜋𝑟 3 𝑡𝛿𝑟 So the moment of inertia of the disc is found by summing over an infinite number of concentric rings, whose widths tend to zero. 𝑅 𝐼 = 𝜌𝑡 (2𝜋 ∫0 𝑟 3 𝑑𝑟) The term in brackets is the second moment of area. Evaluating the integral: 𝐼 = 𝜌𝑡𝜋 𝑅4 2 But the mass of the disc is 𝜌𝑡𝜋𝑅 2 , so the moment of inertia may be written: 1 𝐼 = 2 𝑚𝑅 2 The disc behaves like a ring which has all its mass concentrated at a radius: 𝑟𝑔 = 𝑅 √2 Where rg is known as the radius of gyration. This is a useful concept, because it is much easier to measure the radius of gyration of an arbitrarily shaped body than it is to estimate it by calculation. With practice it becomes possible to estimate the radius of gyration by eye. 6.6 Engineer’s Bending Theory If we consider a section of beam loaded so that it bends into a curve, we see that for the case shown, the lower surface of the beam is in tension, while the upper surface is in compression. If the radius of curvature of the deformation is large compared with the thickness of the beam, it is reasonable to assume that the deformation varies linearly across the depth of the beam. From Hooke’s Law, the stress at any point in the cross section is proportional to the strain, i.e. the deformation. It follows that the stress also varies linearly across the section. The elastic moment resisting the bending moment due to the applied load may be found by summing the moments due to the internal stresses across the depth of the beam. 71 Cybernetics for Systems Engineers – DRAFT F Consider a rectangular beam of breadth b and depth d. For consistency with the rest of this chapter we will take the vertical direction through the beam cross section as the x-axis, so the section is on its side, with the moment applied in the horizontal plane. The origin is taken at the centre of the beam section, because this is where the stress is believed to be zero. The elastic force on a narrow strip of width (actually height) δx is given by: 𝛿𝐹 = 𝜎(𝑥)𝑏 𝛿𝑥 Where σ is the stress at this depth (stress is force per unit area, as the size of the area tends to zero). This will generate a moment about the y axis of: 𝛿𝑀 = 𝑥𝛿𝐹 In order to proceed, we need an expression for the distribution of stress. Figure 6-15 : Small Angle Approximations The length of beam of interest is distorted into a circular arc, which has radius R at the neutral axis (i.e. where the distortion is zero). The angle subtended at the centre of the circle is θ (theta) radians. The length of a section distance x below the neutral axis is: 𝑙(𝑥) = (𝑅 + 𝑥)𝜃 So the strain ε (epsilon) on this section is: 𝜀= (𝑅+𝑥)𝜃−𝑅𝜃 𝑅𝜃 = 𝑥 𝑅 According to Hooke’s Law, the stress is proportional to the strain: 72 Cybernetics for Systems Engineers – DRAFT F 𝜎 = 𝜀𝐸 Where E is called the modulus of elasticity of the material, or Young’s modulus. Figure 6-16 : Beam Bends Locally into a Circular Arc, with Linear Stress Distribution Across the Depth The moment of the elemental strip is, therefore: 𝐸 𝛿𝑀 = 𝑅 𝑏𝑥 2 𝛿𝑥 So the moment is found by summing over all strips and taking the width of the strips to zero. 𝑏 𝑑 𝑀= 𝐸 𝑏 ∫2𝑑 𝑥 2 𝑑𝑥 𝑅 − = 2 𝐸 𝑥3 2 𝑏[3] 𝑏 𝑅 − 2 𝐸 𝑏𝑑 3 = 𝑅 ( 12 ) The term in brackets is the second moment of area, usually written I. The formula for the bending moment is given by: 𝑀 𝐼 𝐸 =𝑅 The maximum stress is at the top or bottom of the beam, which is taken as a distance y from the neutral axis. The maximum stress is given by: 𝐸 𝜎𝑚𝑎𝑥 = 𝐸𝜀𝑚𝑎𝑥 = 𝑅 𝑦 The complete formula for engineer’s bending theory, becomes: 𝜎 𝑦 = 𝑀 𝐼 𝐸 =𝑅 Bearing in mind the fact that σ is the Greek letter s, this lends itself to the rude, but highly memorable mnemonic; ‘Screw Your Mother In the Engine Room’. 73 Cybernetics for Systems Engineers – DRAFT F 6.7 Ship Stability Figure 6-17 : Righting Moment for a Ship Heeled Over If a ship heels over, one side becomes immersed more deeply in the water, whilst the other side is lifted out of the water. By Archimedes Principle, the immersed side experiences an increased upthrust whilst the weight of the ship on the other side is no longer completely supported by buoyancy forces. For a symmetrical waterline planform, The extra up thrust at a point x from the longitudinal axis is the same as the excess weight on the other side, and the ship experiences a righting moment. The magnitude of the righting moment is calculated by summing the righting moments due to an infinite number of vanishingly small longitudinal slices. As an illustrative example, consider a waterline section consisting of a parabolic arc. The equation for this curve is: 𝑥= 𝑤 (1 2 2𝑦 2 −(𝐿) ) Where w is the width, and L the length of the waterline section. This introduces a difficulty because so far we have only dealt with integer powers of x. This equation evidently introduces a square root. The ship is heeled over at an angle φ (phi) so that a longitudinal slice distance x from the centre line is immersed to a depth φx. The volume of water displaced is φxyδx, so the upthrust is φρxyδx, where ρ is the water density. The righting moment due to the infinitesimal longitudinal slice and the corresponding slice which is lifted out of the water on the other side is: 𝛿𝑀 = −2𝜌𝜙𝑥 2 𝑦𝛿𝑥 74 Cybernetics for Systems Engineers – DRAFT F The negative sign indicates that the righting moment is in the opposite direction to the angle of heel. The total righting moment is: 𝑤 𝑀 = −4𝜌𝜙 ∫02 𝑥 2 𝑦𝑑𝑥 The integral is the second moment of area of the waterline section about the centre line. So that in general the righting moment is given by: 𝑀 = 𝜌𝜙𝐼 Where I is the second moment of area of the waterline section. In the example, i ntegrating with respect to x will introduce fractional powers of x, which we have yet to consider . In order to avoid the problem, we shall integrate with respect to y. We have for the waterline section: 𝑥= 𝑤 (1 2 2𝑦 2 𝐿 −( ) ) We found from the analysis of the rectangular beam cross section that the second moment of area of a rectangle about an axis through its centroid is: 𝐼𝑟𝑒𝑐𝑡𝑎𝑛𝑔𝑙𝑒 = 𝑏𝑑 3 12 The second moment of area for a transverse slice of the waterline section, of length δy is, therefore: 𝛿𝐼 = 75 (2𝑥)3 12 𝛿𝑦 = 3 𝑤3 2𝑦 2 (1 − ( ) ) 𝛿𝑦 12 𝐿 Cybernetics for Systems Engineers – DRAFT F So I is found by summing over an infinite number of infinitesimal slices: 𝐿 𝐼= 𝑤3 2 ∫ (1 − 12 −𝐿 2 2𝑦 2 2𝑦 4 2𝑦 6 3 ( 𝐿 ) + 3 ( 𝐿 ) − ( 𝐿 ) ) 𝑑𝑦 From which the second moment of area is: 4 𝐼 = 105 𝑤 3 𝐿 So the wider the ship, the larger the righting moment. But the wider the ship, the more power is required to propel it. The astute reader will probably have noticed that the second moment of area consists of a numerical factor multiplied by the area of the enclosing rectangle, multiplied by the square of the dimension perpendicular to the axis. It is awareness of the expected size of the numerical factor that forms part of the engineer’s intuition as to whether a design ‘looks right’. The excessive influence of academic mathematicians and commercial numerical analysis packages has largely supplanted the true engineer’s intuitive grasp of the problem. I am unrepentant in trying to bring it back. In order to float at all, the average density of the ship must be much less than that of the water, so that the centre of gravity is expected to lie above the waterline. When the ship heels over, the moment of the weight about the centre line of the waterline section tends to cause the ship to heel further, so the moment preventing capsizing is the difference between the hydrostatic righting moment and the moment of weight: 𝑀𝑐 = −(𝜌𝐼 − 𝑊ℎ)𝜙 Where W is the displacement and h the height of the centre of gravity above the waterline. The roll stability is usually characterised by considering the rolling moment to arise completely from the weight of the ship suspended distance hm below the waterline, where: 𝑊ℎ𝑚 = 𝜌𝐼 − 𝑊ℎ This distance is called the height of the metacentre, and unlike the quantities in the original roll equation, it is readily measured by forcing the ship to heel over by adding offset loads and then measuring the resulting angle of heel. However, requiring the ship to be built before its stability can be assessed is hardly satisfactory. Some means is required to estimate the metacentric height long before construction begins. 6.8 Fractional Powers The ship stability problem introduced a requirement to integrate fractional powers of x, which were overcome by integrating the expression for second moment of area with respect to y. In effect, the second moments of area of infinitesimal lateral slices of the ship were taken, and then summed to get the second moment of area of the entire ship. 76 Cybernetics for Systems Engineers – DRAFT F Recall that the area under a parabola 𝑦 = 𝑥 2 is one third the area of the rectangle containing it. The curve 𝑥 = 𝑦 2 is also a parabola, but it is the area between the curve and the y-axis which is one third the area of the enclosing rectangle, the area below the curve must therefore be equal to the remaining two thirds of the enclosing rectangle. 3 1 The enclosing rectangle has area x×y, or 𝑥√𝑥 = 𝑥 2 , so the area under the curve √𝑥 = 𝑥 2 is: 𝑧 1 2 3 ∫0 𝑥 2 𝑑𝑥 = 3 𝑧 2 This result is consistent with the general formula for positive integer powers of x. However, this could be coincidence. In general, the area of the enclosing rectangle is xy, and this must be equal to the integral of the function with respect to y plus the integral of the same function with respect to x. 𝑎 𝑏 ∫0 𝑥 𝑑𝑦 = 𝑎𝑏 − ∫0 𝑦𝑑𝑥 So, if x is raised to a power which is a reciprocal of an integer, the integral can always be evaluated because it will be an integer power of y. The general formula applies to powers which are reciprocals of positive integers. What about more general fractions? Suppose we wish to evaluate: 𝑏 ∫𝑎 𝑥 𝑛 𝑦𝑑𝑥 When 𝑥 = 𝑦𝑚 Both n and m are integers. This is the same as evaluating: 𝑏 1 𝑛+ ∫𝑎 𝑥 𝑚 𝑑𝑥 We solved this in the ship stability example by integrating with respect to y. If x is increased by an amount, δx, then in order to stay on the curve y must also change by an amount δy, such that: 𝑥 + 𝛿𝑥 = (𝑦 + 𝛿𝑦)𝑚 We have found that all terms higher that the first in infinitesimal quantities become zero as we sum over an infinite number of them. So we need only to expand the polynomial in y to the first order term in δy: 𝑥 + 𝛿𝑥 ≈ 𝑦 𝑚 + 𝑚𝑦 𝑚−1 𝛿𝑦 Eliminating x from both sides: 77 Cybernetics for Systems Engineers – DRAFT F 𝛿𝑥 = 𝑚𝑦 𝑚−1 𝛿𝑦 The integral becomes: 𝑑 𝑚 ∫𝑐 𝑥 𝑛 𝑦 𝑚 𝑑𝑦 Where the limits are the values of y corresponding to the limit values for x i.e.: 1 1 𝑐 = 𝑎𝑚 , 𝑑 = 𝑏𝑚 Writing the integral entirely in terms of y: 𝑑 𝑚 𝑚 ∫𝑐 𝑦 𝑚(𝑛+1) 𝑑𝑦 = [(𝑚(𝑛+1)+1) 𝑦 𝑚(𝑛+1)+1 ] 𝑑 𝑐 Substituting for x: 𝑏 𝑛+ 1 ∫𝑎 𝑥 𝑚 𝑑𝑥 = [ 1 1 𝑚 𝑛+ +1 𝑥 1 𝑚 𝑛+ +1 𝑏 ] 𝑎 This is the same result as would have been obtained by applying the general formula to the original integral. 1 𝑚 The general formula therefore applies to powers which are improper fractions of the form (𝑛 + ) This implies that square roots, cube roots, etc may be integrated any number of times using the standard formula. The general fractional index is: 𝑛 𝑦 = 𝑥𝑚 Let 1 𝑧 = 𝑥𝑚 The increment in z corresponding to the increment δx is found from the first order expansion of (𝑧 + 𝛿𝑧)𝑚 : 𝑧 + 𝑚𝑧 𝑚−1 𝛿𝑧 ≈ 𝛿𝑥 Hence: 𝑏 𝑑 𝑚 ∫𝑎 𝑦𝑑𝑥 = 𝑚 ∫𝑐 𝑧 𝑛+𝑚−1 𝑑𝑧 = 𝑛+𝑚 [𝑧 𝑛+𝑚 ]𝑑𝑐 = 1 𝑛 +1 𝑚 𝑛 [𝑥 𝑚+1 ] 𝑏 𝑎 This is the same result as would be obtained from the general formula, so it is concluded that the formula derived for positive integers applies to any positive power of x. 78 Cybernetics for Systems Engineers – DRAFT F Hopefully, by now the reader is no longer daunted by the sight of an integral sign, and sufficient examples have been presented to indicate the utility of the technique of integration to a wide range of problems which would otherwise be intractable. All this capability is at the reader’s fingertips by applying a little ingenuity and the formula; 𝑥 𝑝+1 𝑏 ∫𝑎 𝑥 𝑝 𝑑𝑥 = [ 𝑝+1 ] 𝑏 𝑎 6.9 The Usual Starting Point 6.9.1 Introduction We have introduced a method of dealing with problems which involve continuously changing quantities and demonstrated its utility with a number of examples. The more curious reader may wonder if the process of integration has an inverse process. For example, if we have an expression for distance travelled as a function of time, can we derive the speed at any time, and the acceleration from it? The simple answer is yes, the inverse process exists. It is called differentiation, and is the more usual starting point for texts on calculus. However, it is hard to find examples to motivate interest in the subject matter, which are not clearly contrived for the purpose, and it introduces the concept of a limit which is inherently more difficult to grasp than the idea of a sum over a huge number of tiny strips. It is for these reasons,(i.e. that the student is presented with a difficult to understand concept, and has very little to indicate that there is any benefit in coming to grips with it anyway) that I have chosen to introduce the subject with integration. The standard notation for differentiation is illustrated as follows: If: 𝑏 ∫𝑎 𝑧 𝑑𝑥 = [𝑦]𝑏𝑎 Then: 𝑑𝑦 𝑑𝑥 =𝑧 𝑑𝑦 Where 𝑑𝑥 is called the derivative of y with respect to x. From the formula for integration, it is evident that if: 𝑦 = 𝑥𝑝 Then: 𝑑𝑦 𝑑𝑥 = 𝑝𝑥 𝑝−1 That is all very well, but what does this mean in terms of functions plotted on a graph? 79 Cybernetics for Systems Engineers – DRAFT F We start from the simplest of plots, the straight line: 𝑦 = 𝑚𝑥 + 𝑐 Applying the formula for differentiation: 𝑑𝑦 𝑑𝑥 =𝑚 This is the slope of the line. If the plot is a curve, its slope will vary continuously. Differentiation yields the slope of the tangent at the point of interest. We can calculate an approximate value for the slope at x by incrementing x slightly (by δx) and finding the corresponding increment in y (δy). Dividing the small change in y by the small change in x yields an estimate of the slope of the curve at x. Consider: 𝑦 = 𝑥𝑛 Increment x slightly: 𝑦 + 𝛿𝑦 = (𝑥 + 𝛿𝑥)𝑛 We recall, from summing an infinite number of strips that: (𝑥 + 𝛿𝑥)𝑛 = 𝑥 𝑛 + 𝑛𝑥 𝑛−1 𝛿𝑥 + ⋯ The remaining terms contain squares and higher powers of the increment , so we can anticpate that when we reduce the increment to zero, these terms disappear. The y and xn terms on either side of the expression are equal and so cancel out, leaving an expression for δy: 𝛿𝑦 = 𝑛𝑥 𝑛−1 𝛿𝑥 + ⋯ The slope is given approximately by: 𝛿𝑦 𝛿𝑥 = 𝑛𝑥 𝑛−1 + 𝑡𝑒𝑟𝑚𝑠 𝑖𝑛 𝛿𝑥 0 Now, if δx=0, so does δy, and we are left with the indeterminate result 0. This is the conceptual obstacle we must overcome. It was evident in the case of integration that the more and finer the strips, the more nearly they filled the area, so that with an infinite number, the formula produced became an exact answer, rather than a very close approximation. 80 Cybernetics for Systems Engineers – DRAFT F Figure 6-18 : Differentiation finds the Tangent to the Curve What we say is that as we make δx arbitrarily small, the estimate of the slope becomes ever closer to 𝑑𝑦 nxn-1. Under these conditions we write the ratio of two small but finite changes in y and x as 𝑑𝑥, rather than 𝛿𝑦 . 𝛿𝑥 This is written as: 𝑑𝑦 𝑑𝑥 𝛿𝑦 𝛿𝑥→0 𝛿𝑥 = lim Read this as ‘the limit as δx tends to zero’. ‘lim’ denotes the limiting value where the ratio still remains finite. Just as taking the increment to zero in the case of integration yielded an exact answer, taking it to zero when differentiating yields an exact answer for the slope. To the critical mind this smacks a bit of Emperor’s New Clothes, which is not the best way to introduce a subject which has no immediately obvious applications. I suggest that this is the reason calculus is popularly considered ‘hard’. Few engineers have much patience with arcane discussions about the number of angels which can dance on a pinhead. So whereas integration calculates areas under curves, differentiation finds the slopes of tangents to curves. 6.10 Applications 6.10.1 Newton’s Method Quite often the Laws of physics present us with equations which are not amenable to elementary methods, such as the formula for a quadratic equation. The problem is to find a solution to an equation which takes the form: 𝑓(𝑥) = 0 81 Cybernetics for Systems Engineers – DRAFT F This is the intersection of the function: 𝑦 = 𝑓(𝑥) And the x axis (y=0). Figure 6-19 : Successive Approximations If we start at a rough solution, found for example, by sketching the curve, we can take the tangent to the curve as a straight line approximation in the vicinity of the solution. This straight line fit may be used to get another, hopefully better, approximation to the required solution, although it is not guaranteed that the method will always converge on to the solution. Given an approximate solution xi, the tangent intersects the x-axis at the next iteration xi+1. From the geometry of the curve: 𝑦(𝑥𝑖 ) 𝑥𝑖+1 = 𝑥𝑖 − 𝑑𝑦 𝑑𝑥 𝑥=𝑥 𝑖 Note if we are dealing with the slope at a particular point, rather than deriving the general formula for the slope, it is usual to indicate the fact with a subscript, as shown. As an example, suppose we are seeking a root of the cubic equation: 𝑥 3 + 3𝑥 2 − 4𝑥 + 2 = 𝑦 = 0 The derivative is found by applying the formula for differentiating powers of x term by term: 𝑑𝑦 𝑑𝑥 = 3𝑥 2 + 6𝑥 − 4 Starting with the approximation x=-5.0, the progress of the algorithm can be seen in the following table. 82 Cybernetics for Systems Engineers – DRAFT F Iteration x Y 1 -5.0000 -28.0000 2 -4.3171 -5.2782 3 -4.1141 -0.4015 4 -4.0960 -0.0031 5 -4.0958 0.0000 6 -4.0958 0.0000 In this case the solution is found to an accuracy of four decimal places in five iterations. 6.10.2 Maxima and Minima If we plot a polynomial of degree higher than one, we may notice that in a particular region, y increases with increasing x but further on, y decreases with increasing x, or the slope starts positive and ends up negative. Alternatively, the slope may start negative and become positive. Figure 6-20 : Maxima and Minima Correspond to Turning Points At some point the slope is zero, and this corresponds to a point on the curve where the y has its maximum value (in the first case) or its minimum value (in the second case). So if: 𝑦 = 𝑓(𝑥) These turning points are found where: 83 Cybernetics for Systems Engineers – DRAFT F 𝑑𝑦 𝑑𝑥 =0 This doesn’t tell us what type of turning point it is (maximum or minimum), or whether the turning point is an absolute maximum (minimum) of the function, i.e. a supremum (infimum), when the function has many turning points. A quick sketch of the function in the region of interest will resolve these issues, but we shall see how to distinguish maxima from minima purely analytically, presently. 6.10.3 Body under Gravity Recall that we derived the expression for distance moved under constant acceleration as an illustration of the integration process: 1 2 𝑥 = 𝑢𝑡 + 𝑎𝑡 2 In this case the independent variable t is time. A body which is sufficiently dense is not expected to be influenced by air resistance, especially if the speeds are low, so this equation is a reasonable approximation to the motion of say a cricket ball thrown vertically upwards. The ball is subjected to its weight, which manifests itself as a constant acceleration towards the Earth this acceleration is conventionally denoted g. We shall take the positive direction as upwards, so the gravitational acceleration in the above equation will be denoted by –g. The height h of the ball above the release point is then given by: 1 ℎ = 𝑢𝑡 − 2 𝑔𝑡 2 where u is the speed the ball is thrown upwards with. A question which comes to mind is; what is the maximum height the ball will reach? Differentiating: 𝑑ℎ 𝑑𝑡 = 𝑢 − 𝑔𝑡 So the turning point is given by: 𝑑ℎ 𝑑𝑡 𝑢 = 0, 𝑜𝑟 𝑡 = 𝑔 The maximum height is found by substituting this value of time into the equation for height: 𝑢 1 𝑢 2 1 𝑢2 𝑔 ℎ = 𝑢 (𝑔) − 2 𝑔 (𝑔) = 2 84 Cybernetics for Systems Engineers – DRAFT F 6.10.4 Maximum Area Figure 6-21: Find Proportions for Maximum Area Suppose we have a boundary of fixed length, say L, and we wish to find out the proportions of the rectangle which will enclose the maximum area. Let one side of the rectangle be x, so the adjacent side d must be: 𝑑= 𝐿−2𝑥 2 The area of the rectangle is length times breadth: 𝐿 𝐴 = 𝑥𝑑 = 𝑥 2 − 𝑥 2 The turning point is given by: 𝑑𝐴 𝑑𝑥 𝐿 = 0 = 2 − 2𝑥 𝐿 𝐿 From which it follows that x= 4 and d= 4 , and the rectangle which encloses the maximum area is a square. 6.10.5 Concluding Comments The calculation of turning points has obvious application in problems where we are seeking the best use of resources. The examples show that the application of calculus to ‘given’ problems does not appear particularly difficult; the hard part is invariably formulating real world problems in a form where they are amenable to mathematical analysis. This requires understanding of which terms can be ignored, 85 Cybernetics for Systems Engineers – DRAFT F which can be treated, near enough, as tractable mathematical expressions, and for what range of values of system parameters the current approximation remains adequately valid. It is this step of converting problems into a form to which the methods of mathematics can be applied which is alien to the pure mathematician, but is the essence of the mathematically-inclined engineer’s task. The resulting displacement of the engineer by the pure mathematician, even in university engineering faculties, cannot be considered a healthy development, for this reason. 6.11 Higher Order Derivatives 6.11.1 Introduction When a problem requires us to integrate an expression twice, we simply repeat the integral sign. However, most of the interesting problems requiring calculus involve dependencies on derivatives of order higher than just the first, so it is convenient to indicate the order of a derivative in the notation. The derivative of: 𝑑𝑦 𝑑𝑥 Is written: 𝑑 𝑑𝑦 ( ) 𝑑𝑥 𝑑𝑥 = 𝑑2 𝑦 𝑑𝑥 2 Differentiating n times is denoted: 𝑑𝑛𝑦 𝑑𝑥 𝑛 The following sections introduce some examples to illustrate this point. 6.11.2 Radius of Curvature Approximation If we move a small distance δs around the circumference of a circle, the tangent direction changes by an amount δθ (delta theta). This is the same angle as the radius rotates at the same time, so if δθ is in radians: 𝛿𝜃 = 𝛿𝑠 𝑟 The radius of curvature at a point on a more general curve is equal to that of a circular arc which is tangential to the curve. This circular arc is called an ’osculating curve’ , which literally means ‘kissing’, conveying the notion that the two curves are intimately related. The reciprocal of the radius of curvature is called simply the ‘curvature’ and is given by taking the limit of: 𝑑𝜃 𝑑𝑠 86 𝛿𝜃 𝛿𝑠→0 𝛿𝑠 = lim 1 =𝑟 Cybernetics for Systems Engineers – DRAFT F In order to use this relationship for general curves, we need to explore more functions than positive powers of x. However, if the slope 𝑑𝑦 𝑑𝑥 is small in the region of interest (e.g. less than 5 ̊) we may make the approximations: 𝛿𝑠 ≈ 𝛿𝑥 𝑑𝑦 𝛿𝜃 ≈ 𝑑𝑥 So the curvature is, approximately: 1 𝑟 ≈ 𝑑2 𝑦 𝑑𝑥 2 Near a turning point, the slope is near zero, so this approximation is valid. Now positive curvature implies the slope increases with x, so the turning point is a minimum. Negative curvature implies the slope decreases (i.e. becomes more negative) with increasing x, so the turning point must be a maximum. Having found a turning point by differentiating the function and equating the resulting expression to zero, we find out if it is a maximum or minimum by differentiating again and substitute the value of x at the turning point. If the result is negative, the turning point is a maximum, if positive, it is a minimum. Applying this to the cricket ball example: 𝑑ℎ 𝑑𝑡 = 𝑢 − 𝑔𝑡 𝑑2 ℎ 𝑑𝑡 2 = −𝑔 This is negative, so the solution does indeed correspond to a maximum value for h. Another circumstance in which the slope is expected to be small is the lateral deflection of a loaded beam. From the formula for engineer’s bending theory, we have for the curvature: 1 𝑟 𝑑2 𝑦 1 ≈ 𝑑𝑥 2 = 𝐸𝐼 𝑀 If we have an expression for bending moment as a function of distance along the beam, we can solve for the deflection. As an example consider a beam of length L, rigidly mounted at one end and loaded with weight W at its free end. The bending moment is WL at the root, reducing linearly to zero at the tip: 𝑀 = 𝑊(𝐿 − 𝑥) The deflection pattern may be found by integrating the bending equation twice. 87 Cybernetics for Systems Engineers – DRAFT F 𝑑𝑦 𝑑𝑥 − 𝑑𝑦 𝑑𝑥 𝑥=0 = 𝑊 [𝐿𝑥 𝐸𝐼 − 𝑥 𝑥2 ] 2 0 Figure 6-22 : Loaded Cantilever The slope at the root is zero, so the equation for the slope of the beam is: 𝑑𝑦 𝑑𝑥 = 𝑊𝐿2 𝑥 ( 𝐸𝐼 𝐿 1 𝑥 2 − 2 (𝐿 ) ) The deflection is: 𝑦 − 𝑦(0) = 𝑊𝐿2 1 𝑥 2 [ 𝐸𝐼 2 𝐿 1 𝑥3 − 6 𝐿2 ] 𝑥 0 The deflection at the tip is: 𝑦(𝐿) = 𝑊𝐿3 3𝐸𝐼 6.12 More Exotic Functions So far we have only been able to work with functions which are linear sums of positive powers of the independent variable, typically x. 6.12.1.1 Negative Powers The most obvious gap in the available set of functions is negative powers of x. If: 𝑦 = 𝑥 −𝑛 𝑥𝑛𝑦 = 1 Taking increments in x and y: 88 Cybernetics for Systems Engineers – DRAFT F (𝑥 + 𝛿𝑥)𝑛 (𝑦 + 𝛿𝑦) = 1 Expanding to first order of small quantities: 𝑥 𝑛 𝑦 + 𝑛𝑦𝑥 𝑛−1 𝛿𝑥 + 𝑥 𝑛 𝛿𝑦 = 1 Collecting up terms: 𝑥 𝑛 𝛿𝑦 = −𝑛𝑥 −1 𝛿𝑥 Taking this to the limit as x→0: 𝑑𝑦 𝑑𝑥 = −𝑛𝑥 −𝑛−1 This is precisely the same equation as for positive powers, which implies the integration formula works for negative powers. The formula is: 𝑏 ∫𝑎 𝑥 𝑝 𝑑𝑥 = 𝑥 𝑝+1 𝑝+1 If p=-1, the formula yields a value of infinity, so that: 𝑏1 ∫𝑎 𝑥 𝑑𝑥 Is a special case. 6.12.2 The Integral of the Reciprocal Function It appears our formula for integration works for every power of x apart from -1. We could hope that this particular integral doesn’t crop up very often in practice. Unfortunately, that is a forlorn hope, as the most important phenomena of all are governed by equations of his form. 6.12.2.1 The Tsiolovski Rocket Equation Take for example, the acceleration of a rocket. Matter is accelerated in the rocket motor, and ejected out of the nozzle, the force accelerating the exhaust matter also acts on the rocket itself, according to Newton’s Third Law. If we consider an element of the exhaust material of mass δm, accelerated to the exhaust velocity ue, it will impart an impulse of magnitude: 𝛿𝐼 = 𝑢𝑒 𝛿𝑚 To the rocket. An impulse is a force acting for a very short length of time, so the continuous force acting on the rocket is: 𝑇= 𝑑𝐼 𝑑𝑡 = 𝑢𝑒 𝑑𝑚 𝑑𝑡 Where T denotes thrust. The matter which is ejected originated from the propellant on board the rocket, so the rocket is losing mass at a rate given by this expression. According to Newton’s Second Law, the rocket will accelerate at a rate proportional to the thrust: 89 Cybernetics for Systems Engineers – DRAFT F 𝑚 𝑑𝑢 𝑑𝑡 = 𝑢𝑒 𝑑𝑚 𝑑𝑡 The change in speed over finite time is given by: 𝑚 1 𝑑𝑚 0𝑚 𝑢 − 𝑢(𝑚0 ) = 𝑢𝑒 ∫𝑚 6.12.2.2 The Breguet Range Equation The fuel consumption of a jet propelled aircraft is proportional to the thrust under a particular set of flight conditions. For the cruise, these flight conditions are sufficiently constant for the rate of fuel consumption to be directly proportional to the thrust: 𝑑𝑊 𝑑𝑡 = −𝑐𝑇 Where c is a constant, characterising the propulsion, and is called the thrust specific fuel consumption, T is thrust, W the weight of the aircraft and t is time. The aerodynamic efficiency of the aircraft is characterised by its lift to drag ratio. For steady level flight, thrust equals drag and lift equals weight: 𝑑𝑊 𝑑𝑡 𝐷 = −𝑐 ( 𝐿 ) 𝑊 In order to calculate the endurance, we need to evaluate: 1 𝐿 𝑐 𝐷 𝑊 1 𝑑𝑊 0𝑊 𝑡 = − ( ) ∫𝑊 In order to calculate range, we need the rate of fuel consumption per unit distance. 𝑑𝑊 𝑑𝑥 = 𝑑𝑊 𝑑𝑡 𝑑𝑡 𝑑𝑥 1 𝑑𝑊 𝑑𝑡 =𝑈 Where U is the cruising speed. The result is an integral of the reciprocal of weight, as for the endurance equation. Note: a fast aircraft of inferior fuel consumption and aerodynamic efficiency than a slower one, consumes less fuel because the flight time is shorter. Hence global ranges require jet aircraft, which is counter-intuitive because they are less efficient than their propeller driven alternatives. 6.13 Searching for a Solution These two examples indicate the kind of problem in which this apparently intractable integral raises its head. They are characterised by the rate of change of a quantity being proportional to the quantity itself. Any self-correcting system must have this feature, so we can anticipate that the entire edifice of control engineering must rest on this one equation. Rather than obscure, this particular equation, and the set of functions associated with it, is inconveniently ubiquitous. We proceed as before and glean what we can from summing strips of finite width. Evidently, we are courting disaster if we try to find the value of the integral from zero to an upper value, as the function becomes infinite as x approaches zero. We must therefore evaluate it between two finite limits, say a and b. 90 Cybernetics for Systems Engineers – DRAFT F The width of an individual strip is: 𝑤= (𝑏−𝑎) 𝑛 The x coordinate of the ith strip is: 𝑥𝑖 = 𝑎 + 𝑖𝑤 The area of the ith strip is: 𝑆𝑖 = 𝑤 𝑎+𝑖𝑤 𝑏 = 1 (𝑎−1) 𝑛 1+ 𝑖 (𝑏−1) 𝑛 𝑎 This cannot be written as a constant multiplying a function of i, as in previous examples. The sum depends on 𝑏 𝑎 𝑏 𝑏 as well as on i. The integral is therefore some function of 𝑎, which we shall call 𝑓 (𝑎) The integral between the two limits is also found by substituting the limits in the function and taking the difference: 𝑏1 ∫𝑎 𝑥 𝑑𝑥 = [𝑓(𝑥)]𝑏𝑎 = [𝑓(𝑏) − 𝑓(𝑎)] It follows that the function has the property: 𝑏 𝑓(𝑏) − 𝑓(𝑎) = 𝑓 (𝑎) In the dim and distant past, before the days of calculators, a function with this property would have been recognised immediately. Thankfully, the days of manual calculation are over, and we can concentrate on the more challenging task of formulating real world problems in mathematical form. Let 𝑎 = 𝑐 𝑦(𝑎) , 𝑏 = 𝑐 𝑦(𝑏) , 𝑏 𝑎 =𝑐 𝑏 𝑎 𝑦( ) Where c is a constant. Dividing the expression for b in the above list by that for a, we have: 𝑏 𝑎 = 𝑐 𝑦(𝑏) 𝑐 𝑦(𝑎) = 𝑐 (𝑦(𝑏)−𝑦(𝑎)) The function y appears to have the property we are looking for. It is the power a certain base number must be raised to in order to yield the value of the independent variable. It is called the logarithm, in this case to base c. This is written: 𝑥 = 𝑐𝑦 𝑦 = log 𝑐 𝑥 We have shown that the solution takes the form of a logarithm, but as yet we do not know the base to which it is referred. 91 Cybernetics for Systems Engineers – DRAFT F Evidently, when y=0, x=1 regardless of the base used, and when y=1, x=c, the base of the logarithm generated by the integral. 6.13.1 Finding the Magic Number. We can find the base of the logarithms by evaluating the integral numerically from x=1 to a range of values and find the value of x for which y=1. There are a number of methods which can be used. The obvious approach is to emulate our calculation of infinite sums by taking thin rectangular strips. In practice, this requires a very large number of samples. Fewer strips are needed if they fit the curve better than a rectangle. An obvious choice is to replace the rectangular strips with trapezoidal strips, which leads, believe it or not, to the trapezoidal rule: 1 𝐴 = 𝑤 (2 (𝑦1 + 𝑦𝑛 ) + ∑𝑛−1 𝑦𝑖 ) 2 This may still require too many samples for a manual calculation, but should be good enough for a computer program. Performing the calculation yields a value of about 2.72 for the base of the logarithm obtained from integrating 1/x. This base is denoted e, and like π, it is a fundamental mathematical constant. Since it has been derived from analysis, rather than an arbitrary number convention, e is referred to as the base for ‘natural’ logarithms, and the logarithm to base e is usually written: log 𝑒 𝑥 = ln 𝑥 The inverse function is then: 𝑦 = 𝑒𝑥 (Read this as ‘e to the x’). This is called the exponential function. The derivative may be found as follows: ln 𝑦 = 𝑥 1 𝑦 𝑑𝑥 = 𝑑𝑦 And since the derivative is no more than one infinitesmal quantity divided by another, it is valid to take the reciprocal of both sides: 𝑑𝑦 𝑑𝑥 = 𝑦 = 𝑒𝑥 So the exponential function has the property that it is unchanged by integration or differentiation. In addition and subtraction, the number 0 is probably the most important because it does not change the value of the answer, likewise, the number 1 has particular significance in multiplication and division, because it does not change the answer. The exponential function has a similar importance in calculus. 92 Cybernetics for Systems Engineers – DRAFT F We notice that as x increases, the rate of change of x and all higher derivatives also increase, so that the value of y increases at an ever increasing rate, this is called exponential growth. More interestingly, if x is negative, y decays at an ever reducing rate, so that it approaches the x axis but never actually reaches it. A curve of this form is called an asymptote to the x–axis. 𝑦 = 𝑒 −𝑥 ln 𝑦 = −𝑥 𝑑𝑦 𝑑𝑥 = −𝑦 This equation illustrates feedback in its simplest form. The quantity of interest, which usually needs to be kept near zero determines its own rate of change. If the feedback is positive, the quantity diverges in time without bound. If negative it tends to zero if disturbed. 6.13.2 Cooling of a Bowl of Soup As an illustration, consider the archetypal boring activity of watching soup go cold. Watching paint dry could be described in much the same way. According to Newton’s law of cooling the rate at which the temperature drops is proportional to the difference in temperature between the soup and its surroundings. Let the temperature of the surroundings be Ts and the temperature of the soup be T. Call the temperature difference θ (theta). We have: 𝜃 = 𝑇𝑠 − 𝑇 Differentiating with respect to time (t): 𝑑𝜃 𝑑𝑡 𝑑𝑇 = − 𝑑𝑡 From the law of cooling: 𝑑𝑇 𝑑𝑡 = 𝑘(𝑇𝑠 − 𝑇) Where k is a constant depending on such factors as the size and shape of the bowl, the amount of soup in it and whether the surrounding air is stationary or flowing past it. The equation becomes: 𝑑𝜃 𝑑𝑡 = −𝑘𝜃 Integrating: 𝜃 1 0𝜃 𝑑𝜃 = −𝑘𝑡 ∫𝜃 𝜃 ln (𝜃 ) = −𝑘𝑡 0 93 Cybernetics for Systems Engineers – DRAFT F 𝜃 = 𝜃0 𝑒 −𝑘𝑡 The intelligent person can find something of interest in even the most mundane of phenomena. It is the blasé fool who is bored by anything and everything. 6.13.3 Capacitor Discharge A capacitor is a device for storing electric charge. The amount of charge which can be stored depends on the voltage applied across the device terminals, and the value of the capacitor. Figure 6-23 : Capacitor/Resistor (CR) Circuit The charge is given by: 𝑞 = 𝐶𝑉 Where q is the charge, V the electrical potential difference across the terminals and C the value of the capacitance. In SI units, the charge is in coulombs, the potential difference in volts, and the capacitance in Farads. Practical values of capaicitance are usually in the nanoFarad to microfarad range. A capacitor of capacitance C is charged up to V0 volts, the switch is closed connecting the capacitor to resistor R. The current i through the resistor is given by Ohm’s Law: 𝑉 𝑖=𝑅 Where R is the value of the resistor. Current is the rate of flow of charge, so: 𝑖= 𝑑𝑞 𝑑𝑡 𝑑𝑉 = −𝐶 𝑑𝑡 The negative sign implies the charge is flowing out of the capacitor. 94 Cybernetics for Systems Engineers – DRAFT F The voltage across the capacitor once the switch is closed is governed by: 𝑑𝑉 𝑑𝑡 1 = − 𝐶𝑅 𝑉 This is of the same form as the soup temperature equation, so: 𝑡 𝑉 = 𝑉0 𝑒 −𝐶𝑅 6.13.4 Concluding Comments We have considered problems from space science, aeronautics, heat and electronics, to illustrate the point that exponential functions are to be found everywhere in nature. Many more phenomena may be modelled as approximations to exponential functions. It is the business of the engineer to find a mathematical description which is adequate for the real world problem of interest. New systems raise new and different problems, if they are expected to be superior to last years’ model, so new formulations will always be necessary which are unlikely to be available within existing applications. 6.14 Onwards and Upwards We now have formulas for differentiating and integrating a function which is a linear sum of powers of the independent variable. If we can represent a function as a polynomial, we should be able to apply our standard formula to differentiating it and integrating it. Suppose we can fit a function, near enough, to a polynomial of order n. The nth derivative will be a constant, so it will have the same value wherever we choose to evaluate it. Choose for example x=z: 𝑑𝑛𝑦 𝑑𝑥 𝑛 𝑑𝑛 𝑦 = (𝑑𝑥 𝑛 ) 𝑥=𝑧 Integrating: 𝑑 𝑛−1 𝑦 𝑑𝑥 𝑛−1 𝑑 𝑛−1 𝑦 𝑑𝑛𝑦 = (𝑑𝑥 𝑛−1 ) 𝑥=𝑧 + 𝑥 (𝑑𝑥 𝑛 ) 𝑥=𝑧 If we choose z=0, the nth derivative will not interfere with the n-1th or lower order derivatives, likewise the n-1th derivative will not interfere with the lower derivatives and so on. If we carry on integrating down to terms in x0, we end up with a polynomial approximation to the function: 𝑑𝑦 𝑦 = 𝑦(0) + (𝑑𝑥 ) 𝑥=0 1 𝑑2 𝑦 𝑥 + 2 (𝑑𝑥 2 ) 𝑥=0 1 Where the exclamation denotes a factorial; n(n-1)(n-2)...×2×1 The highest order derivative may be as high as we please. 95 𝑑𝑛𝑦 𝑥 2 ⋯ + 𝑛! (𝑑𝑥 𝑛 ) 𝑥=0 Cybernetics for Systems Engineers – DRAFT F The curve fit (known as a Taylor series expansion) requires us to differentiate the function. However, integration is usually the more difficult task, and this furnishes a method of dealing with particularly awkward functions, as the series may be integrated term by term. Recall that all derivatives of ex are equal to ex, and e0=1, the Taylor series expansion for ex becomes: 𝑒𝑥 = 1 + 𝑥 + 𝑥2 2 𝑥3 + 3×2 ⋯ + 𝑥𝑛 𝑛! The numerical value for e is found by setting x=1: 1 𝑒 = ∑∞ 0 𝑖! Or: 1 1 𝑒 =1+ + 1 1 1 + + ⋯ 2×1 3×2×1 4×3×2×1 This converges quickly on to the value 2.718, to three places of decimals. 6.15 Trigonometrical Functions 6.15.1 Introduction This section is rather long and many readers will have covered the material, I suggest referring back to it only if the material of later sections becomes incomprehensible. However, consistent with a policy of producing a self contained text which does not reference material not already covered, I shall start from first principles. After all, you paid for this, so you might as well get your money’s worth. 6.15.2 Angles From the beginning, I warned the reader that we shall measure angles using radians. If we roll a wheel a distance equal to its radius, any spoke on the wheel rotates through one radian. Alternatively, a radian is the angle subtended at the centre of a circle by an arc of length equal to the radius. Take your pick which definition you prefer; they are equivalent. However, the latter, more usual definition, smacks a bit of obfuscation for my own personal preferences. We don’t always measure an angle this way. When we specify a gradient of a hill, for example, we quote it in terms of the distance we need to travel in order to rise one unit, e.g. 1 in 20 means we must travel 20 metres in order to climb one metre. 96 Cybernetics for Systems Engineers – DRAFT F Figure 6-24 : Gradient of a Path There is usually a pointless argument as to whether the ‘distance travelled’ is the horizontal distance or the distance along the slope. Now, if the gradient were greater than one in three, we would either do something to reduce it, or choose an alternative route. There are very few, if any, practical cases where the gradient will exceed one in three. Within this limitation, if we take the distance travelled as the radius of the circle (r), there is hardly any difference between the arc length (s) and the difference in height (h). There is also hardly any difference between the horizontal distance (x) and the radius. So for typical gradients: 𝑠 𝑟 ℎ ≈𝑥≈ ℎ 𝑟 So for small angles (less than 1/3rd of a radian) the measures are, near enough, the same. Suppose now we extend the gradient approach to defining the angle such that we can apply it to larger angles. We run into two difficulties. The first is one of convention. Since at larger angles x≠r, we must specify whether the gradient refers to distance travelled along the slope, or to horizontal distance travelled. The second problem is the arc length (s) deviates considerably from h as the angle increases. The two ways of measuring gradient are evidently functions of the angle itself. Rather than give them intelligible names (e.g. call h/r ‘gradient’ and h/x ‘slope’) the world of mathematics brings wonderful clarity to the problem by calling the ratio of h/r the ‘sine’ of the angle, and ‘h/x’, the ‘tangent’ of the angle. Well, those are the names chosen for the functions, they are universally accepted, if obscure, so we have no choice but to use them here. 97 Cybernetics for Systems Engineers – DRAFT F Figure 6-25: Angles Become Large - We Need Trigonometry The sine and tangent are two functions of an angle which relate it to the ratio of two perpendicular linear measurements. They are called ‘trigonometric’ functions, I believe named after the trigon or triangular harp of the ancient Greeks, because, I guess, of their association with a right angled triangle. 6.15.3 The Full Set of Trigonometric Functions If we draw a right angled triangle, and call the angle shown θ (theta), denote the side opposite it ‘o’, the side adjacent to it as ‘a’, and the remaining side (the hypotenuse) ‘h’. From the previous discussion: The sine of the angle is written: 𝑜 sin 𝜃 = ℎ The tangent of the angle is: tan 𝜃 = 𝑜 𝑎 The inverse functions are written: tan−1 𝑥 = 𝜃 There is a third possible ratio of the sides of the triangle, which is the ratio of the adjacent to the hypotenuse, this is called the ‘cosine’ , indicating a relationship with the sine function which is not immediately obvious: 𝑎 cos 𝜃 = ℎ This is just a man-made naming convention which cannot be derived from first principles, so we have no alternative but to memorise the function names. The commonest mnemonic is to think of a 98 Cybernetics for Systems Engineers – DRAFT F South Sea island paradise and give it the name SOHCAHTOA, to remember that Sine is Opposite over Hypotenuse, Cosine is Adjacent over Hypotenuse and Tangent is Opposite over Adjacent. If we find ourselves having to memorise excessive numbers of ‘facts’ rather than derive ideas from more fundamental ideas, we must suspect that the associated body of ‘knowledge’ is man-made convention rather than principles which govern reality. As Henri Poincaré put it; ‘facts do not constitute knowledge any more than a pile of bricks constitutes a house’. Knowledge consists of understanding how facts inter-relate, so that we can predict the outcome of circumstances which have not yet been encountered. It is the closest thing we have to a practical gift of prophesy. I have always considered suspect the requirement of religious fanatics to memorise verbatim their respective holy books, without applying reason to what is actually contained in them. In addition to these three basic trigonometric functions, their reciprocals have also been given names. The reciprocal of the cosine (NOT the sine) is given the name ‘secant’, the reciprocal of the sine is called the ‘cosecant’, just to avoid any possibility of confusion, but the reciprocal of the tangent is called the ‘cotangent’. These names are needed because we have used the index -1, usually used to denote a reciprocal, to mean an inverse function in this case. These are written: 1 sec 𝜃 = cos 𝜃 1 csc 𝜃 = sin 𝜃 1 cot 𝜃 = tan 𝜃 The reciprocal of any trigonometric function is itself a defined trigonometric function. For this reason a superscript -1, is not used to signify a reciprocal, but to indicate the inverse of the function, e.g: 𝑥 = sin 𝜃 Implies: 𝜃 = sin−1 𝑥 and similarly for the inverses of all the other trigonometric functions. An alternative notation attaches the prefix ‘arc’ to the function name: If: 𝑥 = sin 𝜃 Then: 𝜃 = 𝑎𝑟𝑐 sin 𝑥 This latter notation, despite being less ambiguous, is not as common as the superscript -1 notation. 99 Cybernetics for Systems Engineers – DRAFT F 6.15.4 Relationhips Between Trigonometric Functions If we take the hypotenuse of the triangle to be unit length, it follows that the opposite side will have length sinθ and the adjacent will have length cosθ. We have immediately: sin 𝜃 tan 𝜃 = cos 𝜃 I have assumed the reader is familiar with Pythagoras Theorem. If not, I apologise for using it earlier in the calculation of the surface of a cone, and probably causing unnecessary confusion. My preferred proof considers a right angled triangle with adjacent side length x and opposite side of length y, the length of the hypotenuse is r. First we must express right angles in radians for consistency with the rest of the text: A right angle is a quarter of a turn of a wheel. We have seen from the derivation of the circumference of a circle that a full turn will advance the wheel by 2π times the radius, so a full turn is 2π radians. A half turn (180 degrees) is π radians, and a quarter turn is π/2 radians. The internal angles of a triangle therefore add up to π radians. The angles of the triangle are 𝜋 2 𝜋 2 therefore θ, and − 𝜃. Figure 6-26: Re-Arrange Four Trianles Inside a Square to Demonstrate Pythagoras' Theorem We start by drawing a square of side (x+y). Four triangles may be arranged as shown within the square . The hypotenuses of the four triangles form a quadrilateral. Examining the angle at a corner of this rectangle yields the result: 𝜋 𝜑 = 𝜋 − (𝜃 + ( 2 ) − 𝜃) = 𝜋 2 Since all its sides are equal in length (=r), and the corners are right angles, the enclosed figure is a square of side r. The area of the square containing the four triangles is four times the area of one triangle plus the area of this square made up of the hypotenuses of the four triangles: 100 Cybernetics for Systems Engineers – DRAFT F 𝑥𝑦 2 (𝑥 + 𝑦)2 = 4 ( ) + 𝑟 2 From which: 𝑥2 + 𝑦2 = 𝑟2 Or, in the usual archaic language: the square on the hypotenuse equals the sum of the squares on the other two sides. For our triangle of unit length hypoteneuse we have, as a relationship between sine and cosine: sin2 𝜃 + cos 2 𝜃 = 1 Also: 1 + tan2 𝜃 = sec 2 𝜃 And: 1 + cot 2 𝜃 = csc 2 𝜃 6.15.5 Special Cases There are a few cases where the trigonometric functions may be found by inspection. Consider first the case θ=45 ̊(π/4 radians). The triangle is isosceles. If the shorter sides are of length b units, the hypotenuse is √𝑏 2 + 𝑏 2 = √2𝑏 It follows that: 𝜋 𝜋 sin ( 4 ) = cos ( 4 ) = 1 √2 And: 𝜋 tan ( 4 ) = 1 Another interesting case is θ=60 ̊(π/3 radians) This is half an equilateral triangle. If the hypotenuse is b, the adjacent side is b/2 and the opposite side is: 𝑏 2 𝑜 = √𝑏 2 − (2) = √3 𝑏 2 From the geometry of the triangle, this also yields the trig. functions for 30 ̊ (π/6): 𝜋 𝜋 √3 2 𝜋 3 𝜋 6 1 2 𝜋 𝜋 sin 3 = cos 6 = cos = sin = tan 3 = cot 6 = √3 101 Cybernetics for Systems Engineers – DRAFT F 6.15.6 Functions of Sums and Differences of Angles The sine, cosine and tangent are readily available as keystrokes on a scientific calculator. The inquisitive reader will, no doubt, wonder how this is done. The Taylor series expansion of the previous section implies that if we can differentiate these functions, we can produce a polynomial series which will enable us to calculate them for specific values of the argument. In order to differentiate, say a sine function, we need to be able to deal with expressions of the form: 𝑦 + 𝛿𝑦 = sin(𝑥 + 𝛿𝑥) In order to evaluate this, we must be able to evaluate the sine of the sum of a pair of angles in terms of functions of the two angles taken separately. e.g. we need to evaluate expressions of the form: sin(𝜃 + 𝜙) , cos(𝜃 + 𝜙) Construct a right angled triangle ABC,such that angle CAB=θ. Draw a line at angle φ to AC and draw a line perpendicular to AC passing through C, meeting this line at D. AD is of length secφ. Figure 6-27 : Trigonometric Functions of Sums of Angles Draw a line perpendicular to AB, passing through D and intersecting AB at N. Then sin(𝜃 + 𝜙) = 𝐷𝑁 𝐴𝐷 = 𝐷𝑁 cos 𝜙 The intersection of AC and DN is point E. Since AEN and DEC are vertically opposite, they are equal, so that angle EDC=angle EAN=θ. Hence: 𝐷𝐶 = 𝐴𝐷 sin 𝜙 = sec 𝜙 sin 𝜙 𝐷𝐸 = 𝐷𝐶 sec 𝜃 = sec 𝜃 tan 𝜙 102 Cybernetics for Systems Engineers – DRAFT F Also: 𝐸𝑁 = 𝐴𝐸 sin 𝜃 = (𝐴𝐶 − 𝐸𝐶) sin 𝜃 = (1 − 𝐷𝐶 tan 𝜃) sin 𝜃 Or: 𝐸𝑁 = (1 − tan 𝜃 tan 𝜙) sin 𝜃 Hence, using the Pythagoras relationship between sine and cosine: 𝐷𝑁 = 𝐷𝐸 + 𝐸𝑁 = sec 𝜃 tan 𝜙 + sin 𝜃 − sec 𝜃 (1 − cos 2 𝜃) tan 𝜙 Or: 𝐷𝑁 = sin 𝜃 + cos 𝜃 tan 𝜙 The required result for the sine of the sum is: sin(𝜃 + 𝜙) = sin 𝜃 cos 𝜙 + cos 𝜃 sin 𝜙 The cosine is: 𝐴𝑁 cos(𝜃 + 𝜙) = 𝐴𝐷 = 𝐴𝐸 cos 𝜃 cos 𝜙 = (1 − tan 𝜃 tan 𝜙) cos 𝜃 cos 𝜙 Or: cos(𝜃 + 𝜙) = cos 𝜃 cos 𝜙 − sin 𝜃 sin 𝜙 And: sin(𝜃+𝜙) tan(𝜃 + 𝜙) = cos(𝜃+𝜙) = sin 𝜃 cos 𝜙+cos 𝜃 sin 𝜙 cos 𝜃𝑐𝑜𝑠𝜙−sin 𝜃 sin 𝜙 Dividing top and bottom by cos 𝜃 cos 𝜑 : tan 𝜃+tan 𝜙 tan(𝜃 + 𝜙) = 1−tan 𝜃 tan 𝜙 We are now in a position to consider the derivatives of these functions. 6.15.7 Derivatives of Trigonometric Functions We proceed as before by considering the change in the value of the function corresponding to a small change in the value of the independent variable. The derivative is what remains as these changes tend to zero. For the sine function: 𝑦 = sin 𝑥 𝑦 + 𝛿𝑦 = sin(𝑥 + 𝛿𝑥) = sin 𝑥 cos 𝛿𝑥 + cos 𝑥 sin 𝛿𝑥 103 Cybernetics for Systems Engineers – DRAFT F Recall our discussion about hill gradients, that if the angle is small, the horizontal distance is almost equal to the distance along the slope, so that the cosine of a small angle is very nearly 1, and the sine of the angle is very close to the angle expressed in radians. These approximations become ever more exact as the angle approaches zero, so we are justified in writing: 𝑦 + 𝛿𝑦 = sin 𝑥 + cos 𝑥 𝛿𝑥 Eliminating y: 𝛿𝑦 = cos 𝑥 𝛿𝑥 In the limit: 𝑑𝑦 𝑑𝑥 = cos 𝑥 The same procedure, using the same small angle approximations, yields for the cosine: 𝑑 (cos 𝑥) 𝑑𝑥 = − sin 𝑥 From the derivatives of the sine and cosine, it is evident that the second derivative must be: 𝑑2 (sin 𝑥) 𝑑𝑥 2 = − sin 𝑥 𝑑2 (cos 𝑥) 𝑑𝑥 2 = − cos 𝑥 Like the exponential function, these functions imply the associated system contains negative feedback; i.e. a tendency to correct itself. The tangent is slightly more complicated: 𝑦 = tan 𝑥 tan 𝑥+𝛿𝑥 𝑦 + 𝛿𝑦 = 1−𝛿𝑥 tan 𝑥 To first order of small quantities: 𝑦 − 𝑦 tan 𝑥 𝛿𝑥 + 𝛿𝑦 = tan 𝑥 + 𝛿𝑥 Or: 𝛿𝑦 = (1 + tan2 𝑥)𝛿𝑥 The Pythagoras relationship for the tangent function yields: 𝛿𝑦 = sec 2 𝑥 𝛿𝑥 The derivative is, therefore: 𝑑𝑦 𝑑𝑥 = sec 2 𝑥 The same approach works for the inverse function. Let: 104 Cybernetics for Systems Engineers – DRAFT F 𝑦 = tan−1 𝑥 tan 𝑦 = 𝑥 (1 + tan2 𝑦)𝛿𝑦 = 𝛿𝑥 𝑑𝑦 𝑑𝑥 = 1 1+𝑥 2 This is of the form: 𝑦 = 𝑦(𝑧) Where: 𝑧 = 𝑧(𝑥) This form is very common, and the derivative is found from: 𝑑𝑦 𝑑𝑥 = 𝑑𝑦 𝑑𝑧 𝑑𝑧 𝑑𝑥 This can be shown by taking finite increments in x,y and z and considering the limits. In the case of the derivative of tan-1x: 𝑑𝑦 𝑑𝑥 1 = 𝑧, 𝑑 𝑑𝑦 ( ) 𝑑𝑧 𝑑𝑥 1 = − 𝑧2 Where: 𝑧 = 1 + 𝑥2 It follows that: 𝑑2 𝑦 𝑑𝑥 2 2𝑥 = − (1+𝑥 2 )2 6.16 Binomial Theorem Consider now the Taylor series expansion of the binomial expression: 𝑦 = (1 + 𝑥)𝑛 From the reasoning of the previous section: 𝑑𝑦 𝑑𝑥 𝑑𝑥 = 𝑛(1 + 𝑥)𝑛−1 𝑑𝑥 = 𝑛(1 + 𝑥)𝑛−1 So that: 𝑑𝑦 (𝑑𝑥 ) 𝑥=0 =𝑛 Repeated differentiation yields: 𝑑𝑘 𝑦 (𝑑𝑥 𝑘 ) 𝑥=0 105 𝑛! = 𝑛(𝑛 − 1) … (𝑛 − 𝑘 + 1) = (𝑛−𝑘)! Cybernetics for Systems Engineers – DRAFT F The kth term of the Taylor series expansion is therefore: 1 𝑑𝑘 𝑦 ( ) 𝑥𝑘 𝑘! 𝑑𝑥 𝑘 𝑥=0 𝑛! = (𝑛−𝑘)!𝑘! 𝑥 𝑘 This result is known as the Binomial Theorem. We have shown from our derivation of the areas under curves, that the formula applies to fractional, as well as negative values of n. 6.17 A Few Tricks Given sufficient ingenuity and patience, the simple formula for differentiation will get there in the end. If all else fails, the derivative may be found by finite increments in the variables. For many functions that is extremely tedious, as we found from the higher derivatives of tan-1 x. There are a couple of tricks which can speed up the evaluation of awkward derivatives. A common case is when the function is a product of two simpler functions of the independent variable. 𝑦 = 𝑢(𝑥)𝑣(𝑥) 𝑦 + 𝛿𝑦 = 𝑢(𝑥 + 𝛿𝑥)𝑣(𝑥 + 𝛿𝑥) 𝛿𝑢 𝛿𝑣 𝑦 + 𝛿𝑦 = (𝑢(𝑥) + 𝛿𝑥 𝛿𝑥) (𝑣(𝑥) + 𝛿𝑥 𝛿𝑥) Expanding to first order of small quantities: 𝛿𝑢 𝛿𝑣 𝑦 + 𝛿𝑦 = 𝑢(𝑥)𝑣(𝑥) + (𝑣(𝑥) 𝛿𝑥 + 𝑢(𝑥) 𝛿𝑥) 𝛿𝑥 The derivative is, therefore: 𝑑𝑦 𝑑𝑥 𝑑𝑢 𝑑𝑣 = 𝑣(𝑥) 𝑑𝑥 + 𝑢(𝑥) 𝑑𝑥 In particular, if v(x)=1/z(x) 𝑑𝑦 𝑑𝑥 1 𝑑𝑢 𝑢(𝑥) 𝑑𝑧 = 𝑧(𝑥) 𝑑𝑥 − 𝑧(𝑥)2 𝑑𝑥 = 𝑧(𝑥) 𝑑𝑢 𝑑𝑦 −𝑢(𝑥) 𝑑𝑥 𝑑𝑥 𝑧(𝑥)2 This form is more suitable for finding the nth derivative of tan-1x. The significance of this series arises from the fact that: 𝜋 4 = tan−1 1 So that expanding tan-1x as a Taylor series, and substituting x=1, we have a means of calculating π, which converges much more quickly than our initial attempt. Our motivation for finding the derivatives of trigonometric functions was to satisfy our curiosity as to how to go about calculating them. We shall see that they crop up regularly in problems which have nothing to do with triangles. 106 Cybernetics for Systems Engineers – DRAFT F 6.18 Differential Equations 6.18.1 Introduction A differential equation is one which contains derivatives and functions of derivatives of the variable of interest, as well as functions of the variable itself. We have met the simplest form of differential equation of the form: 𝑑𝑦 𝑑𝑥 = 𝑓(𝑥) Which is solved directly by integration. The vast majority of differential equations do not have mathematically tractable solutions, and we must revert to numerical solution methods. Numerical methods are fine for finding the solutions in the unknown variable when the parameters defining the equations (e.g. the coefficients of the functions of derivatives) have fixed numerical values. However, the insight we seek is invariably concerned with how the parameters affect the solution. Thus hundreds, if not thousands of cases must be run before such insight can be gained into the problem by handle-cranking a computer model. All too often huge sums are invested in producing a high fidelity computer simulation of a system, generally by managers and software engineers who are ignorant of just how useless they are at providing the insight needed to support design or take remedial action in the event of performance shortfall. Only after the design work is over, the dust has settled, and it is time to demonstrate the system meets its requirements are such simulations required. During the process of getting there, they are more trouble than they are worth. Most of the time we are not really interested in what the answer is, but what determines the features of the solution. There exists a class of equation, which probably doesn’t really represent any real world system exactly, but with sufficient ingenuity, may be used as a working approximation for practically any self-correcting system within clearly defined bounds of validity. That the representations are not exact does not worry us unduly, because a self-correcting systems which use negative feedback as the correction process (like the cooling bowl of soup), are inherently robust to uncertainties in their mathematical descriptions. In any case, there exist methods which can quantify the level of uncertainty which can be tolerated, before this inherent robustness is lost. 6.18.2 First Order Equation The differential equations of interest are called linear homogenous differential equations. We met one when considering soup going cold. 6.18.2.1 The Bowl of Soup The differential equation took the form: 𝑑𝑇 𝑑𝑡 107 = 𝑘(𝑇𝑠 − 𝑇) Cybernetics for Systems Engineers – DRAFT F We converted it into a simple integral by noticing that if the temperature of the surroundings is constant, its rate of change would be zero, so it would not affect the derivative. However, this would not be valid if the temperature of the surroundings were not constant. More generally, this would take the form: 𝑑𝑇 𝑑𝑡 + 𝑘𝑇 = 𝑘𝑇𝑠 The solution for temperature difference was an exponential, so that seems a reasonable guess for the actual temperature: 𝑇 = 𝑏(𝑡)𝑒 −𝑎𝑡 𝑇𝑒 𝑎𝑡 = 𝑏(𝑡) Differentiating: 𝑑𝑇 𝑒 𝑎𝑡 𝑑𝑡 + 𝑎𝑒 𝑎𝑡 𝑇 = 𝑑𝑏 𝑑𝑡 Comparing coefficients: 𝑑𝑇 𝑑 𝑒 𝑘𝑡 ( 𝑑𝑡 + 𝑘𝑇) = 𝑑𝑡 (𝑇𝑒 𝑘𝑡 ) = 𝑒 𝑘𝑡 𝑘𝑇𝑠 Integrating: 𝑡 [𝑇𝑒 𝑘𝑡 ]𝑡0 = ∫0 𝑘𝑇𝑠 𝑒 𝑘𝑡 𝑑𝑡 In the case of constant temperature surroundings: 𝑇𝑒 𝑘𝑡 − 𝑇(0) = 𝑇𝑠 𝑒 𝑘𝑡 − 𝑇𝑠 (𝑇𝑠 − 𝑇) = (𝑇𝑠 − 𝑇(0))𝑒 −𝑘𝑡 This is expected from our earlier derivation. 6.18.2.2 Charging a Capacitor We have seen that the discharging of a capacitor was governed by the exponential function. Changing the circuit slightly, we can consider the process of charging. In this case, the resistor R and the capacitor C are connected in series, and closing the switch connects the combination across the terminals of the battery. 108 Cybernetics for Systems Engineers – DRAFT F Figure 6-28 : CR Circuit Again The objective is to find the voltage across the capacitor as a function of time. Remember that the current passing into the capacitor is given by: 𝑑𝑉 𝑖 = 𝐶 𝑑𝑡 And the voltage drop across the resistor is: 𝑑𝑉 𝐸 − 𝑉 = 𝑖𝑅 = 𝐶𝑅 𝑑𝑡 The resulting equation is: 𝑑𝑉 𝑑𝑡 1 1 + 𝐶𝑅 𝑉 = 𝐶𝑅 𝐸 This is exactly the same form as the equation governing the cooling of the soup. The left hand side is: 𝑡 𝑡 𝑑 𝑒 −𝐶𝑅 (𝑑𝑡 (𝑉𝑒 𝐶𝑅 )) The solution is, therefore: 𝑡 𝑡 1 𝑡 𝑡 [𝑉𝑒 𝐶𝑅 ] = 𝐶𝑅 ∫0 𝐸𝑒 𝐶𝑅 𝑑𝑡 0 If E is constant: 𝑡 𝑡 𝑉𝑒 𝐶𝑅 − 𝑉(0) = 𝐸 (𝑒 𝐶𝑅 − 1) The capacitor is initially discharged, so V(0)=0, and the charging voltage is: 109 Cybernetics for Systems Engineers – DRAFT F 𝑡 𝑉 = 𝐸 (1 − 𝑒 −𝐶𝑅 ) 6.18.2.3 Hot Air Balloon A hot air balloon flies because the weight of air it displaces is equal to its own weight. Ascent and descent is controlled by using a heat source to adjust the density of the gas in the balloon, or by discharging ballast. The lift of the gas bag is given by: 𝐿 = (𝜌𝑎𝑖𝑟 − 𝜌𝑔𝑎𝑠 )𝑔𝑉 Where: ρ denotes local air density, g the acceleration of gravity and V the volume of the gas bag. The change of atmospheric pressure with height is given by the hydrostatic equation: 𝑑𝑝 𝑑ℎ = −𝜌𝑔 Where p is pressure, h is height. This states that pressure falls with height by an amount equal to the weight of gas per unit area. The pressure is related to density and absolute temperature (T) by the gas equation: 𝑝 = 𝜌𝑅𝑇 Where: R is the gas constant for the gas of interest. For air it has the value287 J kg-1 K-1. Differentiating the gas equation: 𝑑𝑝 𝑑ℎ 𝑑𝜌 𝑑𝑇 = 𝑅𝑇 𝑑ℎ + 𝜌𝑅 𝑑ℎ In the troposphere, the temperature drops at a rate of 6.5 degrees Centigrade per kilometre. This is called the lapse rate and is denoted α (alpha). The hydrostatic equation, the gas equation and the lapse equation are combined to yield an expression for air density as a function of altitude: 𝑅(𝑇(0) − 𝛼ℎ) 𝑑𝜌 𝑑ℎ 𝜌 𝑑𝜌 𝑅 ∫ (𝛼𝑅−𝑔) 𝜌(0) 𝜌 = ∫0 − 𝜌𝑅𝛼 = −𝜌𝑔 Or: ℎ 𝑑ℎ 𝑇(0)−𝛼ℎ The right hand side is integrated by letting 𝑧 = 𝑇(0) − 𝛼ℎ 𝑑𝑧 = −𝛼 𝑑ℎ ℎ ∫0 Integrating: 110 𝑑ℎ 𝑇(0)−𝛼ℎ 1 𝑇(0)−𝛼ℎ 𝑑𝑧 = − 𝛼 ∫𝑇(0) 𝑧 Cybernetics for Systems Engineers – DRAFT F 1 𝑔 (1− ) 𝛼𝑅 ln 𝜌 𝜌(0) 𝑇(0) ) 𝑇(0)−𝛼ℎ = ln ( Or: 𝛼ℎ ( 𝜌 = 𝜌(0) (1 − 𝑇(0)) 𝑔 −1) 𝛼𝑅 The sea level temperature is about 288 K, so if we are considering heights less than 1km, αh is very much less than T(0), so we are justified in treating the change in density with height as a nearly linear function: 𝑔 𝛼ℎ 𝜌 ≈ 𝜌(0) (1 − (𝛼𝑅 − 1) 𝑇(0)) (Apply the Binomial theorem expansion to the original formula and neglect all but the first term to get this). A reduction in the lift gas density, will cause the balloon to rise to a new equilibrium level. In order to avoid cluttering the equations with extraneous symbols, let: 𝑔 𝛼 𝛽 = (𝛼𝑅 − 1) 𝑇(0) (β is the Greek letter beta). The change in lift gas density (assuming the balloon volume remains constant), is equal to the difference in air density between the current height and the new equilibrium height. The upwards excess buoyancy force (F) is therefore: 𝐹 = (𝜌(ℎ2 ) − 𝜌(ℎ1 ))𝑔𝑉 Expressed in terms of height difference: 𝐹 = 𝛽(ℎ2 − ℎ1 )𝜌(0)𝑔𝑉 As the balloon is rising through the air, it will be subject to an aerodynamic drag force opposing its motion. At low speeds for objects the size of a typical hot air balloon we should expect the drag to be proportional to the square of the speed. 1 𝐷 = 2 𝜌𝑈 2 𝑆𝑐𝐷 Where U is the upwards velocity magnitude, S is a reference area (typically the cross-section of the body) and cD is a drag coefficient based on the reference area chosen. We assume that the difference in density between the start and end heights is small, so that the density may be considered constant in this equation. The balloon obeys Newton’s Second Law: 111 Cybernetics for Systems Engineers – DRAFT F 𝑚 𝑑𝑈 𝑑𝑡 = 𝐹−𝐷 Where m is the mass of the balloon and payload and t is time. This is not a particularly useful form of the equation. It is easier to solve if h is the independent variable: 𝑑𝑈 𝑑𝑡 = 𝑑𝑈 𝑑ℎ 𝑑ℎ 𝑑𝑡 1 𝑑(𝑈 2 ) 𝑑ℎ 𝑑𝑈 = 𝑈 𝑑ℎ = 2 The result is an equation in U2: 𝑑(𝑈 2 ) 𝑑ℎ + 𝜌𝑆𝑐𝐷 2 𝑈 𝑚 = 2𝛽(ℎ2 −ℎ)𝜌(0)𝑔𝑉 𝑚 This takes the form of a first order linear homogenous equation, like the other two examples, but illustrates the point that most of the effort is expended in formulating the problem. Unlike the mathematician, problems are not ‘given’ to the engineer. A formulation appropriate to the task in hand, making reasonable assumptions, must be found, and this is rarely a trivial exercise. Evidently the left hand side of the equation of motion is: 𝑒− 𝜌𝑆𝑐𝐷 ℎ 𝑚 𝑑 (𝑑ℎ (𝑈 2 𝑒 𝜌𝑆𝑐𝐷 ℎ 𝑚 )) The solution is found from: 𝑈2𝑒 𝜌𝑆𝑐𝐷 ℎ 𝑚 = 2𝛽𝜌(0)𝑔𝑉 ℎ2 ∫ℎ (ℎ2 𝑚 1 − ℎ)𝑒 𝜌𝑆𝑐𝐷 ℎ 𝑚 𝑑ℎ This can be solved for height as a function of time: 𝑑ℎ 𝑑𝑡 =𝑈 6.18.3 Integration by Parts Our hot air balloon example introduced an integral of the product of two elementary functions: 𝑏 ∫𝑎 𝑥 𝑒 𝑥 𝑑𝑥 We could expand this as a series and try to interpret the result in terms of familiar functions. However, when we last encountered an (at the time) intractable integral, we found that it could be evaluated by considering integrating to the left of the curve of interest, and subtracting the resulting area from the rectangle enclosing the curve. Recall that the resulting formula was: 𝑏 𝑦(𝑏) ∫𝑎 𝑦𝑑𝑥 = [𝑥𝑦(𝑥)]𝑏𝑎 − ∫𝑦(𝑎) 𝑥𝑑𝑦 In the case of the motion of the balloon, the function of interest is: 𝑦 = 𝑥𝑒 𝑎𝑥 Differentiating, using the formula for the derivative of a product: 112 Cybernetics for Systems Engineers – DRAFT F 𝑑𝑦 𝑑𝑥 = 𝑒 𝑥 + 𝑎𝑥𝑒 𝑎𝑥 This appears to make matters worse, as it introduces an integral of the form: 𝑏 ∫𝑎 𝑥 2 𝑒 𝑎𝑥 𝑑𝑥 This situation arises quite frequently when using this method. All is not lost, however. Consider the more general function: 𝑦 = 𝑥 𝑛 𝑒 𝑎𝑥 The derivative is: 𝑑𝑦 𝑑𝑥 = 𝑛𝑥 𝑛−1 𝑒 𝑎𝑥 + 𝑎𝑥 𝑛 𝑒 𝑎𝑥 Denoting, for convenience: 𝑏 ∫𝑎 𝑦 𝑑𝑥 = 𝐼𝑛 Then: 𝐼𝑛 = [𝑥𝑦(𝑥)]𝑏𝑎 − 𝑛𝐼𝑛 − 𝑎𝐼𝑛+1 Or: 1 𝐼𝑛+1 = ([𝑥𝑦(𝑥)]𝑏𝑎 − (𝑛 + 1)𝐼𝑛 ) 𝑎 Setting n=0, we have: 𝑦(𝑥) = 𝑒 𝑥 𝑏 𝑏 1 ∫𝑎 𝑥𝑒 𝑎𝑥 𝑑𝑥 = 𝑎 ([𝑥𝑒 𝑎𝑥 ]𝑏𝑎 − ∫𝑎 𝑒 𝑎𝑥 𝑑𝑥 ) In the example of the balloon, we need to evaluate: ℎ ∫ℎ (ℎ2 − ℎ)𝑒 𝜌𝑆𝑐𝐷 ℎ 𝑚 1 𝑑ℎ To reduce clutter, let: 𝜌𝑆𝑐𝐷 𝑚 = 1 ℎ𝑐 The integral becomes: ℎ ℎ ℎ ℎ ℎ ℎ2 ∫ℎ 2 𝑒 ℎ𝑐 𝑑ℎ − ∫ℎ 2 ℎ𝑒 ℎ𝑐 𝑑ℎ = ℎ2 ℎ𝑐 [𝑒 ℎ𝑐 ] The solution for the speed squared is: 113 ℎ2 ℎ ℎ − ℎ𝑐 [ℎ𝑒 ℎ𝑐 ] ℎ2 ℎ ℎ ℎ2 + ℎ𝑐2 [𝑒 ℎ𝑐 ] ℎ Cybernetics for Systems Engineers – DRAFT F 2𝛽𝜌(0)𝑔𝑉 ((ℎ2 𝜌(ℎ1 )𝑆𝑐𝐷 𝑈2 = − ℎ + ℎ𝑐 ) − (ℎ2 − ℎ1 + ℎ𝑐 )𝑒 (ℎ1 −ℎ) ℎ𝑐 ) The reader is welcome to try and solve this explicitly for h as a function of time. If h=h1, the vertical velocity is zero. The apogee is reached when this velocity is again zero. Let: ℎ1 −ℎ ) ℎ𝑐 𝑧=( The apogee height is given by: ℎ2 −ℎ1 )+ ℎ𝑐 (( 1) (1 − 𝑒 𝑧 ) + 𝑧 = 0 Thinking about the problem, we notice that balloons tend to be rather large, say about 10m in diameter. Also they usually are spherical or some other bluff shape when moving vertically, so the drag coefficient based on cross section will not be much less than 1. Also the air density near sea level is expected to be about 1.23 kg m-3. A typical value for hc is about: ℎ𝑐 = 𝜌𝜋𝑅 2 𝑐𝐷 ≈ 1.23 × 𝜋 × 102 × 1.0 = 386𝑚 The second order term in the Taylor series expansion depends is: 1 ℎ1 −ℎ 2 ( ℎ ) 2 𝑐 If we are typically dealing with height changes of the order of 30m (≈100ft), ignoring this term introduces an error of about 0.3%, so we shouldn’t introduce much error if we expand the exponential term to second order. The apogee height is given to a reasonable approximation by: ℎ2 −ℎ1 )+ ℎ𝑐 (( 1) (1 − 1 − 𝑧 − 𝑧2 )+ 2 The result is z=0, or: ℎ2 −ℎ1 )+ ℎ𝑐 (( 𝑧=− 𝑧 2 1) (1 + ) − 1 = 0 ℎ −ℎ 2( 2 1 ) ℎ𝑐 ℎ −ℎ (( 2ℎ 1 )+1) 𝑐 There will be no overshoot in height if: ℎ2 −ℎ1 ) ℎ𝑐 𝑧 = −( This will occur if: 114 𝑧=0 Cybernetics for Systems Engineers – DRAFT F ℎ2 − ℎ1 = ℎ𝑐 Notice how the engineer brings in a wider understanding of the problem space which is not available to the mathematician. This awareness of the relative magnitudes of the quantities, and the simplifications it permits, is usually dismissed disparagingly by the pure mathematician as ‘ad hoc’. All too often it is easy to mistake one who can merely manipulate the algebra for one who actually understands the problem. Similarly, it does not follow that, because an individual can implement equations in computer code, they have any understanding of the subject matter whatsoever. All too often the credit for the technical staff’s work goes to the software engineers who merely implemented it. So, within the bounds of validity we have defined: 𝑈2 = 2𝛽𝜌(0)𝑔𝑉 ((ℎ2 𝜌(ℎ1 )𝑆𝑐𝐷 ℎ1 −ℎ 1 ℎ1 −ℎ 2 ) + ( ℎ ) )) ℎ𝑐 2 𝑐 − ℎ + ℎ𝑐 ) − (ℎ2 − ℎ1 + ℎ𝑐 ) (1 + ( 6.18.4 More Exotic Forms The solution for the hot air balloon vertical velocity is given by an expression of the form: 𝑈 2 = 𝑎ℎ2 + 𝑏ℎ + 𝑐 Where a, b and c are constants, the equation for height as a function of time is found from: 𝑑ℎ 𝑑𝑡 = √𝑎ℎ2 + 𝑏ℎ + 𝑐 The closest we have come to functions of this form is with the Pythagoras relationship between trigonometric functions. The term under the square root may be written: 𝑏 2 𝑏2 𝑎ℎ2 + 𝑏ℎ + 𝑐 = (√𝑎ℎ + 2 𝑎) + 𝑐 − 4𝑎 √ From which: 𝑏2 √𝑎ℎ2 + 𝑏ℎ + 𝑐 = √𝑐 − 4𝑎 √𝑥 2 + 1 Where: 𝑥= 𝑏 ) 2√𝑎 2 𝑏 (√𝑎ℎ+ √𝑐− 4𝑎 Note that if the denominator took the common form : √1 − 𝑥 2 , we could use the substitutions: 𝑥 = sin 𝜃 𝑑𝑥 = 𝑐𝑜𝑠𝜃 𝑑𝜃 Also, for the form: 1 + 𝑥 2 we could use: 𝑥 = tan 𝜃 115 Cybernetics for Systems Engineers – DRAFT F 𝑑𝑥 = sec 2 𝜃 𝑑𝜃 And for the form: 1 − 𝑥 2 , we can factorise into one term in 1-x and another in (1+x): 𝑏 𝑑𝑥 1−𝑥 2 ∫𝑎 𝑏 𝑑𝑥 1−𝑥 1 = 2 (∫𝑎 𝑏 𝑑𝑥 ) 1+𝑥 + ∫𝑎 These common forms are derived directly from the properties of trigonometric functions, and elementary algebra. Many physical problems may be approximated using them. Unfortunately, the motion of the balloon is not one of them. It takes the form: 𝑓(𝑥) = √𝑥 2 + 1 If we are confident using complex numbers, we could use the substitution: 𝑥 = 𝑗 sin 𝜃 𝑑𝑥 = 𝑗 cos 𝜃 𝑑𝜃 Where 𝑗 = √−1 . However, I assume the reader has either not yet encountered complex numbers or has forgotten about them over the years. We are seeking a function y(z) such that: 𝑦2 = 𝑥2 + 1 𝑑𝑥 𝑑𝑧 =𝑦 Which implies: 𝑑𝑦 𝑑𝑥 𝑦 𝑑𝑧 = 𝑥 𝑑𝑧 𝑑𝑦 𝑑𝑧 =𝑥 The second derivative is: 𝑑2 𝑦 𝑑𝑧 2 𝑑𝑥 = 𝑑𝑧 = 𝑦 The required function is governed by the equation: 𝑑2 𝑦 𝑑𝑧 2 =𝑦 Remember, from the derivative of sine and cosine, that the equation: 𝑑2 𝑦 𝑑𝑧 2 = −𝑦 Has a solution involving sine and cosine terms. The solution with a positive coefficient of y is in terms of functions, which by analogy are called hyperbolic sine (sinh) and hyperbolic cosine (cosh). 116 Cybernetics for Systems Engineers – DRAFT F We notice that the equation governing our desired function exhibits feedback. The feedback is positive, so we should expect the solution to diverge as the independent variable z increases. From our discussion about e, we note that if y=ez or y=e-z: 𝑑2 𝑦 𝑑𝑧 2 = 𝑒 𝑧 𝑜𝑟 𝑒 −𝑧 It follows that: 𝑦 = 𝐴𝑒 𝑧 + 𝐵𝑒 −𝑧 𝑑𝑦 𝑑𝑧 = 𝑥 = 𝐴𝑒 𝑧 − 𝐵𝑒 −𝑧 Where A and B are constants. Now: 𝑦2 = 𝑥2 + 1 𝐴2 𝑒 2𝑧 + 2𝐴𝐵 + 𝐵2 𝑒 −2𝑧 = 𝐴2 𝑒 2𝑧 − 2𝐴𝐵 + 𝐵2 𝑒 −2𝑧 + 1 By convention A=B, so 4A2=1; A=½. The result is the definition of the hyperbolic functions: cosh(𝑧) = 𝑒 𝑧 +𝑒 −𝑧 2 sinh(𝑧) = 𝑒 𝑧 −𝑒 −𝑧 2 (Read this as ‘cosh z, shine z’. So the hot air balloon height against time, and any other integral containing the √𝑥 2 + 1 factor, takes the form of an inverse hyperbolic function. 6.19 The Lanchester Equation of Combat A rather grim example of a divergent solution is furnished by the equation of combat, formulated by Frederic W Lanchester, the automobile and aviation pioneer, in 1916. Consider two opposing armies engaged in combat. We shall call them blue and red. The rate of loss of blue troops is assumed proportional to the number of red troops and vice versa. We have: 𝑑𝐵 𝑑𝑡 = −𝑘𝑅 𝑅 𝑑𝑅 𝑑𝑡 = −𝑘𝐵 𝐵 Where t is time, B and R are the numbers of surviving troops in the blue and red armies respectively, kB and kR are the effectiveness of the blue and red army respectively. Considering blue’s fortunes (the equation for red is similar): 117 Cybernetics for Systems Engineers – DRAFT F 𝑑2 𝐵 𝑑𝑡 2 = 𝑘𝐵 𝑘𝑅 𝐵 From the previous section, this has solution: 𝐵 = 𝑎𝑒 √𝑘𝑅𝑘𝐵 𝑡 + 𝑏𝑒 −√𝑘𝑅𝑘𝐵𝑡 When time =0, blue starts the battle with B0 troops. 𝑎 + 𝑏 = 𝐵0 𝑑𝐵 𝑑𝑡 = −𝑘𝑅 𝑅 = √𝑘𝐵 𝑘𝑅 (𝑎𝑒 √𝑘𝐵𝑘𝑅𝑡 − 𝑏𝑒 −√𝑘𝐵 𝑘𝑅𝑡 ) When time =0, red starts the battle with R0 troops. 𝑘 𝑎 − 𝑏 = −√𝑘𝑅 𝑅0 𝐵 Solving for a and b: 1 2 𝑘𝑅 𝑅 ) 𝑘𝐵 0 𝑎 = (𝐵0 − √ 1 𝑘 𝑏 = 𝐵0 − 𝑎 = 2 (𝐵0 + √𝑘𝑅 𝑅0 ) 𝐵 We have: 1 2 𝑘𝑅 𝑅 ) 𝑒 √𝑘𝐵 𝑘𝑅𝑡 𝑘𝐵 0 𝐵 = (𝐵0 − √ 1 2 𝑘𝑅 𝑅 ) 𝑒 −√𝑘𝐵 𝑘𝑅𝑡 𝑘𝐵 0 + (𝐵0 + √ Or, more neatly: 𝑘 𝐵 = 𝐵0 cosh(√𝑘𝑅 𝑘𝐵 𝑡) + √𝑘𝑅 𝑅0 sinh(√𝑘𝑅 𝑘𝐵 𝑡) 𝐵 This is correct, but it doesn’t shed much light on the problem in this form. For convenience, and the avoidance of writer’s cramp, we shall make the substitutions: 𝑘 √𝑘𝑅 = 𝛼 𝐵 𝑒 √𝑘𝐵 𝑘𝑅𝑡 = 𝑥 The equation now becomes: 1 2 1 2 𝐵 = (𝐵0 − 𝛼𝑅0 )𝑥 + (𝐵0 + 𝛼𝑅0 ) 1 𝑥 Similarly: 1 1 1 1 1 𝑅 = 2 (𝑅0 − 𝛼 𝐵0 ) 𝑥 + 2 (𝑅0 + 𝛼 𝐵0 ) 𝑥 118 Cybernetics for Systems Engineers – DRAFT F If we square both these equations and then subtract one from the other, we end up with B as a function of R: 𝐵02 − 𝐵2 = 𝛼 2 (𝑅02 − 𝑅 2 ) We can derive an expression for x from either equation: 1 (𝐵0 2 1 − 𝛼𝑅0 )𝑥 2 − 𝐵𝑥 + 2 (𝐵0 + 𝛼𝑅0 ) = 0 This is a quadratic equation having solution: 𝑥= 𝐵±√𝐵2 −𝐵02 +𝛼2 𝑅02 (𝐵0 −𝛼𝑅0 ) When time=0, x=1, B=B0, so the square root must be negative. The solution becomes: (𝐵−𝛼𝑅) 𝑥 = (𝐵 0 −𝛼𝑅0 ) √𝑘𝐵 𝐵 − √𝑘𝑅 𝐵 = (√𝑘𝐵 𝐵0 − √𝑘𝑅 𝑅0 )𝑒 √𝑘𝑅𝑘𝐵𝑡 The difference in the two armies increases exponentially as the battle proceeds, so that the side which is weaker will eventually be wiped out. There is a stage in the battle when it becomes evident that the situation is hopeless, and the rational commander would surrender. The history of warfare indicates that commanders feared the charge of cowardice more than catastrophic and pointless losses. It is difficult to understand why the bards wax so lyrical in celebrating military incompetence. In the First World War, Admiral Troubridge faced a court martial for failing to expend his cruiser squadron in a futile attempt to prevent the German battle cruiser Goeben from crossing the Mediterranean to Turkey. Evidently the Navy was determined not to be beaten by the Army in the reckless stupidity stakes. If the effectiveness of the two sides in the context of the battle, were known beforehand, the rational action would be for the weaker side to surrender without a fight, unless the purpose of the battle were to delay the enemy whilst reinforcements are being sent. Nearly all military tactics are geared to gaining an advantage locally, so that the enemy is outnumbered at a particular point of contact. This can only be done by introducing potential weakness elsewhere. Winning or losing depends on denying accurate knowledge on the disposition of one’s own forces to the enemy. Napoleon relied heavily on a network of spies able to provide intelligence of potential enemies raising armies. A pre-emptive strike whilst the French Army held the advantage in numbers not only defeated the unprepared enemy, but also tacked their country on to the French Empire. Hannibal destroyed a Roman force four times the size of his own at the Battle of Cannae, by encircling it, ensuring only the relatively small number of Roman soldiers on the perimeter could actually make contact, most of the army was confined inside the crowd and couldn’t bring weapons 119 Cybernetics for Systems Engineers – DRAFT F to bear. Similarly, tales of heroes holding bridges or narrow passes illustrate the importance of maintaining superiority only where the opposing forces met. Modern wars rely heavily on advanced remote sensing from aircraft and satellites as well as spies on the ground, to gain accurate knowledge of enemy disposition and movements, whilst at the same time denying such intelligence to the enemy by decoys, camouflage and concealment. To quote Sun Zi, the earliest published authority on war from the 5th Century BC; ‘where you are strong appear weak, where you are weak appear strong’. If the two sides are evenly matched, the duration of the battle tends to infinity, and even the winning side loses nearly all its troops. The rational behaviour would be to avoid such a conflict by making sure one can match the military power of a potential aggressor. ‘ Let him who seeks peace, prepare for war’ (Flavius Vegetus Renatus). The major problem is the build up to war is itself an unstable process, as during this stage opposing forces will try to deploy to more favourable positions for the conflict. Such action is inevitably viewed as hostile, encouraging the side whose advantage is being eroded to fire the first shot. 6.20 Impulse Response If we consider the derivation of the differential equations considered so far, they take the form: ∑𝑛0 𝑎𝑖 𝑑𝑖𝑦 𝑑𝑡 𝑖 = 𝑓(𝑡) Where t is an independent variable, which the balloon example shows, is not necessarily time, although usually it is. The ai are constants. Evidently, the left hand side is derived from the dynamic behaviour of the object of interest, whilst the right hand side is a disturbance applied to the object. The right hand side is the description of the input, the equation itself is a mathematical model of the objects behaviour, and the dependent variable y is the output. The response to one particular input tells us very little about how the object will respond to other inputs. We need to find an elementary input from which the output to any input may be derived. The idea of an impulse was mentioned with the example of rocket acceleration. My apologies for introducing a concept without first explaining it. The impulse of a force is the product of the force and the time it is applied for, as the duration of the application of the force becomes insignificant. In essence, the force may be treated as impulsive if there is no significant motion during its application. Examples include hitting a golf ball with a club. We may treat a continuous input function approximately as a sequence of impulses of finite duration; the approximation converges on to the function as the duration tends to zero. If the equation is linear, the complete solution may be found by combining the outputs resulting from a sequence of impulses. The starting point is to discover the response to an impulsive input. 120 Cybernetics for Systems Engineers – DRAFT F In general the output in response to a unit impulse applied at time t=0, may be written as h(t), so that an impulse applied at a later time (t1) will produce an output h(t-t1) . The amplitude of the input is f(t1)δt, where δt is the duration of the impulse. The output is: ∞ 𝑦(𝑡) ≈ ∑𝑛1 𝑓(𝑖𝛿𝑡)ℎ(𝑡 − 𝑖𝛿𝑡) 𝛿𝑡 → ∫0 𝑓(𝑡1 )ℎ(𝑡 − 𝑡1 )𝑑𝑡1 If the impulse response is known, the response to any other input should be calculable from the above integral. As an example, consider the charging capacitor: 𝑑𝑣 𝑑𝑡 1 1 + 𝐶𝑅 𝑣 = 𝐶𝑅 𝐸(𝑡) 1 𝑣𝑒 𝐶𝑅 = 𝑡 1 𝛿𝑡 𝐶𝑅 𝑑𝑡 𝐸(𝑡)𝑒 ∫ 𝐶𝑅 0 We define δt as short compared with the characteristic time of E(t), so over the interval of integration; E(t)=E(0). 𝛿𝑡 𝑡 𝑣 = 𝐸(0) [𝑒 𝐶𝑅 − 1] 𝑒 −𝐶𝑅 Expanding the exponential in the square brackets to first order: 𝛿𝑡 𝑡 𝑣 = 𝐸(0) 𝐶𝑅 𝑒 −𝐶𝑅 E(0)δt is a voltage applied very briefly, or a voltage ‘spike’. The effect of a spike depends on both its amplitude and duration. We are interested in the case where the product of the two has unit value (1 Volt sec), so that it may be scaled by an arbitrary input. The shorter the spike, the larger its amplitude. Replacing the input spike strength E(0)δt, with unity, we have the system response to a unit input impulse: 1 1 𝑣 = 𝐶𝑅 𝑒 −𝐶𝑅 The astute reader will notice that this equation is apparently dimensionally inconsistent. The left hand side has units of volts and the right hand side has units sec-1. This is usually glossed over, or ignored completely, since the student can get away with not fully understanding the physical meaning when manipulating the maths. The engineer, on the other hand, must interpret the maths in terms of real world behaviour. Ideally, the pure number 1 should be replaced with a symbol, such as I, to indicate that it has dimensions consistent with the output and the independent variable, but unity magnitude. We usually hide this conceptual difficulty by defining a function δ(T) which is zero everywhere apart from where time=T. ∞ 𝑡 ∫0 𝐸(𝑡) 𝑒 𝐶𝑅 𝛿(0)𝑑𝑡 = 𝐸(0) 121 Cybernetics for Systems Engineers – DRAFT F This delta function allows us to write the answer for the impulse response immediately, provided we don’t comment on the dimensional inconsistency of the result. Comparing the result for the impulse response with the response with no input, i.e. starting with an initial voltage across the capacitor: 𝑡 𝑉 = 𝑉0 𝑒 −𝐶𝑅 This should not be surprising, since the impulse response may be written: 𝑑𝑣 𝑑𝑡 Or: + 1 𝑣 𝐶𝑅 = 1 𝐸(𝑡)𝛿(0) 𝐶𝑅 𝑑𝑣 𝑣0 = 𝐸(0) − 𝐶𝑅 ( 𝑑𝑡 ) 𝑡=0 So that starting the system with an initial value of output has the same effect as applying an impulse to the input. The response following an initial displacement is called the transient response. Note that if there is no feedback, the system will remain in its initial state indefinitely. Evidently, for any system governed by a first order homogenous differential equation of the form governing capacitor discharge, cooling soup or hot air balloons, i.e: 𝑑𝑦 𝑑𝑡 + 𝑎𝑦 = 𝑓(𝑡) The transient response, and the impulse response take the form: 𝑦 = 𝑦0 𝑒 𝜆𝑡 Where λ (lamda) is a constant. 𝑦0 𝑒 𝜆𝑡 (𝜆 + 𝑎) = 𝑓(𝑡)𝛿(0) This holds for time>0. Also the exponential cannot be zero, and since we are considering transient response, neither is y0. The transient response is given immediately by: 𝜆+𝑎 =0 This is called the characteristic equation. We see that if it has a negative root, the feedback is negative, and the state will converge on to zero. If positive, the solution diverges indefinitely. Most of the time we are only concerned with whether the solution is stable or unstable, we are rarely interested in solutions for specific cases expressed as functions of time. This information is immediately furnished by the characteristic equation. 6.20.1 Second, and Higher Order The relationship between the transient response and the impulse response may be extended to equations of any order, since if: ∑𝑛0 𝑎𝑖 122 𝑑𝑖𝑦 𝑑𝑡 𝑖 = 𝑓(𝑡) Cybernetics for Systems Engineers – DRAFT F 𝑦0 = 1 𝑓(𝑡)𝛿(0) − 𝑎0 𝑑𝑖𝑦 ) 𝑑𝑡 𝑖 𝑡=0 ∑𝑛1 ( The Lanchester Equation showed that a second order equation of the form: 𝑑2 𝑦 𝑑𝑡 2 = 𝑎𝑦 Has solutions in terms of the exponential function. The characteristic equation has the form: 𝜆2 − 𝑎 = 0 This has one positive root, so the solution diverges. The solution may be written in the form: 𝑦 = 𝐴 cosh(√𝑎𝑡) + 𝐵 sinh(√𝑎𝑡) Recall, from the differentiation of the trigonometric functions, the equation: 𝑑2 𝑦 𝑑𝑡 2 = −𝑎𝑦 Has solution of the form: 𝑦 = 𝐴 cos(√𝑎𝑡) + 𝐵 sin(√𝑎𝑡) This is not obviously an exponential form. We shall introduce the constant 𝑗 = √−1 , mainly to give a physical meaning to it in terms of the solutions to these equations. After all, we know the form the solutions take; it is the number j which looks a bit fishy. It is more sensible to define what we don’t know in terms of what we do know, than vice versa. The second order equation becomes: 𝑑2 𝑦 𝑑𝑡 2 = −𝑎𝑦 = 𝑗 2 𝑎𝑦 This has solution: 𝑦 = 𝐴 cosh(𝑗√𝑎𝑡) + 𝐵 sinh(𝑗√𝑎𝑡) When t=0, y=y(0), so A=y(0) 𝑑𝑦 𝑑𝑥 = 𝑗√𝑎(𝐴 sinh(𝑗√𝑎𝑡) + 𝐵 cosh(𝑗√𝑎𝑡)) From which: 𝐵=− 𝑗 𝑑𝑦 ( ) √𝑎 𝑑𝑥 𝑡=0 The solution in terms of trigonometric functions is: 𝑦 = 𝑎 cos(√𝑎𝑡) + 𝑏 sin(√𝑎𝑡) 123 Cybernetics for Systems Engineers – DRAFT F From which: 𝑎 = 𝐴, 𝑏 = 𝑗𝐵 We can now write sine and cosine functions in terms of exponentials: cos 𝑥 = cosh 𝑗𝑥 = sin 𝑥 = sinh 𝑗𝑥 𝑗 = 𝑒 𝑗𝑥 +𝑒 −𝑗𝑥 2 𝑒 𝑗𝑥 −𝑒 −𝑗𝑥 2𝑗 When the roots of the characteristic equation contain square roots of negative numbers, it means that the solution involves an oscillatory function of time. In this case y will vary sinusoidally in time at a frequency given by √𝑎 radians per second, or 1 √𝑎 2𝜋 Hz. By introducing j, we can characterise all possible solutions as exponential functions. 6.20.1.1 Torsion Pendulum Figure 6-29: Torsion Pendulum A wheel of moment of inertia I is fixed rigidly to a stiff axle. A spiral spring is attached to the axle, and its free end is rigidly fixed so that it is kept stationary; it generates a restoring moment proportional to the angular deflection θ of the wheel. The equation of motion is: 𝑑2 𝜃 𝐼 𝑑𝑡 2 = −𝑘𝜃 Where k is the stiffness of the Torsion bar (the torque per unit deflection). The characteristic equation is: 𝑘 𝜆2 + 𝐼 = 0 124 Cybernetics for Systems Engineers – DRAFT F 1 𝑘 The wheel will oscillate back and forth at frequency 2𝜋 √ 𝐼 Hz. 6.20.2 Resonant Circuit Figure 6-30 : Inductance/Capacitor (LC) Circuit The circuit shown consists of an inductance L, connected in parallel with a capacitor C, a short pulse is applied across the combination. The voltage across the inductor is proportional to the rate of change of current through it. The rate of change of voltage across the capacitor is proportional to the current charging or discharging it. We have, for the inductance: 𝑑𝑖 𝑣 = −𝐿 𝑑𝑡 For the capacitor: 𝑑𝑣 𝑑𝑡 𝑖 =𝐶 Differentiating: 𝑑2 𝑣 𝑑𝑡 2 =− 1 𝑣 𝐿𝐶 In the absence of resistance, or radiation by means of an aerial, the circuit will ring indefinitely at a frequency 1 2𝜋√𝐿𝐶 Hz. 6.20.3 Pin Jointed Strut Consider a problem which baffled the Ancient Greeks. How slender can a column be before it buckles and collapses? 125 Cybernetics for Systems Engineers – DRAFT F At any point along the strut, the bending moment is given by the end load multiplied by the deflection. From engineer’s bending theory the elastic bending the elastic restoring is proportional to the radius of curvature, the moment at any point along the strut is given by: 𝑑2 𝑦 𝑀 = 𝐸𝐼 𝑑𝑥 2 + 𝑃𝑦 Where P is the end load. Figure 6-31 : Buckling Load of a Strut There will be a further moment arising because each particle of the beam is being accelerated laterally as the beam deflects. This manifests itself as a lateral loading on the beam. The loading is given by: 𝑤 = −𝜌 𝑑2 𝑦 𝑑𝑡 2 Where ρ is the mass per unit length of the beam. Considering the equilibrium of an element of beam, we find the rate of change of shear force Q is equal to the loading: 𝑤= 𝑑𝑄 𝑑𝑥 Also from equilibrium considerations, the shear force is equal to the rate of change of bending moment along the beam: 𝑑𝑀 𝑑𝑥 𝑑2 𝑀 𝑑𝑥 2 = 𝑄, 𝑑2 𝑥 = 𝑤 = −𝜌 𝑑𝑡 2 The equation of motion is: 𝑑2 𝑦 𝑑2 𝑑2 𝑦 𝜌 𝑑𝑡 2 + 𝑑𝑥 2 (𝐸𝐼 𝑑𝑥 2 + 𝑃𝑦) = 0 126 Cybernetics for Systems Engineers – DRAFT F For a strut which is free at the top, but prevented from lateral displacement at the base, the lateral deflection is expected to look like: 𝜋𝑥 ) 2𝐿 𝑦 = 𝑦𝑡𝑖𝑝 cos ( Where L is the length of the strut, and ytip is the lateral deflection at the tip. 𝜋 2 𝑑2 𝑦 𝑑𝑥 2 = −𝑦𝑡𝑖𝑝 (2𝐿) cos ( 2 𝐿 ) 𝑑2 𝑦 𝑑𝑡 2 = 𝜋𝑥 Also: 𝑑 2 𝑦𝑡𝑖𝑝 𝑑𝑡 2 𝜋𝑥 cos ( 2 𝐿 ) The equation of motion becomes: 𝑑 2 𝑦𝑡𝑖𝑝 𝑑𝑡 2 1 𝜋 2 𝜋 2 + 𝜌 (2𝐿) ((2𝐿) 𝐸𝐼 − 𝑃) 𝑦𝑡𝑖𝑝 = 0 The characteristic equation is: 1 𝜋 2 𝜋 2 𝜆2 + 𝜌 (2𝐿) ((2𝐿) 𝐸𝐼 − 𝑃) = 0 If the load is less than: 𝑃= 𝜋2 𝐸𝐼 4𝐿2 The strut will vibrate back and forth when subjected to a disturbance. A load greater than this will cause the tip deflection to diverge, and the strut will buckle. This is a more interesting stability problem because it determines circumstances where a stable system can become unstable. 6.20.4 Mass – Spring – Damper The examples considered so far containing just second and zeroeth order derivatives either diverge or exhibit a simple harmonic motion. Neither can really be described as ‘stable’ responses because there is no tendency for the dependent variable to decay to zero as in the first order case. The hot air balloon appeared to furnish an example in which height could be controlled without this perpetual oscillation above and below the desired height. Unfortunately, the governing equations were not linear, and do not form the basis of a general theory of control. However, they do illustrate the point that non-linear controllers can be quite satisfactory. In the physical examples above, we notice that strain energy in a spring is converted to kinetic energy in a mass and back again, or energy is stored alternately in the magnetic field of an inductance and the electric field between the capacitor’s plates. The result is a system which oscillates indefinitely. Such systems have a repeatable period of oscillation and form the basis of a clock. 127 Cybernetics for Systems Engineers – DRAFT F However we would not expect a real torsion pendulum to oscillate forever, and the oscillations of a resonant circuit are expected to decay rapidly in practice. The real world contains friction and electrical resistance which convert ordered energy, such as electrical, kinetic or elastic into disordered energy (heat), which is then lost to the surroundings. It must be emphasised that not every system governed by a second order equation represents a transformation of ordered energy. However, the presence of dissipation in these examples indicates how we may cause a more general system to converge. If the wheel of the torsion pendulum were immersed in a viscous fluid, the oscillations would die out quickly. A reasonable model for viscous drag sets the drag force or moment directly proportional to the linear or angular velocity. The torsion pendulum immersed in thick oil is now governed by an equation of the form: 𝑑2 𝜃 𝑑𝜃 𝐼 𝑑𝑡 2 = −𝑐 𝑑𝑡 − 𝑘𝜃 Where c is the viscous drag term, or the ‘damper’ term. The characteristic equation now takes the form: 𝑐 𝑘 𝜆2 + ( 𝐼 ) 𝜆 + ( 𝐼 ) = 0 From which: 1 2 𝑐 𝐼 𝑐 2 𝐼 𝑘 𝐼 𝜆 = [− ( ) ± √( ) − 4 ( )] The viscous damper provides feedback of angular velocity, whilst the stiffness provides feedback of deflection. If either introduces positive feedback, one of the roots will become positive, and the system will be unstable. If the oil is really thick, or the stiffness weak: 𝑐 2 𝑘 (𝐼 ) ≫ 4 ( 𝐼 ) The quantity under the square root is positive, so there are no complex roots. Both roots correspond to exponential decays. Unfortunately, the smaller root of the equation would tend to zero, implying an infinite settling time. Bearing in mind the fact that the exponential decay never actual reaches zero, it might be desirable to ensure the solution retains its oscillatory behaviour. In the absence of damping a second order system is characterised by a constant amplitude oscillation. The frequency of the oscillation is called the undamped natural frequency, denoted ω0 (omega nought). In this particular case: 𝜔02 = 128 𝑘 𝐼 Cybernetics for Systems Engineers – DRAFT F It is desirable to have a standard form of the equation for a second order system, so that our description of the behaviour applies to all systems, and not just this specific example. The coefficient of the damping term has dimensions of frequency, so it is reasonable to express it as proportional to the undamped natural frequency. 𝑐 𝐼 = 2𝜁𝜔0 Where ζ (zeta) is called the damping ratio. The characteristic equation has the general form: 𝜆2 + 2𝜁𝜔0 𝜆 + 𝜔02 = 0 This is a true system representation, as it deals with behaviour independently of the actual hardware exhibiting the behaviour. Such self-corrective behaviour is to be expected at any system level. It is not, as I have heard all too often from the ignorant and innumerate, a ‘detail of implementation’. The roots of the characteristic equation in terms of the undamped natural frequency and damping ratio become: 𝜆 = (−𝜁 ± 𝑗√1 − 𝜁 2 )𝜔0 The explicit solution is of the form: 𝜃 = 𝐴𝑒 −𝜁𝜔0 𝑡 sin(√1 − 𝜁 2 𝜔0 𝑡) + 𝐵𝑒 −𝜁𝜔0 𝑡 cos(√1 − 𝜁 2 𝜔0 𝑡) The sinusoidal response is multiplied by an exponential decay, so that its amplitude reduces exponentially with time. Figure 6-32 : Damping of 2nd Order System - 0.7 Seems Optimum 129 Cybernetics for Systems Engineers – DRAFT F The value of damping coefficient which appears to yield the shortest settling time is about 0.7. Actually, when the roots of the characteristic equation are equal, the solution ceases to be an exponential function, real, complex or otherwise, but introduces terms of the form: 𝑡 𝑛 𝑒 𝛽𝑡 Where β is the value of the root, and n is the number of times it is repeated. However, negative real parts still ensures the solution is stable. Although the form of the solution changes, the characteristic equation remains a valid means of assessing stability. The equation for the resonant circuit with a resistor in series with the inductor takes exactly the same form. 6.20.5 Higher Order Systems We can tell all we need to know about the nature of the solution of a linear homogenous equation from its characteristic equation. If we consider a third order system, we should aim to factorise the characteristic equation it into a quadratic and a simple factor, i.e: (𝜆2 + 𝑎𝜆 + 𝑏 )(𝜆 + 𝑐) = 𝜆3 + (𝑎 + 𝑐)𝜆2 + (𝑎𝑐 + 𝑏)𝜆 + 𝑏𝑐 If the coefficients of the cubic are real, then a,b and c are real. So the simple root must be real. This reasoning may be extended to a polynomial of any order, so it follows that any nth order characteristic equation may be factorised into simple and quadratic factors. The complete solution is the sum of the individual exponential functions. If we have the solution for a first order and a second order linear homogenous equation, we have the solution for any order, provided we can factorise the characteristic equation. The motion associated with each simple or quadratic factor is called a mode. We can define the behaviour required of a self correcting system in terms of the time constants of exponential decays, and the undamped natural frequency and damping ratios of second order system responses, without in any way concerning ourselves with how the behaviour is to be implemented in hardware. 6.21 Conclusions We have presented calculus in a restricted form, as applicable to the ideas behind the subject of automatic control. We started off by explaining integral and differential calculus, and derived a formula which applied to every mathematical function but the one most relevant to self-correcting systems, which is the one most relevant to the problem domain. At each stage we considered an example which introduced a feature not yet covered, and some care has been taken to ensure ideas are not used before they are explained, and illustrated with an example. This ground work is itself a useful introduction to the wider subject of calculus. 130 Cybernetics for Systems Engineers – DRAFT F After introducing methods of solving common types of differential equations, we discovered that in most cases the effort would be nugatory, as we are not usually interested in explicit solutions for specific systems, but in the general form the solution takes for entire classes of system. The information we sought, we discovered, can be found from an algebraic equation, with no need at all to worry about solving the governing differential equation explicitly. 131 Cybernetics for Systems Engineers – DRAFT F 7 More Advanced Calculus 7.1 Introduction This chapter does not introduce any new concepts, it is ‘advanced’ in the sense that solutions are seldom of the form of simple polynomials, but require more obscure functions. Perhaps ‘more obscure’ calculus might have been a better title. Usually we must resort to numerical solutions, to produce the ‘right’ answer, typically however with absolutely no understanding or insight into the associated problem. 7.2 Arc Length and Surface areas of Revolution We calculated the surface areas of a few special cases by calculating the difference in volume of two similar shapes, one enclosed inside the other. This was only possible in the special case of the sphere or the cone, where the thickness of the space between the two surfaces was uniform. In general, that is not the case. For a finite small change in the x coordinate(δx), the curve is very nearly a straight line, so the distance along the arc (δs) may be found by applying Pythagoras’ theorem: (𝛿𝑠)2 = (𝛿𝑥)2 + (𝛿𝑦)2 Where δy is the change in the dependent variable corresponding to the change in the independent variable (δx). This contains second order terms, which we usually set to zero, but as long as we are dealing with finite increments we may divide through by (δx)2: 𝛿𝑧 2 𝛿𝑦 2 (𝛿𝑥) = 1 + (𝛿𝑥 ) Now from our initial thoughts on the derivative, we know that: 𝛿𝑦 𝛿𝑥 → 𝑑𝑦 𝑑𝑥 As δx→0. Since 1 is a constant, it is legitimate to claim that: 𝛿𝑠 𝛿𝑥 𝑑𝑦 2 → √1 + (𝑑𝑥 ) As δx→0. The length of an arc may be approximated by summing finite lengths corresponding to equally spaced small intervals in the x coordinate: 𝑑𝑦 𝑆 = ∑𝑛𝑖=0 𝑖𝑤√1 + ((𝑑𝑥 ) 𝑥=𝑖𝑤 132 2 ) Cybernetics for Systems Engineers – DRAFT F Where w is the width of the x interval. As the width tends to zero, the sum becomes an integral, yielding an expression for the arc length, which is now exact: 2 𝑏 𝑑𝑦 𝑆 = ∫𝑎 √1 + (𝑑𝑥 ) 𝑑𝑥 Similarly, the surface area of a body of revolution may be found from: 2 𝑏 𝑑𝑦 𝑆 = 2𝜋 ∫𝑎 𝑦√1 + (𝑑𝑥 ) 𝑑𝑥 Also, the curvature at any point is given by: 1 𝑅 = 𝑑𝜃 𝑑𝑠 = 𝑑𝜃 𝑑𝑥 𝑑𝑥 𝑑𝑠 = 𝑑𝜃 𝑑𝑥 2 √1+(𝑑𝑦) 𝑑𝑥 Where θ is the angle the tangent to the curve makes with respect to the x axis, and R is the radius of curvature. Its tangent is equal to the slope of the curve. 𝑑𝑦 𝑑𝑥 = tan 𝜃 ; 𝑑2 𝑦 𝑑𝑥 2 𝑑𝜃 𝑑𝑦 2 𝑑𝜃 = sec 2 𝜃 𝑑𝑥 = (1 + (𝑑𝑥 ) ) 𝑑𝑥 The curvature is, therefore: 1 𝑅 = 𝑑2 𝑦 𝑑𝑥2 (1+( 3 𝑑𝑦 2 2 ) ) 𝑑𝑥 Typically, we deal with small values of slope (bending of beams, streamlines around a thin aerofoil), and the denominator is taken as unity. We can estimate the error in ignoring the denominator by calculating it for the maximum value of slope, and finding the effect on the particular problem solution. The presence of the square root complicates the finding of analytical solutions to this type of problem. Fortunately, we have already encountered functions which are related via square roots in this way. Recall the solution to Lanchester’s Equation, and the hot air balloon example, introduced the hyperbolic sine and cosine: sinh 𝑥 = 𝑒 𝑥 −𝑒 −𝑥 2 cosh 𝑥 = 𝑒 𝑥 +𝑒 −𝑥 2 From which: cosh2 𝑥 − sinh2 𝑥 = 1 as can be seen by direct substitution. Also, we have met their derivatives: 133 Cybernetics for Systems Engineers – DRAFT F 𝑑 (sinh 𝑥) 𝑑𝑥 = cosh 𝑥 𝑑 (cosh 𝑥) 𝑑𝑥 = sinh 𝑥 Another function which might help us is the tangent, which we met in our discussion of trigonometry: 1 + tan2 𝑥 = sec 2 𝑥 The tangent function has derivative: 𝑑 (tan 𝑥) 𝑑𝑥 = sec 2 𝑥 Similarly, whenever expressions of the form: √1 − 𝑥 2 Crop up in integrals, we might consider using the relationships between sine and cosine functions. 7.2.1 Why Bother? We may evaluate finite integrals using a computer, so why do we search for analytical solutions? The answer is that we do not seek numerical answers to specific problems, what is much more valuable is the understanding of how the solution is affected by the problem parameters. Very few problems are ever formulated as a once-through ‘sausage machine’ producing the ‘right’ answer. Most of our analysis is concerned with finding out how things are interrelated so that we have some means of improving on each iteration. Without this insight, we are shooting in the dark, with no means of deciding in which aspect of the system the change is required. For this reason an analytical approximation to the real world problem is worth a thousand handlecranked computer generated numbers, and the months spent getting largely irrelevant code to work. One of the tragedies of our age is the widespread failure to grasp the difference between data and useful information. We seek the inter-relationships between facts, because this is what constitutes knowledge. Facts on their own are akin to memorising the telephone directory – requiring much effort but of little use to man or beast. 7.3 The Catenary A heavy chain supported at either end hangs in a curve in the vertical plane, and curiosity motivates us to wonder what shape the curve is (after all, we are engineers). If the question doesn’t interest you, may I suggest a career in administration or accountancy, where an under-developed imagination might prove advantageous? 134 Cybernetics for Systems Engineers – DRAFT F A short element of the chain is in equilibrium between the change in the vertical component of tension in the chain. If the weight per unit length of the chain is ρ N m-1, the weight of a short element is ρδs, where δs is the length of the element. If the slope of the chain at this element is θ, the equilibrium of the element requires: (T + δT)sin(𝜃 + 𝛿𝜃) − 𝑇 sin 𝜃 = 𝜌𝛿𝑠 Also, the element must be in equilibrium horizontally: (𝑇 + 𝛿𝑇) cos(𝜃 + 𝛿𝜃) − 𝑇 cos 𝜃 = 0 Where T is the tension in the chain. To first order of small quantities, the equilibrium equations become: 𝑇 cos 𝜃 𝛿𝜃 + 𝛿𝑇 sin 𝜃 = 𝜌𝛿𝑠 −𝑇 sin 𝜃 𝛿𝜃 + 𝛿𝑇 cos 𝜃 = 0 Eliminating the change in slope over the element we have an expression for the change in tension: 𝛿𝑇 = 𝜌 sin 𝜃 𝛿𝑠 135 Cybernetics for Systems Engineers – DRAFT F But: sin 𝜃 = 𝛿𝑦 𝛿𝑠 From which we deduce that: 𝑑𝑇 𝑑𝑦 =𝜌 𝑇 = 𝜌(𝑦 − 𝑦0 ) Where y0 is the length of chain whose weight is the same as the minimum tension in the chain. The shape of the curve is found from the equation for horizontal equilibrium: 𝜌𝛿𝑦 cos 𝜃 = 𝜌(𝑦 − 𝑦0 ) sin 𝜃 𝛿𝜃 𝛿𝑦 = (𝑦 − 𝑦0 ) tan 𝜃 𝛿𝜃 But: tan 𝜃 = 𝛿𝑦 𝛿𝑥 𝑑𝑦 𝑑𝑥 → From which: 𝑦𝛿𝜃 = 𝛿𝑥 Differentiating the equation for the gradient: 𝑑 (tan 𝜃) 𝑑𝑥 = sec 2 𝜃 𝑑𝜃 𝑑𝑥 = (1 + tan2 𝜃) 𝑑𝜃 𝑑𝑥 𝑑𝑦 2 𝑑𝜃 𝑑𝑥 𝑑𝑥 = (1 + ( ) ) = 𝑑2 𝑦 𝑑𝑥 2 Substituting: (𝑦 − 𝑦0 ) Let : 𝑑𝑦 𝑧 = 𝑑𝑥 ; 𝑑𝜃 𝑑𝑥 𝑦 = 𝑑2 𝑦 2 𝑑𝑦 2 (1+( ) ) 𝑑𝑥 𝑑𝑥 𝑑2 𝑦 𝑑𝑥 2 𝑑𝑧 𝑑𝑦 𝑑𝑧 =1 𝑑𝑧 1 𝑑 = 𝑑𝑥 = 𝑑𝑥 𝑑𝑦 = 𝑧 𝑑𝑦 = 2 𝑑𝑦 (𝑧 2 ) Substituting for z yields an equation for the slope in terms of the height above the lowest point: 𝑑 (𝑧 2 ) 𝑑𝑦 (1+𝑧 2 ) 0) = 2 (𝑦−𝑦 This may be integrated directly: 𝑧 2 𝑑𝑤 1+𝑤 ∫0 𝑦 = 2 ∫0 𝑑𝑦 (𝑦−𝑦0 ) The lower limits are zero on both sides because the slope is zero at the lowest point. 𝑦 ln(1 + 𝑧 2 ) = 2 ln (1 − 𝑦 ) 0 i.e. 136 𝑑𝑦 𝑑𝑥 𝑦 2 = √(1 − 𝑦 ) − 1 0 Cybernetics for Systems Engineers – DRAFT F Let: 𝑝 = ± (1 − 𝑦 ); 𝑦0 𝑑𝑝 𝑑𝑥 =∓ 1 𝑑𝑦 𝑦0 𝑑𝑥 With this substitution: 𝑑𝑝 −𝑦0 𝑑𝑥 = √𝑝2 − 1 If: 𝑑𝑝 𝑑𝑥 Hence: √𝑝2 − 1 = ± sinh 𝑎𝑥 𝑝 = cosh 𝑎𝑥 ; = ±𝑎 sinh 𝑎𝑥 1 𝑎=𝑦 0 ± (1 − 𝑦 ) 𝑦0 𝑥 𝑦0 = ± cosh ( ) For the chain to curve upwards at its centre, both the cosh and y must have the same sign. If p is taken as positive, and the right hand side negative, the shape of the suspended chain becomes: 𝑦 𝑦0 𝑥 𝑦0 = cosh ( ) − 1 Resolving the sign ambiguity in this way, y=0 when x=0. This shape is known as a catenary. The scale factor (y0) is determined by the tension. 𝑑𝑇 𝑑𝑦 = 𝜌; ⇒ 𝑑𝑇 𝑑𝑥 𝑑𝑦 𝑑𝑇 𝑥 = 𝑑𝑥 𝑑𝑦 = ±𝜌 sinh (𝑦 ) 0 By symmetry, the tension must increase as we move away from the lowest point in the chain, so the rate of change with horizontal coordinate must be an even function, i.e: 𝑑𝑇 𝑑𝑥 𝑥 = 𝜌 |sinh (𝑦 )| 0 The integration is taken over the positive x range, but we note the solution applies to negative x as well: 𝑥 𝑇 = 𝜌𝑦0 cosh |𝑦 | + 𝑇0 0 Where T0 is the tension in the centre of the span. If the semi span of the gap is L, the tension will be a maximum there. As we are typically concerned with how strong the chain must be to cross a given gap, it seems sensible to base the calculation on the strength of the chain. 𝐿 𝑇𝑚𝑎𝑥 = 𝜌𝑦0 cosh (𝑦 ) + 𝑇0 0 𝐿 𝑇0 = 𝑇𝑚𝑎𝑥 − 𝜌𝑦0 cosh (𝑦 ) 0 137 Cybernetics for Systems Engineers – DRAFT F The chain can never go into compression, so the limiting value of y0 is found from the condition that the tension in the chain is zero at the centre of the span. 𝐿 𝑦0 cosh (𝑦 ) = 0 𝑇𝑚𝑎𝑥 𝜌 This would be solved iteratively for the case of interest, and the sag of the cable found from the catenary equation. Another possibility is for the sag in the chain to be restricted, as for a power line suspended between pylons, which requires a minimum ground clearance. The value of y0 is than found from the catenary equation, and the required cable strength is then found. Finally, the most likely question to be asked is ‘how much cable will be needed?’ To answer that, we need the arc length. We derived the general expression for the length of an arc in the previous section: 𝑑𝑦 2 𝐿 𝑆 = ∫−𝐿 √1 + (𝑑𝑥 ) 𝑑𝑥 In this case: 𝑑𝑦 𝑑𝑥 𝑥 𝑦0 = sinh ( ), so that: 𝐿 𝑥 𝐿 𝑥 𝑥 0 0 𝑆 = ∫−𝐿 √1 + sinh2 (𝑦 ) 𝑑𝑥 = ∫−𝐿 cosh (𝑦 ) 𝑑𝑥 = 𝑦0 [sinh (𝑦 )] 0 Or: 𝐿 −𝐿 𝐿 𝑆 = 2𝑦0 sinh (𝑦 ) 0 7.3.1 Comment Once we have familiarised ourselves with the exponential function in all its guises (trigonometric functions, hyperbolic functions and logarithms) we are well placed to produce approximate mathematical descriptions for a vast range of potential problems. The difficulty lies not in the mechanical process of evaluating integrals, but in deciding how to approximate a real world problem, and estimate what we think the domain of validity must be. In this case, some thought had to be applied to the resolution of sign ambiguities, which required an understanding of the problem in hand, rather than the abstract mathematics of a particular function. It is for this reason that the student is usually given a range of example problems to work through, in order to develop the skill of mapping real world problems, even those considered intractable, into an approximation which is adequate for the purpose in hand. What is missing in this approach is the overall objective which is being sought in formulating the mathematical representation in the first place. Usually one is asked to produce the ‘right’ answer, which tests the mechanical skills of manipulating formulas, but does not address the more important skill of finding a mathematical representation which is adequate for the current stage of study. 138 Cybernetics for Systems Engineers – DRAFT F Pure mathematicians are always presented with ‘given’ problems. The engineer, applied mathematician and analyst require the ability to formulate the problem in the first place and use features of the problem to assist in finding solutions. 7.4 About Suspension Bridges The catenary is an adequate approximation to a cable hanging between two support points such as a power cable between two pylons. It is tempting to claim that this represents the shape of the cables on a suspension bridge. However, the bridge differs from the suspended cable in that the weight of the support cables is insignificant compared with the weight of the deck. The supported weight depends on the horizontal component of position rather than the distance along the arc. The equilibrium of an element of support cable is more nearly (T + δT)sin(𝜃 + 𝛿𝜃) − 𝑇 sin 𝜃 = 𝜌𝛿𝑥 As before, the element must be in equilibrium horizontally: (𝑇 + 𝛿𝑇) cos(𝜃 + 𝛿𝜃) − 𝑇 cos 𝜃 = 0 This assumes that the cables are close enough together to effectively transfer the weight of the deck continuously to the cables. This is obviously an approximation, but it would be a pretty lousy design if this were not the case. The equilibrium equations, expanded to quantities of first order, become: 𝛿𝑇 sin 𝜃 + 𝑇 cos 𝜃 𝛿𝜃 = 𝜌𝛿𝑥 𝛿𝑇 cos 𝜃 − 𝑇 sin 𝜃 𝛿𝜃 = 0 Eliminating the increment in the slope yields an equation in for the tension: 𝑑𝑇 𝑑𝑥 139 = 𝜌 sin 𝜃 Cybernetics for Systems Engineers – DRAFT F Eliminating the change in tension: 𝑇 𝑑𝜃 𝑑𝑥 = 𝜌 cos 𝜃 Eliminating the increment in horizontal position: 1 𝑑𝑇 𝑇 𝑑𝜃 = tan 𝜃 This is integrated using the substitution: 𝑧 = cos 𝜃 ; 𝑑𝑧 𝑑𝜃 = − sin 𝜃 𝜃 𝜃 sin 𝜃 cos 𝜃 ∫0 tan 𝜃 𝑑𝜃 = ∫0 cos 𝜃 𝑑𝑧 𝑑𝜃 = − ∫1 𝑧 = − ln(cos 𝜃) The left hand side of the equation becomes: 𝑇 𝑑𝑇 0 𝑇 𝑇 𝑇0 = ln ( ) ∫𝑇 So the variation of tension with slope is given by: 𝑇 𝑇 = cos0 𝜃 Differentiating with respect to x: 𝑑𝑇 𝑑𝑥 = 𝑇0 cos2 𝜃 𝑑𝑥 sin 𝜃 𝑑𝜃 𝑑𝜃 𝑑𝑥 = 𝜌 cos 2 𝜃 𝑇0 ⇒ sec 2 𝜃 𝑑𝜃 𝑑𝑥 = 𝜌 𝑇0 We recall that: tan 𝜃 = 𝑑𝑦 𝑑𝑥 Differentiating with respect to x: 𝑑𝜃 𝑑𝑥 sec 2 𝜃 𝑑2 𝑦 = 𝑑𝑥 2 The shape of the curve is therefore determined by: 𝑑2 𝑦 𝑑𝑥 2 𝜌 𝑇0 = Taking the reference point as the centre of the span, where x=y=dy/dx = 0: 𝜌 𝑦 = 2𝑇 𝑥 2 0 The tension is found from: tan 𝜃 = 140 𝑑𝑦 𝑑𝑥 = 𝜌 𝑥; 𝑇0 1 sec 𝜃 = √1 + tan2 𝜃 = 𝑇 √𝑇02 + 𝜌2 𝑥 2 0 Cybernetics for Systems Engineers – DRAFT F i.e: 𝑇 = √𝑇02 + 𝜌2 𝑥 2 The arc length is: 𝜌𝑥 2 𝑇0 𝐿 𝑆 = ∫−𝐿 √1 + ( ) 𝑑𝑥 In order to evaluate this, try the substitution: 𝜌𝑥 𝑇0 sinh 𝑧 = The integral becomes: 𝑆= 𝑇0 𝛼 ∫ cosh2 𝑧 𝜌 −𝛼 𝑑𝑧 𝜌𝐿 Where: 𝛼 = sinh−1 ( 𝑇 ) 0 There are several ways to evaluate the integral, when all else fails it is probably worth writing it as exponential functions: 𝑒 𝑧 +𝑒 −𝑧 2 ) 2 cosh2 𝑧 = ( = 𝑒 2𝑧 +2+𝑒 −2𝑧 4 Integrating: 𝜌 𝑆 = 4𝑇 [ 0 𝑆= Or: (𝑒 2𝑧 −𝑒 −2𝑧 ) 2 𝛼 + 2𝑧] −𝛼 𝜌 = 𝑇 (𝛼 + (𝑒 𝛼 +𝑒 −𝛼 ) (𝑒 𝛼 −𝑒 −𝛼 ) 0 2 2 ) 𝜌 𝜌𝐿 (sinh−1 ( ) + cosh 𝛼 sinh 𝛼) 𝑇0 𝑇0 𝜌 𝜌𝐿 0 0 𝜌 2 𝜌𝐿 2 𝑆 = 𝑇 sinh−1 ( 𝑇 ) + (𝑇 ) 𝐿√1 + ( 𝑇 ) 0 0 If your calculator doesn’t have hyperbolic functions, the inverse sinh may be written as a logarithm: 𝑦 = sinh−1 𝑥 ⇒ 𝑥 = sinh 𝑦 = 𝑒 𝑦 −𝑒 −𝑦 2 𝑒 2𝑦 − 2𝑥𝑒 𝑦 − 1 = 0 𝑒 𝑦 = 𝑥 ± √1 + 𝑥 2 𝑦 = ln(𝑥 ± √1 + 𝑥 2 ) 7.5 Arc Length of an Ellipse The circumference of a circle was found by differencing the areas of concentric circles to find the area of the annulus between them, and then reducing the width of the annulus. That method cannot be applied to the ellipse because the space between similar ellipses is not of uniform width. 141 Cybernetics for Systems Engineers – DRAFT F The equation for an ellipse is: 𝑥 2 𝑦 2 (𝑎) + (𝑏 ) = 1 Where a and b are the semi-axes lengths. Differentiating: 𝑑𝑦 𝑑𝑥 𝑏 2𝑥 𝑦 = − (𝑎) √1 + 𝑑𝑦 2 (𝑑𝑥 ) 𝑏 2 = − (𝑎) 𝑥 𝑥 2 𝑎 𝑏√1−( ) 𝑏 2 𝑥 2 𝑎 𝑎 𝑥 2 1−(1−( ) )( ) =√ 1−( ) 𝑎 We could evaluate this using mathematical functions called (believe it or not) elliptical integrals. Now there is a whole range of somewhat specialised mathematical functions, which emerged before the days of personal computers. These were calculated numerically and presented in tables. These functions are nothing like as ubiquitous as the exponential function and its related functions, so we would only really familiarise ourselves with them if they cropped up frequently in our current investigations. At the level the systems engineer works, the most we really need to know is the geometric shape of plots of these functions, and how the geometric features are affected by changing parameters. Our desire is to visualise the problem, and not necessarily solve explicitly for a specific case. What we need to know is how the problem changes as we alter values of parameters; under which circumstances it is sensitive, and when it is robust to small changes. When confronted with an apparently intractable problem, we have three courses of action: Look it up in a standard reference, such as Abromowitz and Stegun . Evaluate it numerically, using a fairly trivial code Find a tractable approximation which is ‘good enough’ The first is of value if we are interested in the function in its own right. If we find ourselves repeatedly encountering it in our current piece of work, it might be worthwhile becoming familiar with the background mathematics. The second is fine for a particular instance, but requires numerical results for specific cases to check the operation of the code. Also, writing code brings with it its own problems which are not relevant to the task in hand, and naive solution methods frequently generate inaccurate results. To quote a crude but relevant aphorism; ‘when you are up to your arse in alligators, you tend to forget that you are here to drain the swamp’. The third approach requires us to sit back and consider exactly what we are trying to achieve. We expect the form of the solution to be some function of the slenderness ratio of the ellipse multiplied by a representative length: 142 Cybernetics for Systems Engineers – DRAFT F 𝑏 𝑎 𝑆 = 𝑎𝑘( ) We look for regions of the function k where we can get a reasonable approximate answer. For example, we know that when b/a = 1, the ellipse becomes a circle, and k=2π, and as b/a →0, k→4. From the nature of the problem, we know that the arc length cannot lie outside these bounds. Two points are insufficient to sketch a curve, but we could, as a first approximation, draw a straight line between these points. However, this isn’t much use without some estimate of the error in the approximation. Intuition would tell us that the error can’t be too great, but then intuition would soon have us burning witches and sacrificing virgins to the Moon Goddess. We know that the arc length cannot be greater than that of the enclosing rectangle, for which: 𝑏 𝑆 = 4(𝑎 + 𝑏) = 4𝑎 (1 + 𝑎) This is a reasonable estimate for slender ellipses, but becomes ever more inaccurate as the ellipse tends to a circle. Also, the arc length is not expected to be shorter that the hypotenuse of the triangle drawn between the intersection s with the x and y axes: 𝑏 2 𝑎 𝑆 = 4√𝑎2 + 𝑏 2 = 4𝑎√1 + ( ) As before, the estimate error is a maximum for the quadrant of the circle. Using the straight line estimate, we know the estimate could be up to 27% too great or 10% too small. 143 Cybernetics for Systems Engineers – DRAFT F We cannot say whether this is adequate or not, unless we know how accurate our analysis must be. At a first iteration, straight line approximations are desirable, as they simplify the analysis considerably. However, without knowledge of the error implicit in the simplification, the analysis ceases to be ‘top-level’ but merely self-deception, and not worth the paper it is written on. Arc Length Error Bounds 2.5 Arc Length 2 1.5 1 0.5 0 0 0.2 0.4 0.6 0.8 1 Slenderness The upper and lower error bounds with the box (blue) and lozenge (red) approximations to the ellipse. We know that the upper bound must be π/2 for a single quadrant, so that an approximation which exceeds this value must be invalid, regardless of how loose the tolerances on the calculation might be. Taking the arithmetic mean f the two estimates yields an improved estimate: 𝑏 2 𝑏 𝑆 = 2𝑎 (1 + 𝑎 + √1 + (𝑎) ) For which the error will not exceed: 𝑏 𝑏 2 ∆𝑆 = 2𝑎 (1 + 𝑎 − √1 + (𝑎) ) This may be up to 17%. 7.5.1 Towards a better solution The estimate is at its worst s the ellipse approaches the circular quadrant, so evidently a better estimate in this region is: 𝑆 = 2𝜋 144 Cybernetics for Systems Engineers – DRAFT F In order to reduce the clutter a bit, let: 𝑥 𝑏 𝜂 = 𝑎, and 𝛽 = 𝑎 The arc length of the ellipse is given by: 𝑏 2 𝑆= 𝑥 2 𝑎 1−(1−(𝑎) )(𝑎) 4 ∫0 √ 1−𝑥 2 𝑑𝑥 The equation becomes: 1 𝑆 = 4𝑎𝑘(𝛽) = 4𝑎 ∫0 √ 1−(1−𝛽 2 )𝜂2 𝑑𝜂 1−𝜂2 For values of β approaching unity, we the numerator square root may be approximated as a Taylor series: 1 √1 − (1 − 𝛽 2 )𝜂 2 ≈ 1 − 2 (1 − 𝛽 2 )𝜂 2 1 1−(1−𝛽 2 )𝜂2 𝑑𝜂 1−𝜂2 ∫0 √ 1 ≈ ∫0 1 1 √1−𝜂2 1 (1−𝛽 2 )𝜂2 𝑑𝜂 − 2 ∫0 √1−𝜂2 𝑑𝜂 The first term evaluates to the circular quadrant solution, so the second term may be considered an estimate of the error in using the circular quadrant estimate. The presence of the square root indicates a sine function substitution might be appropriate: 𝜂 = sin 𝜃 𝑑𝜂 𝑑𝜃 = cos 𝜃 1 (1−𝛽2 )𝜂2 ∫0 √1−𝜂2 𝜋 𝑑𝜂 = (1 − 𝛽 2 ) ∫02 sin2 𝜃 cos 𝜃 cos 𝜃 𝑑𝜃 In order to deal with the square of the sine, we recall from the formula for the cosine of the sum of two angles: cos 2𝜃 = cos 2 𝜃 − sin2 𝜃 = 1 − 2 sin2 𝜃 It follows that: 𝜋 ∫02 sin2 𝜃 𝑑𝜃 = 𝜋 1 2 ∫ (1 − 2 0 1 𝜋 cos 2𝜃) = 2 [𝜃 − sin 2𝜃]02 = 𝜋 4 The error in using the circular arc estimate is about: 𝜋 ∆𝑆 ≈ 2 𝑎(1 − 𝛽 2 ) We can claim that the circular arc approximation is superior to that of our average of the box and lozenge approximations: 145 Cybernetics for Systems Engineers – DRAFT F ∆𝑆 ≈ 2𝑎(1 + 𝛽 − √1 + 𝛽 2 ) There exists a value of β for which the two are equal, above which it is better to use the circular arc approximation. Comparison of Errors 30 25 error % 20 15 10 5 0 0 0.2 0.4 0.6 0.8 1 slenderness Comparing the two errors it is evident that by taking the circular arc approximation at values above β=0.62, the error is unlikely to exceed 15%. By implication we could continue this process of taking more terms in the series expansion to get a more accurate estimate of the arc length. However, the astute reader will have noticed that we have assumed that the sum of all the terms which we have ignored may be neglected. To see that this is not always justified, consider an expansion which could be used in the vicinity of β=0: 1 1−(1−𝛽 2 )𝜂2 𝑑𝜂 1−𝜂2 ∫0 √ 𝛽2 𝜂2 1 1 𝛽2 1 = ∫0 √1 + 1−𝜂2 𝑑𝜂 = √1 − 𝛽 2 ∫0 √1 + (1−𝛽2 ) 1−𝜂2 𝑑𝜂 Expanding the square root: 𝛽2 1 1 1 1 𝛽2 1 ∫0 √1 + (1−𝛽2 ) 1−𝜂2 𝑑𝜂 ≈ ∫0 1 − 2 (1−𝛽2 ) 1−𝜂2 𝑑𝜂 Integrating the second term: 1 𝑑𝜂 1−𝜂2 ∫0 = 1 1 1 ∫ ( 2 0 1−𝜂 1 + 1+𝜂) 𝑑𝜂 = 1 1+𝜂 1 ( )] [ln 2 1−𝜂 0 This evaluates to infinity, which even to the sloppiest standards, cannot be considered small. 146 Cybernetics for Systems Engineers – DRAFT F 7.5.2 About Errors This is as far as we shall go with the elliptical arc problem. We can reduce the error by taking more terms, but that is just mechanical algebraic manipulation (the alligators and swamp), which adds nothing to the approach employed. It serves as an example to illustrate what the engineer does when confronting an apparently intractable problem. The mathematician seeks an ‘exact’ solution for what is itself an approximation to physical phenomena, because mathematicians do not solve problems within any specific context. Historically, these exact solutions have proved surprisingly close fits to real world behaviour. The medicine man uses approximations, even if they are little better than guesses, without regard to the error introduced, and can therefore ‘prove’ whatever the end user wishes to hear. This appeals to legal minds which all too often dominate corporate thinking, and which seek to ‘prove’ a particular result to be true rather than find what the actual truth is. The engineer assumes from the outset that most of the real world problems are messy, and only approximate formulations are possible, especially at the initial stages of system analysis. It is at this stage where the errors in analysis have greatest effect, where an early misjudgement will steer the project in a catastrophically ill-conceived direction, that the uncertainties in the analysis, and their causes, must be fully understood. All too often we flow down precise system specifications from guesses and speculations disguised as analysis. When presented with an apparently intractable problem we seek an approximation which is ‘good enough’. What is meant by ‘good enough’ depends strongly on the context in which the analysis is being conducted. If it is for part of the system which is not critical to overall performance, the tolerances can be wide. Critical elements may need to be specified to much tighter tolerances. System engineering analysis is concerned primarily with establishing the relationships between component tolerances and overall system behaviour, without which attempts to optimise or balance the design are little better than guesses. Without this analysis, it is impossible to decide which are critical issues, and which do not matter. It is fair to say that most of the neat analytical solutions were found years ago, so that in his/her professional life the engineer is unlikely to meet problems which have a straightforward analytical solution. However, there exists a standard solution of some form which is adequate for practically all apparently intractable problems, but it is necessary to be aware of how much additional error the approximation introduces into the analysis. Ultimately, when the aircraft crashes or the bridge collapses, it is the engineer who must attend the enquiry. We must be satisfied that the consequences of the worst combination of tolerances are understood long before the product enters service. 7.6 Convergence We know our results are wrong, that’s life, but we must know how wrong they are. One common method of approximation is the series expansion of a function into more familiar functions. For this to be possible we must be certain that we have taken sufficient terms, and that accuracy does actually improve as we take more terms. 147 Cybernetics for Systems Engineers – DRAFT F In addition, we need to be sure that the sum of all the terms that have been ignored is no greater than the final term of the approximation. If we take a strip of paper and cut it in half, take one of the halves and cut it in half, and repeat the process, it is evident that no matter how many times we cut the remaining piece in half, the sum of all the pieces cannot exceed the original strip. In other words we know that the geometric series: 1 𝑆 = ∑𝑁 𝑖=0 2𝑖 cannot exceed unity, no matter how large N becomes. Furthermore: ∑𝑁 𝑛 1 2𝑖 ≤ 1 2𝑛 The argument would still be valid if we cut the strip at 1/3rd of its length. Indeed, if we chose a different fraction at each stage, the pieces still could not exceed the original strip nor could the remainder at any stage exceed the previous partial strip. The ratio of consecutive strips need not be constant, provided the ratio of any term to its preceding term has magnitude less than 1 for all terms, the series will be convergent, and we are justified in claiming that the error is no greater than the magnitude of the final term of our approximation. This reasoning indicates that if the ratio of the magnitudes of consecutive terms is less than unity for all terms, the series will be convergent. 148 Cybernetics for Systems Engineers – DRAFT F In the elliptical arc example the expansion is: 1 1−(1−𝛽 2 )𝜂2 𝑑𝜂 1−𝜂2 1 ≈ ∫0 ∫0 √ 1 1 √1−𝜂2 (1 − 2 (1 − 𝛽 2 )𝜂2 ⋯ ) 𝑑 𝜂 The ith term is: 1 1 3 1 (− 2) (− 2) ⋯ (2 2 1 1 − (𝑖 − 1)) 𝑖! ∫0 (1 − 𝛽 2 )𝑖 𝜂2𝑖 √1−𝜂2 𝑑𝜂 In order to proceed we must evaluate the integral: 1 ∫0 𝜋 𝜂2𝑖 √1−𝜂2 𝜋 𝑑𝜂 = ∫02 (sin2 𝜃)𝑖 𝑑𝜃 = ∫02 sin2i 𝜃 𝑑𝜃 (this uses the sine function substitution, which we have already met). We don’t need to evaluate the integral to prove convergence. We know that sinθ will lie in the range 0 to 1 over the limits of the integral, so the integral itself cannot evaluate to greater than π/2. Also, as i increases, the area under the curve reduces. The ratio of the integrals in successive terms is therefore less than unity. The ratio of the magnitudes of consecutive terms is therefore less than: 1 2 ( −𝑖)(1−𝛽 2 ) 𝑟=| 𝑖 | As i→∞, r →(1-β2), which is less than unity, so we are justified in treating the expansion as convergent. 7.7 Concluding Comment The mathematically inclined may not find this intellectually satisfying. It is so easy to become distracted from our primary quest by distractions, such as spending too long trying to derive an ‘exact’ solution to an inexact problem. We must recognise the fact that such results are rare in the real world, and so the engineer uses the few analytical solutions derived in the textbooks to provide hints as to the form the best approximation may take. By the same token it is inexcusable to employ an approximation without some understanding of the possible errors introduced by it. Intuition and guesswork are not good enough, although in recent years these appear to be preferred to proper engineering analysis. Most of our effort is rightly spent deciding how much we can trust our results, rather than seeking results which are absolute truth. It is better to find an appropriate approximation and know the truth within defined error bounds, than it is to insist on an absolutely correct answer, and learn nothing about the truth. It should be clear that few people are more aware of the limitations of theoretical models than scientists and engineers. Nobody expects absolute correspondence with reality, but we must know the error bounds on results produced. Without such error bounds the results are useless. 149 Cybernetics for Systems Engineers – DRAFT F I find it amusing how much authority and kudos is so often ascribed to cost benefits analysis whose variability in the parameters swamps any marginal differences between the options. The main merit of this approach is it tells the sponsoring authority what they want to hear, whilst exposing dissenters to a largely innumerate public as irrational and ignorant. This ploy has been used by economists for years. What is less amusing, is the tendency for important policy decisions to be based on such pseudoscientific nonsense. 150 Cybernetics for Systems Engineers – DRAFT F 8 The Calculus of Variations When we first considered using calculus to find maxima and minima, the function whose turning points we sought was ‘given’. We could solve problems such as finding the best proportions of a given shape to enclose the maximum area, but what we couldn’t do was answer the more fundamental question of what shape encloses the greatest area with a given perimeter? Before we consider that problem, let us address a simpler problem; what is the shortest path between two points on a plane? In other words, find a function y(x) which minimises: 2 𝑏 𝑑𝑦 𝑆 = ∫𝑎 √1 + (𝑑𝑥 ) 𝑑𝑥 The approach is the same as we adopted in all other calculus problems, only instead of incrementing a variable by a small quantity, we modify a function by adding to it another function which is continuous and everywhere ‘small’ compared with the original function. Let the solution be y(x), a trial solution near this is y(x)+ε(x). 𝑌(𝑥) = 𝑦(𝑥) + 𝜀(𝑥) As ε(x) is everywhere close to y(x), and is assumed smooth, we are justified in claiming its derivatives are also small quantities at all points. The difference in arc length of the incremented function is given by 2 2 𝑏 𝑏 𝑑𝑌 𝑑𝑦 𝛿𝑆 = ∫𝑎 √1 + (𝑑𝑥) 𝑑𝑥 − ∫𝑎 √1 + (𝑑𝑥 ) 𝑑𝑥 The increment in the integral; δS is called the ‘variation’, hence the ‘calculus of variations’. Expanding the first term: 2 2 𝑏 𝑏 𝑑𝑦 𝑑𝜀 𝑑𝑦 ∫𝑎 √1 + (𝑑𝑥 + 𝑑𝑥) 𝑑𝑥 ≈ ∫𝑎 √1 + (𝑑𝑥 ) (1 + 2 𝑑𝑦 𝑑𝜀 𝑑𝑥 𝑑𝑥 2 √1+(𝑑𝑦) 𝑑𝑥 The variation becomes: 𝑏 𝛿𝑆 = ∫𝑎 𝑑𝑦 𝑑𝜀 𝑑𝑥 𝑑𝑥 √1+(𝑑𝑦) 2 𝑑𝑥 𝑑𝑥 Integrating by parts: 𝑏 𝛿𝑆 = [𝜀 𝑑𝑦 𝑑𝑥 √1+(𝑑𝑦) 𝑑𝑥 151 𝑏 𝑑 ] − ∫𝑎 𝜀 𝑑𝑥 ( 2 𝑎 𝑑𝑦 𝑑𝑥 √1+(𝑑𝑦) 𝑑𝑥 2 ) 𝑑𝑥 1 2 ) 𝑑𝑥 Cybernetics for Systems Engineers – DRAFT F The function ε(x) is zero at x=a and at x=b, so the first term is identically zero. Between the end points ε(x) has a finite value, so the variation can only be zero if: 𝑑 ( 𝑑𝑥 𝑑𝑦 𝑑𝑥 2 √1+(𝑑𝑦) )=0 𝑑𝑥 This has the solution that dy/dx is a constant. In other words the path is a straight line, as we all could have guessed. However, it is not the result itself which interests us, it is the means of deriving it which is of value, because it is evident that this may apply to more awkward problems. More generally, we wish to find a function of x which minimises an integral of the form: 𝑑𝑦 𝑑 2 𝑦 𝑏 𝐽 = ∫𝑎 𝐹 (𝑦. 𝑑𝑥 , 𝑑𝑥 2 ⋯ ) 𝑑𝑥 In order to avoid writer’s cramp, let us use a more compact notation: 𝑑𝑦 𝑑𝑥 𝑑2 𝑦 = 𝑦 ′ , 𝑑𝑥 2 = 𝑦 ′′ , etc. We follow the same procedure as for the shortest path problem: 𝑏 𝜕𝐹 𝜀 𝜕𝑦 𝛿𝐽 = ∫𝑎 ( + 𝜕𝐹 ′ 𝜀 ) 𝑑𝑥 𝜕𝑦 ′ The second term in the integrand is found by integrating by parts: 𝑏 𝜕𝐹 ′ 𝜀 𝜕𝑦 ′ ∫𝑎 𝑑𝑥 = [𝜀 𝜕𝐹 𝑏 ] 𝜕𝑦 ′ 𝑎 𝑏 𝑑 𝜕𝐹 ( ) 𝜀𝑑𝑥 𝑑𝑥 𝜕𝑦 ′ − ∫𝑎 As before, the increment is zero at either limit (the path must pass through these points), so we have, for the variation: 𝑏 𝜕𝐹 𝑑 𝜕𝐹 𝛿𝐽 = ∫𝑎 (𝜕𝑦 − 𝑑𝑥 (𝜕𝑦′ )) 𝜀𝑑𝑥 The function ε(x) is finite in the interval, so the variation can only be zero if: 𝜕𝐹 𝜕𝑦 𝑑 𝑑𝐹 − 𝑑𝑥 (𝑑𝑦 ′ ) = 0 This expression is known as the Euler-Lagrange Equation. 8.1.1 Application to Maximum Area The problem we set ourselves was to determine the shape of the contour which encloses the greatest area. The area under a curve is given by: 𝑏 𝐴 = ∫𝑎 𝑦 𝑑𝑥 This must be maximised subject to the constraint: 152 Cybernetics for Systems Engineers – DRAFT F 2 𝑏 𝑑𝑦 𝑠 = ∫𝑎 √1 + (𝑑𝑥 ) 𝑑𝑥 is fixed. This differs from the path between points as we have introduced a constraint into the problem. This is the more usual form of an optimisation problem. If we were solving this problem by trial and error, we would guess a function to find A, and see if it were greater or smaller than the previous guess, then we would see the effect of our guess on the arc length. If the area were greater, but the arc length increased, we would try something else. In effect, we would seek to maximise a combined expression which includes the constraint: 2 𝑏 𝑏 𝑑𝑦 𝐻 = ∫𝑎 𝑦 𝑑𝑥 + 𝜆 (𝑠 − ∫𝑎 √1 + (𝑑𝑥 ) 𝑑𝑥 ) Where λ is a weighting , usually called a Lagrange multiplier. An expression of this form, with the constraints adjoined with undetermined multipliers is called a Hamiltonian. Differentiating with respect to λ yields the constraint equation: 𝜕𝐻 𝜕𝜆 2 𝑏 𝑑𝑦 = 𝑠 − ∫𝑎 √1 + ( ) 𝑑𝑥 = 0 𝑑𝑥 We find the turning point in H by calculating the variation in: 𝑑𝑦 2 𝑏 𝐽 = ∫𝑎 (𝑦 − 𝜆√1 + (𝑑𝑥 ) ) 𝑑𝑥 This is given by the Euler-Lagrange Equation: 𝜕 (𝑦 𝜕𝑦 𝑑 𝜕 ( (𝑦 𝑑𝑥 𝜕𝑦 ′ − 𝜆√1 + (𝑦 ′ )2 ) − − 𝜆√1 + (𝑦 ′ )2 )) = 0 This yields a set of potential solutions, which may or may not meet the perimeter constraint. 𝑑 1 + 𝜆 𝑑𝑥 ( 𝑑𝑦 𝑑𝑥 √1+(𝑑𝑦) 2 )=0 𝑑𝑥 Integrating: 𝑑𝑦 𝑑𝑥 2 √1+(𝑑𝑦) 𝑑𝑥 𝑥 =𝑐−𝜆 Where c is a constant. 𝑑𝑦 2 𝑥 2 𝑑𝑦 2 (𝑑𝑥 ) = (𝑐 − 𝜆 ) (1 + (𝑑𝑥 ) ) 153 Cybernetics for Systems Engineers – DRAFT F i.e: 𝑑𝑦 𝑑𝑥 = 𝑥 𝜆 (𝑐− ) √1−(𝑐−𝑥) 2 𝜆 𝑦=∫ 𝑥 𝜆 (𝑐− ) √1−(𝑐−𝑥) 2 𝑑𝑥 𝜆 Let 𝑥 (𝑐 − 𝜆 ) = cos 𝜃 1 𝑑𝑥 𝜆 𝑑𝜃 = sin 𝜃 𝑥 2 𝑦 = ∫ 𝜆 cos θ 𝑑𝜃 = 𝜆 sin 𝜃 + 𝑑 = 𝜆√1 − (𝑐 − 𝜆 ) + 𝑑 Where d is a constant. The Hamiltonian is, therefore: 𝑥 2 𝜆 𝑑 𝜆 𝐻 = 𝜆 ∫ √1 − (𝑐 − ) + 𝑑𝑥 + 𝜆 (𝑠 − ∫ 1 √1−(𝑐−𝑥) 2 𝑑𝑥) 𝜆 Evidently, c and d determine the position of the figure, which is irrelevant to its shape, so we may set them both to zero. 𝐻 = ∫ √𝜆2 − 𝑥 2 𝑑𝑥 + 𝜆𝑠 − 𝜆2 ∫ 𝑑𝑥 √𝜆2 −𝑥 2 From the first integral it is evident that: 𝑦 = √𝜆2 − 𝑥 2 The curve is a circle of radius λ. But all we have shown is that a circle encloses either the minimum or the maximum area. The curve must be closed, otherwise it doesn’t form a boundary at all, so the perimeter is four times the arc length of a quadrant. The arc length becomes: 𝜆 𝑠 = 4 ∫0 𝜆 √𝜆2+𝑥2 𝑑𝑥 = 2𝜋𝜆 As expected. In order to show this is a maximum, we only need to show it not to be the minimum, by considering the area enclosed by any shape figure we choose which has the perimeter 2πλ, and show that its area is less than that of the circle. The obvious example is a square, which will have area: 𝜋 2 𝐴𝑠𝑞𝑢𝑎𝑟𝑒 = ( 2 ) 𝜆2 154 Cybernetics for Systems Engineers – DRAFT F Compared with a circle: 𝐴𝑐𝑖𝑟𝑐𝑙𝑒 = 𝜋𝜆2 Since the area of the square is less, the circle cannot be the figure enclosing the smallest area, ergo it encloses the largest. 8.1.2 Comment The calculus of variations is seldom taught at an undergraduate level to engineering students, not because it is intrinsically difficult, but because in practice it has very few applications. The problem does not need to become very complicated before it becomes intractable. Just about all the possible analytical solutions may be found somewhere in the literature. With the advent of high speed computers it was hoped that the analytical difficulties would go away, and methods based on calculus of variations would be used to optimise everything. This hope was soon dashed because the constraints introduce a combinatorial problem, which quickly exceeds the capacity of the fastest computers. There remain methods such as dynamic programming which apply variational methods to approximations of the actual constraint functions. These generally make assumptions about the nature of the function which is being maximised. As an example, the function may contain multiple peaks, yet if all we are doing is searching for a turning point, we have no way of knowing whether or ‘maximum’ is a local maximum, or the supremum, which we are seeking. There is something juvenile and naive about seeking ‘the best’. It pre-supposes we know what ‘the best’ actually is. Too often, the criteria which ‘obviously’ characterised the optimum solution are merely relegated to the dustbin of ‘it seemed a good idea at the time’. 155 Cybernetics for Systems Engineers – DRAFT F In order to apply optimisation techniques we must understand the problem we are dealing with thoroughly, and must assess the appropriateness and limitations of the algorithm we are using. These are the very features of the problem modern software packages hide from the end user. In this world of spin and hype, nobody is prepared to admit that their product has limitations. We try to look up ‘bounds of validity’ in the documentation, only to draw a complete blank. The manual waxes lyrical about the wonders one can achieve, yet the first serious problem causes it to generate garbage. The naive user can indeed use the tool to generate meaningless numbers, only to prove the old adage that a fool with a tool remains a fool. We study mathematics not for the pragmatic utility of the results, but because the subtlety of thought in the more elegant proofs teaches us how to think clearly and logically. The engineer arms himself with a set of standard solutions which characterise as wide a range of physical phenomena as possible, and uses them to get a best fit on the current problem presented. Consequently the true engineer can tackle problems originating in any domain. 156 Cybernetics for Systems Engineers – DRAFT F 9 Cybernetics and Automatic Control 9.1 Introduction The fundamental idea which emerges from Ross-Ashby’s work, which is presented in the early chapters of this book is the Principle of Requisite Entropy. Implicit in much of the literature on systems engineering is the assumption that a complex system is characterised by complexity which increases with component count and diversity of components. Such is the case if the methods are not based on cybernetics. Firstly, cybernetics doesn’t care how the behaviour of the components comes about. The same behavioural models can be made to represent a vast array of possible real world entities. In this way the diversity in the components is accommodated. Secondly, the purpose of adding components is to reduce the complexity of the behaviour of the system, not to increase it. If complexity does increase, we are not dealing with a system at all, merely a collection of parts in close formation. There may be interactions at the secondary and tertiary function level, but not at the primary function level. An approach which is not based on behaviours can only pre-judge the solution, since if we are restricted to organisation and structure, we must have physical entities to organise and arrange before we can commence. Indeed, the methods have difficulty identifying what constitutes primary function, and excessive emphasis on what colour the box should be appears to take precedence of whether the system will fulfil its primary function. The cybernetic view is that a system is defined by what it does, and behaviour can be specified mathematically completely independently of the physical objects which achieve it. Requisite entropy decrees that a system, or subsystem, is like a piece of a jig saw. It is specified completely by the shape of the hole it must fit into, and the matching up of the picture on its surface. 9.2 Implications for System Specification At some relatively late stage of development we may be able to define our system by its components and sub-systems, in general that is not how we begin. Real world systems are complex in the behavioural sense, and there is no one to one mapping of behaviours on to widgets, as appears to be the underlying assumption in software-oriented methods. We cannot form any kind of hierarchical decomposition without the premise that processes map to widgets. When dealing with World-class technology suppliers, the end user is not competent to specify anything. Rather than prescribe the system requirements directly, it is the end user’s responsibility to specify the environment with which the system interacts, and request the supplier to produce a list of costed options, which represent the best that the current state of the art can deliver, rather than prescribe actual performance. The Law of Requisite Entropy states that the potential design solutions, if they exist at all, will be dictated by the environment with which the system must interact. 157 Cybernetics for Systems Engineers – DRAFT F This is called concept study. Note it is done in support of the requirements, and not in the forlorn hope of searching for a technical solution to meet the end user’s wishes, and which may not exist. Once the customer has seen the options, and hopefully had the concept work peer reviewed preferably by a rival, or a competent consultancy, so that there is reasonable confidence in the figures, the next stage is feasibility. This costs more because it introduces mock ups, test rigs, breadboards and boiler plate prototypes, whilst concept will involve modelling and paper studies. Finally, if there appears a worthwhile solution, the process moves on to Project Definition, at which point we have requirements which both customer and supplier are prepared to sign up to. 9.3 On Feedback The finite state machine is often taken as the paradigm for the ‘higher level’ control option in a system, as usually its function is to specify parameters or operating points or select the lower level functions required at any time. The reason we began studying the finite state machine is it is easy to understand. Actual finite state machines have been around since the 18th Century in the form of clockwork automata. Since then, automatic gun loaders, capstan lathes, even engine valve gear emerged as examples. Implementing them in digital computers is truly a ’detail of implementation’ as far as cybernetics is concerned. Whether mechanical cams or linkages, hydraulic or pneumatic valves, or a modern computer is used, they are conceptually identical. The problem arises, when considering systems which are self-adapting, is the finite state machine is inherently ‘brittle’ in its operation, to the extent that it can only be used practically in a highly controlled environment. A robot used in a factory environment works well, repeating a sequence of operations which would bore a human being silly. If the cat gets in the way, it keeps on spraying regardless. Even a knowledge-based system, so beloved of the Artificial Intelligence community, can only select from a finite set of options. Provided the set which the environment presents is closed, that is fine. Unfortunately, the environment is not a closed system. Also, the Turing machine presents problems when we consider adaptation. As Roger Penrose pointed out in his book ‘the Emperor’s New Mind’, changing a single bit at random does not cause the code to change its function slightly but destroys it altogether. The program in the machine must change in its entirety or not at all, which indicates that the Turing machine is a poor model for intelligence evolving by natural selection. There may be an algorithm which can run on the Turing machine which can adapt through random mutation, but that would be an emulation of the adaptive process, not the process itself. Ross Ashby side-stepped the issue by adding the ‘vital state’ to the set of machine states. This provided the trigger for the adaptation process. He illustrated this with his ‘homeostat’. This is merely a threshold below the fatal condition, which gives the machine time to adapt. Essentially the two states constitute quantised levels in what is really a single state. 158 Cybernetics for Systems Engineers – DRAFT F The homeostat was an electronic circuit which modelled an organism interacting with the environment. It was possible to inject disturbances which could render the organism/environment system unstable. When the vital state moved out of its acceptable bounds, the ‘organism’ part of the machine selected resistors at random until the combination again became stable. Ross Ashby’s homeostat illustrated the power of feedback. The random selection process indicated that how the feedback is done really doesn’t matter, provided it is possible to determine when adaptation has taken place. Nor is it necessary for the organism to have exact knowledge of the environment. Provided its action results in cancellation of the disturbance, the organism doesn’t care why it works. The Principle of Requisite Entropy states that the environment throws entropy at the organism, and for adaptation to be possible, the organism must have sufficient internal entropy to destroy it. This can come about by having a pre-defined reaction to counter every conceivable action of the environment. However, Requisite Entropy tells us that if we can detect when the entropy has been reduced to zero, no further action need be taken. This latter strategy is called feedback, and unlike the former, it is effective even if the set of environmental actions is open. The adaptation could use knowledge based systems or fuzzy logic, genetic algorithms, neural nets, or as in the homeostats case, random selection, or any other ingenious technique which may occur to the reader. The feedback could serve to build up an internal model of the environment, in which responses could be tested, but the only way to adapt to an open environment is through feedback. Note feedback is not just the interconnection of two processes, it is specifically a means of correcting the state of the system in response to a disturbance to drive the system entropy to zero. Implicit in feedback is the assumption that the system is observed for long enough for the effect of the corrective action to become manifest, if we attempt to take corrective action in a shorter timescale than this, we run the gauntlet of instability. 9.4 About Intelligence The reader is warned that this section is philosophical in nature, and may be skipped unless such speculations are of interest. It contains the views of the author, which are not widely accepted at present, but it is left to the reader to judge their merits. My objective is to try and de-mystify ‘intelligence’, and put it on the same sound objective footing as any other phenomenon, such as time, length or temperature. Few subjects are as important, or as ill-defined as ‘intelligence’. The position adopted here is that if a property exists, there should be an operational procedure for measuring it and by implication units by which it can be quantified, otherwise it has no place in science. For example, if it were suggested that, say, length should be measured using different units and definitions depending on whether we are measuring human beings, animals or inanimate objects, we would have serious doubts whether ‘length’ was a real concept at all. Still worse, if we claimed that temperature caused by heating with a candle was a somehow inferior kind of temperature than heating with an electric element, we would quickly lose the attention of our audience. Yet, that is precisely how intelligence is described. 159 Cybernetics for Systems Engineers – DRAFT F The world of Artificial Intelligence defines intelligence in terms of specific mental processes which must be present. Thus adaptation must take place in a specific way to be ‘intelligent’. This is akin claiming we can only use a candle flame for our definition of temperature, all other sources are not ‘real’ heat. Our definition must be eclectic, in the sense that it does not pre-suppose any particular method of implementation. Most systems which show goal seeking behaviour, and which work, use the methods of automatic control, although the software engineer responsible for coding the algorithms usually takes the credit for them. Requisite variety furnishes us with a clue as to how we can describe this concept without polluting our science with academic narcissism and imprecision. In order to get an objective measure, it must apply to man, machines and animals equally well, and must not depend on any specific form of processing. Words like ‘thinking’ have no objective definition, so they are excluded from the outset as meaningless. Likewise we do not distinguish adaptive behaviour which arises from the spinal cord from that involving the brain. How adaptation is implemented has no relevance whatsoever to defining what it is. That one form of adaptation is less effective than another is understood, and that is why we must have a unit with which to measure it. We do not dispense with numbers less than say, ten, as not ‘big enough’ to be considered ‘proper numbers’, beneath our dignity to consider. Indeed, it is usually the examples characterised by the elementary values that give us insight into the phenomenon, from which we stand a chance of progressing. Starting with the grandiose objective of simulating human intelligence is little more than academic megalomania. Entropy is needed to destroy entropy. However, intelligence brings with it not only the destruction of entropy, but a timescale within which the entropy is reduced to zero. The intelligent subject is the one who solves the problem more quickly than the rest. This implies that we should measure intelligence in bits per second, characterising the rate at which entropy is destroyed. This implies that we should expect intelligence to emerge in interesting times. We should not be surprised that homo sapiens emerged during a time of climatic turmoil, whilst dinosaurs reigned stupid and supreme for millions of years in steady climatic conditions. It has been claimed that civilisation requires calm and stability to develop and survive. Actually, if the thesis of this section is correct, quiet comfortable senility is the last thing required for civilisation. War and natural disaster provides the motivation to strive to improve. Civilisations grow and expand to a point where external strife is brought to a standstill by military prowess and internal strife by law and order; the result is senility and collapse. Current technology has its roots in the two World Wars, The Space Race and the Cold War, we have reached the limits of performance of the technologies which emerged during these periods, so that products must be sold as fashion items, rather than for their pragmatic utility. At a time we desperately need the truly novel, all we have is tinkering. Indeed, the idea of novelty is so alien to the modern mind, that tinkering and flogging dead horses appear to be the only 160 Cybernetics for Systems Engineers – DRAFT F recognised approaches to engineering innovation. What little innovation there is, usually falls foul of the vanity of the self-important. As the Earth’s climate becomes more hostile to human existence, new technologies may become necessary which could be used equally well for survival in the hostile environments of other planets. Indeed, the Earth may cease to have any special merit as far as human habitation is concerned. Without challenges to develop intelligence, the result will be regression and eventual extinction. Exploration is essential to find new questions, and hopefully lead to new discoveries. When the current technology reaches saturation, it is so tempting to believe that modern science has answered all the important questions and further discovery is impossible. I’m sure the ancient Greeks thought the same about their science, practically all of which has now been shown to be so much nonsense. 9.5 Automatic Control Adaptation by feedback can use any method the ingenuity of the human mind can devise. Provided the principle of zeroing the entropy error is employed (i.e. the scheme is based on the principle of Requisite Entropy) it will have the ability to adapt to an unknown environment. The greatest systematic body of knowledge which applies this principle to practical systems is called automatic control. It deals with systems which evolve in time and can be approximated by linear differential, or difference, equations. The inherent robustness of feedback ensures that quite gross approximations can be tolerated in the formation of the equations describing the interaction of the organism with its surroundings. Most ‘systems engineering’ tends to deal with the object (usually called the ‘system’) and the environment as two separate entities. Control considers the interaction of the object with the environment as the ‘system’, the Law of Requisite Entropy explains why this is the valid approach. A control system may find itself in any start position with equal probability but converges in time to the reference values specified at the input. In other words it begins in a high entropy state and converges on to a zero entropy state. It destroys entropy in finite time so, according to the eclectic definition of the previous section, it exhibits intelligence. The principles which emerge from this study apply to all self correcting time evolving systems, although only the mathematically rigorous analysis permits us to find the upper bounds of potential performance. That these principles do not appear to apply to more general systems merely reflects the mediocrity of performance which is currently achieved with state space machine approaches. 161 Cybernetics for Systems Engineers – DRAFT F 10 Stability 10.1 Introduction Automatic control deals with systems that may be described by sets of linear homogenous differential, or difference, equations. In reality few systems meet this criterion, but the inherent robustness of feedback alleviates the effect of disparities between the equations used to design the controller, and the actual plant behaviour. Indeed, automatic control seeks to quantify this mismatch and determine bounds on practical system performance based on this uncertainty. The system described by the equations is usually called the ‘plant’. In a few cases, the approximation may be accurate, and the feedback controller may be augmented with the feed forward control characteristic of typical finite state machine implementations. One application for feedforward is in tracking a distant target with a narrow beam, high resolution sensor. Knowledge of the approximate position of the target can be used to reduce the duration of the initial search for it. The a priori knowledge on which the feedforward controller is based normally comes from the designer’s understanding of system behaviour. In principle, it could be acquired by an outer loop observing the long term behaviour of the basic system, this could be repeated in hierarchical fashion any number of times. The design methods are becoming ever more mechanised, to the extent that a system using currently available methods could characterise the plant using system identification processes, and design controllers, without human intervention. The potential for hierarchical control systems in which the states of one level form the parameters of the next level, has not been investigated extensively. Existing techniques merely extend the order of basic systems until they are overwhelmed by ill-conditioning of the associated matrices. There seems to be a failure to grasp the fundamental idea that what are variables to one level are parameters to the next. This inevitably must introduce non-linear effects, which are expected to introduce chaos, as well as instability. However, before considering potential future developments, we require a firm understanding of the basics of control theory, and the methods used. 10.2 Additional Background Starting from elementary algebra, chapters 5 through 8 derive the ideas from calculus needed to understand basic ideas in automatic control. After deriving the basic ideas behind calculus, we reach the conclusion that we don’t actually need it at all to examine the stability of a system. The dynamic behaviour can be analysed by considering the roots of the characteristic equation. 162 Cybernetics for Systems Engineers – DRAFT F Excessive emphasis on control itself is probably misplaced, as it is the formulation of an adequate description of the plant which is intrinsically more difficult, which is why my emphasis is on examples. The coefficients of the characteristic equation are determined by the plant parameters, so that most of the time we use them to furnish insight into why the basic system is ill-behaved. In most cases the fault may be corrected by selecting appropriate values of system parameters. The upper bound of system performance is usually dictated by stability problems. It is known that a slender ship can sail faster than a tub, but if too slender, it simply turns turtle. Trying to get trains to run fast by simply installing more power leads to uncontrolled swaying of the carriages, risking de-railing and damage to the track. Trying to minimise the material in a structure for reasons of economy or minimising weight introduces the risk of all manner of structural instabilities. It is reasonable to suppose that most systems are not really describable as linear homogenous differential equations. We can usually find approximations which characterise an operating region, with some form of higher level function detecting which operating region the system is in and changing the parameters accordingly. Before proceeding, we shall introduce notation which reduces writer’s cramp. We shall invariably be dealing with quantities which vary in time, so we will be considering terms of the form: 𝑑𝑥 𝑑 2 𝑥 𝑑 3 𝑥 , , ,⋯ 𝑑𝑡 𝑑𝑡 2 𝑑𝑡 3 It is common practice to use the more compact form: 𝑑𝑥 𝑑𝑡 = 𝑥̇ (read this as ‘x dot’). The number of dots above the variable is equal to the order of the differential. For example, the generic second order system: 𝑑2 𝑥 𝑑𝑡 2 𝑑𝑥 + 2𝜁𝜔0 𝑑𝑡 + 𝜔02 𝑥 = 𝑓(𝑡) Is written: 𝑥̈ + 2𝜁𝜔0 𝑥̇ + 𝜔02 𝑥 = 𝑓(𝑡) Before the latter part of the20th Century understanding of control amounted to studying how feedback affected the solution of this type of equation. More typically, it is convenient to describe the plant as a set of first order simultaneous differentia equations, which may be written and manipulated in matrix form. For that reason it is necessary to revise the basics of matrix algebra. 163 Cybernetics for Systems Engineers – DRAFT F 10.3 Matrix Algebra 10.3.1 Definition and Basic Operations Figure 10-1 : Illustration of Matrix Multiplication Suppose we have a list of standard components, which can be used in different combinations to produce a range of possible widgets. In order to find out how much each widget costs, we could list the price of each widget in a column in a table, and on each row we could enter the numbers of each component needed for each widget. We find the cost of parts for any particular widget by taking the terms in the row corresponding to it and multiplying each term by its corresponding cost entry in the component price column. If we imagine a set of several suppliers, each offering an attractive loyalty discount, to the extent that it is only worthwhile sourcing from a single supplier, we can add price list columns for each, and generate the cost of widgets made using each supplier’s components, by multiplying row by column term by term. The process amounts to multiplying one table (the component count per widget) by another table (the cost of components), to yield another table (the cost of parts for each widget). We could write this as a single expression which just encapsulates the general idea that we multiply cost of components by numbers to get component costs for each widget: 𝑊=𝑁𝐶 Where the terms W, N and C are the tables containing the data, they are usually called matrices (the single of matrices is matrix). 164 Cybernetics for Systems Engineers – DRAFT F If the prices change by P, we can calculate the new cost per widget by adding this increment on to the prices matrix and repeating our row by column multiplication. This is a term by term addition, so obeys the usual arithmetic laws of addition, in particular it doesn’t matter in which order the terms are taken: 𝐶+𝑃 =𝑃+𝐶 However, our row/column procedure, although written the same way we would represent multiplication by ordinary numbers (scalars), is not the same. Logic would dictate that this process should have its own operator, to avoid confusion with algebraic multiplication, which like addition, would be applied element by element. However matrix multiplication is written as algebraic multiplication, by convention, and a notation is used to distinguish matrices from scalars. Paradoxically, there are occasions when we do wish to multiply element by element, and a special operator is introduced to indicate this. When we write two matrices next to each other we indicate this row/column operation takes place. For this to be possible the number of elements in the rows of the first matrix must match the number of elements in the columns in the second. Note that matrices are rectangular, all rows of a matrix are the same length, all the columns are the same length. Elements containing zero are added to pad out rows and columns as necessary. If two matrices are rectangular, as is likely in the example, because it is distinctly unlikely that the number of suppliers would equal the number of items in their catalogues, it is obvious that the operation: 𝐶𝑁 =? is undefined as the matrices rows and columns mismatch. I defy you to attach a meaning to this matrix product, in the cost example. If the rows and columns are equal in length, the matrix is, quite reasonably, called ‘square’. Evidently, we can apply the matrix multiplication in either order. Call the matrices A and B. For the product AB, any element of the result matrix is given by the sum of the product of row elements of A with corresponding column elements of B. For the product BA, the row elements of B multiply the column elements of A. The result is quite different, depending on the order of the two matrices, or: 𝐴𝐵 ≠ 𝐵𝐴 10.3.2 Additional Operations There is no compelling reason why our component cost data should be organised in columns under the ‘supplier’ headings. The same data could be organised in rows. However in order to maintain consistent dimensions (in this case meaning compatible rows and columns, rather than fundamental units), the numbers per widget data would have to be written in columns, with ‘widget’ headings. A matrix which has its rows and columns interchanged is called the transpose of the original matrix. We shall use the superscript T to denote this operation. 165 Cybernetics for Systems Engineers – DRAFT F If we wish to calculate our component costs per widget matrix using these transposed matrices, we note that firstly the order of multiplication is reversed, and the result is the transpose of the product of the original matrices. i.e: (𝑁𝐶)𝑇 = 𝐶 𝑇 𝑁 𝑇 So the transpose of a product is equal to the product of the transposes of the original matrices taken in reverse order. A matrix which is equal to its own transpose is called ‘symmetrical’. 10.3.3 The Identity Matrix and the Matrix Inverse Having applied the label ‘multiplication’ to an operation which is nothing of the sort, we imply an inverse ‘division’ process exists. When we consider multiplying scalars we make use of the fact that there exists a particular number, which under multiplication does not affect the value. For scalars, that is the number one. The equivalent question is whether there exists a matrix which has this property under matrix multiplication. If the matrices are rectangular, the result has as many rows as the first matrix and as many columns as the second, so there is no matrix which will maintain the dimensions of either matrix, unless both matrices are square. It follows that we can only have an equivalent to a unity quantity for square matrices. Consider pre-multiplying this unity matrix by an arbitrary square matrix. To ensure the first column isn’t affected by this operation, the first element in each row must be multiplied by unity and the rest by zero. It follows that the first column of the unity matrix has one as its first entry, and zero for all the rest. Considering the second column of the first matrix , the second column of the unity matrix consists of zeros apart from the second element which must be unity, similarly for all the other columns. The ‘unity’ matrix consists of zero elements everywhere but on the lead diagonal, which consists of ones. Post multiplying this by an arbitrary matrix shows that it retains its ‘unity’ property for both pre- and post- multiplication. The unity matrix is called the ‘identity’ matrix to avoid confusion with a matrix whose elements are all unity. It is usually denoted by the symbol I, or In if it is necessary to specify the size of the matrix. Having found the matrix equivalent to the scalar unity, the next question is, given an arbitrary square matrix, does there exist a matrix such that the product of the two equals the identity matrix? If such a matrix exists we have the equivalent of an arithmetic reciprocal function, and hence a matrix equivalent of division. Evidently we cannot treat a matrix as a whole if we are trying to find out if an inverse exists. We are looking for relationships between the elements themselves. A set of simultaneous equations is a classic example of a problem which may be written more compactly in matrix form: 𝐴𝑥 = 𝑏 166 Cybernetics for Systems Engineers – DRAFT F Where A is n×n, and x and b are n×1, sometimes called n-vectors by analogy with 3-vectors used to describe quantities in 3D (Euclidean) space. A solution for x can only exist if A is square. If the inverse exists, the solution is found: 𝐴−1 𝐴𝑥 = 𝐴−1 𝑏 𝑥 = 𝐴−1 𝑏 Where the superscript -1 denotes a matrix inverse. The problem reduces to finding the circumstances for which the set of simultaneous equations cannot be solved. Evidently if any row or column consists entirely of zeros it can be removed from the matrix, as it will not affect the matrix multiplication. The result is a non-square matrix which has no inverse. More generally, if any row consists of a linear sum of two or more other rows, it would be reduced to a row of zeros by applying the row operations used to solve sets of simultaneous equations. It follows that the inverse requires all rows of the matrix to be linearly independent. A matrix which has no inverse is called singular. As an example, if we post multiply an n×1 matrix by its transpose, the result is an n×n matrix, but the rows cannot be linearly independent, as each row is a multiple of the first. In general, if we (post) multiply an n×m matrix (with m<n), by its transpose, the resulting matrix cannot have an inverse. A matrix of this form is called a dyadic. Note that the inverse of the inverse is our original matrix, which post multiplies the inverse. It follows that unlike general matrices, the inverse may pre- or post-multiply the matrix. The inverse of a product obeys the same rules as the transpose of a product. [𝐴𝐵]−1 = 𝐵−1 𝐴−1 Since: 𝐵−1 𝐴−1 [𝐴𝐵] = 𝐵−1 [𝐴−1 𝐴]𝐵 = 𝐵−1 𝐵 = 𝐼 When dealing with matrix algebra, we cannot assume a square matrix will always have an inverse. If a result depends on the existence of an inverse, this must be stated explicitly. The results presented in this section are the very basics of matrix algebra. Additional ideas will be introduced as necessary, in the context of realistic problems. 10.3.4 Checking for Linear Independence - Determinants Examining a matrix by eye and deciding if one row is the sum of multiples of other rows is an unreliable process and something more methodical is needed. Consider the pair of simultaneous equations: 𝑎𝑥1 + 𝑏𝑥2 = 𝛼 167 Cybernetics for Systems Engineers – DRAFT F 𝑐𝑥1 + 𝑑𝑥2 = 𝛽 In order to solve for x1, we multiply the first equation by d and the second by b and subtract one from the other: (𝑎𝑑 − 𝑏𝑐)𝑥1 = 𝛼𝑑 − 𝛽𝑏 Similarly, solving for x2; (𝑎𝑑 − 𝑏𝑐)𝑥2 = 𝛽𝑎 − 𝛼𝑐 Evidently, if (ad-bc) is zero, the equation set cannot be solved. Extending this idea to a 3rd order system of equations: 𝑎𝑥1 + 𝑏𝑥2 + 𝑐𝑥3 = 𝛼 𝑑𝑥1 + 𝑒𝑥2 + 𝑓𝑥3 = 𝛽 𝑔𝑥1 + ℎ𝑥2 + 𝑘𝑥3 = 𝛾 Note: we skipped i, which is usually used to denote an index. Eliminating x2 and x3, from the last two equations: 𝑒𝑥2 + 𝑓𝑥3 = 𝛽 − 𝑑𝑥1 ℎ𝑥2 + 𝑘𝑥3 = 𝛾 − 𝑔𝑥1 Using the result for the 2×2 matrix: 𝑥2 = 𝑘(𝛽−𝑑𝑥1 )−𝑓(𝛾−𝑔𝑥1 ) (𝑒𝑘−ℎ𝑓) 𝑥3 = 𝑒(𝛾−𝑔𝑥1 )−ℎ(𝛽−𝑑𝑥1 ) (𝑒𝑘−ℎ𝑓) Substituting for x2 and x3 in the first equation yields the solution for x1: (𝑎(𝑒𝑘 − ℎ𝑓) − 𝑏(𝑘𝑑 − 𝑓𝑔) + 𝑐(ℎ𝑑 − 𝑒𝑔))𝑥1 = 𝑐𝑜𝑛𝑠𝑡𝑎𝑛𝑡 The coefficient would be the same whichever pair of unknowns was eliminated. It is evident that repeated elimination of unknowns from any number of equations will generate a single coefficient of this form, which if zero indicates that the set of equations doesn’t have a solution. The quantity: ad-bc is conventionally written: 𝑎 | 𝑐 𝑏 | 𝑑 𝑎 And is called the determinant of matrix ( 𝑐 168 𝑏 ). 𝑑 Cybernetics for Systems Engineers – DRAFT F Since the set of equations can be solved by multiplying both sides by the matrix inverse, a determinant which evaluates to zero indicates that the inverse does not exist. The set of three equations yields a definition which may be extended to higher order matrices: 𝑎 |𝑑 𝑔 𝑏 𝑒 ℎ 𝑐 𝑓 | = 𝑎 |𝑒 ℎ 𝑘 𝑑 𝑓 |−𝑏| 𝑔 𝑘 𝑓 𝑑 |+𝑐| 𝑔 𝑘 𝑒 | ℎ This is the first row multiplied by the determinants of the matrices remaining after the row and column containing the corresponding row element have been removed. These reduced order matrices are called the minors. The sign depends on the indices of the row element. In this case the entry b is in row one, column two. If the sum of the row and column indices is odd, the term is negative. The same determinant can be evaluated from any row of the matrix. The determinant furnishes a convenient means of deciding whether a matrix has an inverse: If the determinant is zero, the inverse does not exist. 10.4 State Space Representation Since linear homogenous equations are characterised by feedback, they, not surprisingly, form the basis of feedback control. They may be written in matrix form by making each derivative a state variable. Consider the second order damped oscillator: 𝑥̈ + 2𝜁𝜔0 𝑥̇ + 𝜔02 𝑥 = 𝑓(𝑡) Let: 𝑢 = 𝑥̇ The equation may now be written: 𝑢̇ + 2𝜁𝜔0 𝑢 + 𝜔02 𝑥 = 𝑓(𝑡) This is now a pair of first order equations: 𝑢̇ = −2𝜁𝜔0 𝑢 − 𝜔02 𝑥 + 𝑓(𝑡) 𝑥̇ = 𝑢 This may be written in matrix form: 𝑢̇ ( ) = (−2𝜁𝜔0 𝑥̇ 1 −𝜔02 ) (𝑢) + (1) 𝑓(𝑡) 𝑥 0 0 Or more compactly: 𝑥̇ = 𝐴𝑥 + 𝐵𝑢 Where x is understood to be the two element vector of system states. u is the input to the system represented by the second order equation. A is called the ‘system’ matrix and B the control matrix. 169 Cybernetics for Systems Engineers – DRAFT F If we consider impulse responses, u=0, so the plant equation becomes: 𝑥̇ = 𝐴𝑥 Differentiating with respect to time: 𝑥̈ = 𝐴𝑥̇ = 𝐴2 𝑥 𝑑𝑛𝑥 𝑑𝑡 𝑛 = 𝐴𝑛 𝑥 This is the property of the exponential function. The above relationships imply that we are justified in defining the matrix exponential as the same Taylor expansion as that for the scalar exponential. 1 1 𝑒 𝐴𝑡 = 𝐼 + 𝐴𝑡 + 2! 𝐴2 𝑡 2 + 3! 𝐴3 𝑡 3 ⋯ Interestingly, although the matrix exponential is defined in this way, the individual elements of the matrix exponential are not necessarily exponential functions. Consider the case of uniformly accelerated motion. Denoting the acceleration f, the speed u and position z, the state vector is: 𝑓 𝑥 = (𝑢) 𝑧 The system matrix is; 0 𝐴 = (1 0 0 0 0 0) 1 0 0 0 𝐴 = (0 0 1 0 0 0 0 3 0), 𝐴 = (0 0 0 0 0 2 0 0) 0 Evidently, all higher powers of A are also zero. The solution is: 1 𝑓 (𝑢) = ( 𝑡2 𝑡 𝑧 2 0 1 0 𝑓(0) 0) (𝑢(0)) 𝑡 1 𝑧(0) This is the same as was derived by integration of the acceleration with respect to time. If A were a scalar, stability would be determined by whether it is positive or negative. What do we mean by ‘positive’ when applied to the matrix as a whole? 10.4.1 ‘Sign’ of a Matrix We cannot say much about the ‘sign’ of a matrix without some general definition of what we mean by ‘positive’. We can claim something is positive if it doesn’t change the sign of a positive quantity. 170 Cybernetics for Systems Engineers – DRAFT F If x is n×1, the product: 𝑥𝑇𝑥 Is the sum of the squares of the elements of x and is therefore positive. In general, the individual elements are not pure numbers, so this product would be dimensionally inconsistent; we cannot add cabbages to kings. It is understood that we always scale our state variables to avoid this difficulty. Implicit in the above product is a scaling of the original state variables: 𝑥 = 𝑄𝑧 Where Q is a usually diagonal matrix whose elements are the reciprocals of the maximum desirable values of the original state variables (z). If we multiply a system matrix A by x, the product: 𝐴𝑥 still has dimensions of 1/time, or frequency. We must multiply the product by an arbitrary positive time to convert it to a vector of pure numbers. This frequency scaling is understood, and as it doesn’t affect the result, is assumed to be unity. Quite often, the time used in the plant equation is itself scaled to some important time constant characterising the motion, so that this frequency scaling is unnecessary. Pre-multiplying the result by xT, the matrix becomes a weighting on the squares and products of the elements of x. By convention, the matrix is considered positive, if the product remains positive. i.e. if: 𝑥 𝑇 𝐴𝑥 𝑥𝑇𝑥 >0 The matrix A is considered positive. If the matrix were diagonal, some of the elements could be negative, and there would be values of the elements of x which would yield a negative result. The criterion must be strengthened, such that the expression is positive regardless of the values the elements of x may take. If this is the case, the matrix is called ‘positive definite’. If there exists values of x for which the result is zero, it is called ‘positive semi-definite’. Reversing the criterion introduces the idea of negative definite and negative semi-definite. If the sign depends on the values of x, the matrix is ‘indefinite’. 10.4.2 Eigenvalues and Eigenvectors The only type of matrix which we can characterise in this way is a diagonal matrix. Let: 𝑥 = 𝑇𝑦 171 Cybernetics for Systems Engineers – DRAFT F Where the rows of T are linearly independent. Each element of y is a linear combination of the elements of x, and vice versa. Furthermore, we do not wish to change the value of the denominator of the test function: 𝑥 𝑇 𝑥 = 𝑦 𝑇 𝑦 = 𝑥 𝑇 𝑇 𝑇 𝑇𝑥 So that: 𝑇 𝑇 = 𝑇 −1 As an example of a matrix of this type consider the 2×2 example: cos 𝜃 𝑇=( − sin 𝜃 sin 𝜃 ) cos 𝜃 Where θ is an angle. This orthogonality requirement restricts our discussion to symmetrical matrices, which is unfortunate because the system matrix is rarely symmetrical. We can interpret the rows as the projections of a pair of perpendicular axes which have been rotated with respect to the reference set. Higher order matrices of this form can be similarly considered as a set of n perpendicular axes which have been rotated with respect to the initial reference axes. Since the rows represent lines in nspace which are perpendicular to each other (whatever that means in spaces of dimension higher than 3), this type of matrix, whose inverse is equal to its transpose is called an ‘orthogonal’ matrix. It can be thought of as the rotation of an n-vector. The numerator of the test function is: 𝑥 𝑇 𝐴𝑥 = 𝑦 𝑇 𝑇 𝑇 𝐴𝑇𝑦 Taking any pairs of rows (r) of T, the orthogonality condition implies: 𝑟𝑖 𝑟𝑗𝑇 = { 1 𝑖=𝑗 0 𝑖≠𝑗 With a similar relationship for the columns. We shall assume that a rotation exists such that each row becomes aligned with its corresponding axis, so that it takes on a scalar value, rather than the vector of the arbitrarily oriented original matrix row. In other words, the transformation results in a diagonal matrix. The elements of this diagonal matrix are called the eigenvalues. Multiplying a vector by a matrix may be considered to consist of three stages firstly a rotation into the eigenvalue axis set, followed by component-wise scaling by the eigenvalues, followed by a rotation back into the original reference frame. This may be written as the relationship: 172 Cybernetics for Systems Engineers – DRAFT F Λ = 𝑇 𝑇 𝐴𝑇 Where Λ (lamda) is the diagonal matrix of eigenvalues. The rows of T define the orientation of the eigenvalue axes, and are called eigenvectors. As with the inverse, the process of finding the eigenvalues requires processing at the element level, so as far as matrix algebra is concerned, we assume they exist, and don’t concern ourselves with how to go about calculating them. Using this diagonalisation we can propose the following definitions: If all the eigenvalues are positive – the matrix is positive definite If the eigenvalues are positive or zero – the matrix is positive semi-definite If all the eigenvalues are negative – the matrix is negative definite If the eigenvalues are negative or zero – the matrix is negative definite. Eigenvalues can occur as complex conjugate pairs. When this occurs the elements of T are also complex, so that when calculating the weighted sum squares of x, the imaginary parts cancel. The above statements apply to the real part of the eigenvalue. The calculation of eigenvalues is covered in the following sections. For the special case of a symmetric matrix, the simplest test for positive definiteness is to expand the determinants of increasing order from the lead diagonal. If all are positive, the matrix is positive definite. 10.4.3 Calculation of Eigenvalues We can extend the idea to non-symmetrical matrices, such as typical system matrices, by noting that the change of variable: 𝑥 = 𝑇𝑦 Transforms the system equation into: 𝑇𝑦̇ = 𝐴𝑇𝑦 The new system matrix is: Λ = 𝑇 −1 𝐴𝑇 In the special case where A is symmetric, the transformation matrix T is orthogonal. In either case, we can set up an equation for each of the rows of A. 𝜆𝑖 𝑣𝑖 = 𝐴𝑣𝑖 Where λi is the diagonal element corresponding to the row of interest. It follows that the eigenvalues may be calculated for a general (square) matrix by solving: |𝐼𝜆 − 𝐴| = 0 173 Cybernetics for Systems Engineers – DRAFT F If the matrix is not symmetric, the eigenvectors will no longer be orthogonal. So we reach the conclusion that the condition for the matrix exponential eAt to be stable, is for A to be negative definite. This is curiously similar to the stability requirement for the roots of the characteristic equation. 10.5 Stability – a Different Perspective Consider a system characterised by a state variable x. We know that any even function of x can be arranged such that it will have positive values regardless of the value of x. For example: 𝑉 = 𝑥2 Will always have positive values. Its evolution in time is: 𝑉̇ = 2𝑥𝑥̇ If this is negative, the system will move back to the origin if displaced. As an example, if the speed is proportional to the sign of x: 𝑥̇ = −𝑈 𝑠𝑖𝑔𝑛(𝑥) ̇ 𝑉̇ = −2𝑈𝑥 𝑠𝑖𝑔𝑛(𝑥) = −2𝑈|𝑥| This is always negative, regardless of the value of x. We are mainly interested in the exponential case, with its feature of linear feedback. 𝑥̇ = −𝑎𝑥 𝑉̇ = −2𝑎𝑥 2 = −2𝑎𝑉 The square of the state variable is also governed by an exponential decay. For the multi-variable case, we may define a ‘length’ of the state vector: |𝑥| = √(𝑝1 𝑥1 )2 + (𝑝2 𝑥2 )2 ⋯ Where the xi are the individual elements of x and the pi are weightings needed to ensure the calculation is dimensionally consistent (in the units sense, not the matrix size sense). The values of the pi are frequently chosen as the reciprocals of the maximum values of the corresponding state variables. The function: 𝑉 = |𝑥|2 174 Cybernetics for Systems Engineers – DRAFT F Appears a reasonable analogue of the scalar example. It may be written: 𝑉 = 𝑥 𝑇 𝑃𝑥 In using the ‘length’ (usually called the ‘Euclidean norm’, because there are many metrics which could reasonably be described as the ‘length’). Actually, P doesn’t need to be diagonal, if it is positive definite, V will still be positive. Differentiating: 𝑉̇ = 𝑥̇ 𝑇 𝑃𝑥 + 𝑥 𝑇 𝑃𝑥̇ In order to ensure the system always drives itself to zero, we require, by analogy with the scalar case: 𝑉̇ = 2𝜆𝑉 Where λ (lamda) is a constant, which is negative for a stable system. We could use this criterion to determine the stability of non-linear systems. For the moment, we shall restrict the discussion to the linear case: 𝑥̇ = 𝐴𝑥 𝑉̇ = 𝑥 𝑇 𝐴𝑇 𝑃𝑥 + 𝑥 𝑇 𝑃𝐴𝑥 = 2𝑥 𝑇 𝜆𝑃𝑥 As x is not identically zero, we may write this as: 𝐴𝑇 𝑃 + 𝑃𝐴 + 𝑄 = 0 This equation can be solved for P. If Q is positive definite, and the equation yields a positive definite solution for P, the system is stable. Since we are dealing with symmetric matrices, this is a straightforward test. Note that the equation must be solved for P. There exists an infinite number of choices for P, but we would be lucky indeed if they all yielded a positive definite value for Q. The matrix equation for P is called Lyapunov’s Equation, and the method of demonstrating stability this way is called Liapunov’s direct method. It may be extended to non-linear systems. It may be written, in terms of 𝜆 : (𝐴𝑇 𝑃 − 𝜆𝑃)𝑇 + (𝐴𝑇 𝑃 − 𝜆𝑃) = 0 i.e: (𝐴𝑇 − 𝜆𝐼)𝑃 = 0 Now P is positive definite, so the equation requires the matrix (AT-λI) to be singular. Since I is obviously symmetrical, this is the same as requiring (λI-A) to be singular. The similarity transformation for A yields: 175 Cybernetics for Systems Engineers – DRAFT F 𝑇𝜆 − Λ𝑇 𝑖𝑠 𝑠𝑖𝑛𝑔𝑢𝑙𝑎𝑟 If λ is equal to one of the eigenvalues of A, the rows will cease to be linearly independent, and the matrix will be singular. There exist as many solutions as there are eigenvalues, and if any of the solutions has a positive real part, the system will be unstable. The test for linear independence is for the determinant of the matrix to be zero. Hence the eigenvalues are found from: |𝐼𝜆 − 𝐴| = 0 As we have seen. 10.5.1.1 Second Order and Higher Systems We have seen that the damped linear oscillator equation may be written in state space form: 𝑢̇ ( ) = (−2𝜁𝜔0 𝑧̇ 1 −𝜔02 ) (𝑢) 𝑧 0 The determinant is: |𝜆 + 2𝜁𝜔0 −1 𝜔02 | = 𝜆2 + 2𝜁𝜔 𝜆 + 𝜔2 = 0 0 0 𝜆 This is exactly the same as the characteristic equation. More generally, for an nth order equation: −𝑎𝑛−1 ⋯ −𝑎1 −𝑎0 1 0 ⋯ 0 𝑥̇ = ( ⋮ ⋮ ⋮ ⋮ )𝑥 0 0 1 0 The bottom left sub-matrix is the n-1 identity matrix, the top row consists of the coefficients of the order given by the subscript. The final column, apart from its first entry, consists of zeros. This form of the system matrix is called a companion form. More general matrices may be transformed into this form by a series of elementary row operations, amounting to a change of variable to a linear combination of the original state variables. Hence the determinant becomes: 𝜆𝑛 + 𝑎𝑛−1 𝜆𝑛−1 ⋯ 𝑎1 𝜆 + 𝑎0 = 0 This is the characteristic equation for an nth order linear homogenous equation. We conclude that the eigenvalues are the same as the roots of the characteristic equation. However, the matrix approach shows us that each root is associated with a specific direction. The plant equations are rarely derived in companion form, so that evaluating the determinant furnishes a more convenient method of deriving the characteristic equation than transforming first into companion form. 176 Cybernetics for Systems Engineers – DRAFT F 11 Examples 11.1 Introduction Before we consider using artificial feedback to improve the behaviour of a system, our first task is to find out why it misbehaves in the first place. In the vast majority of cases adequate behaviour is obtainable by changing the values of the system parameters (e.g. adding bilge keels, changing wheel profiles, or changing the wing dihedral angle, depending on context). In many organisations the relationship between the dynamics and the parameters of the physical components tends to fall between the mechanical, aeronautical or electronic design department and the department responsible for control system design. Identifying the fundamental reasons for the stability problems is all too often the responsibility of neither. We have considered a few stability examples in Chapters 5 to 8. This section will introduce a range of problems from different engineering domains. This is to illustrate the point that the control engineer is not a specialist in any single domain. The control engineer must be a polymath, with sufficient knowledge of all domains to be able to communicate with specialists in each on their own terms. 11.2 Vibrating String The earliest eigenvalue problem was the vibration of a stretched string. Pythagoras noted that the pitch of notes was determined by the length of the string, with harmonious sounds resulting from strings whose lengths bore a simple ratio, like 2:3. Figure 11-1 : Vibrating Stretched String If we consider a string under constant tension T which is large compared with any change in tension due to the vibration, the equation of motion of a short length δl of the string is found from equating the force applied to it the product of its mass and acceleration. If θ is the local slope of the string, The lateral force is given by: 177 Cybernetics for Systems Engineers – DRAFT F 𝐹 = 𝑇(sin(𝜃 + 𝛿𝜃) − sin 𝜃) The deflection of the string is small compared with its length, so the slope will be a small angle, so: 𝐹 = 𝑇 𝛿𝜃 If the mass per unit length of the string is ρ, the mass of the element is ρδl, so the equation of motion for the element of string is: 𝜌𝑦̈ = 𝑇 𝑑𝜃 𝑑𝑙 If the shape of the distorted string is given by the curve y=f(x), 𝑑𝑦 𝛿𝑙~𝛿𝑥, 𝜃~ 𝑑𝑥 → 𝑑𝜃 𝑑 2 𝑦 ~ 𝑑𝑥 2 𝑑𝑙 If the shape of the distortion is: 𝑥 𝑦 = 𝑦𝑚𝑥 sin (𝑛𝜋 𝐿 ) Where n is an integer and L is the length of the string, the curvature is given by: 𝑑2 𝑦 𝑑𝑥 2 = −𝑦𝑚𝑎𝑥 𝑛 2 𝜋2 𝑥 sin (𝑛𝜋 𝐿 ) 𝐿2 From which the equation of motion becomes: 𝑦̈𝑚𝑎𝑥 = −𝑛2 𝜋2 𝑇 𝑦 𝜌𝐿2 𝑚𝑎𝑥 Since n may have any value, there are infinite possibilities for the shape of the distorted string, In other words there is an infinite number of possible modes, all of different frequency. The noises made by a novice violinist illustrate this point all too clearly. However, the skilful player will only excite a set of modes which gives the instrument its characteristic tone colour. Since the modes are orthogonal, it is possible to excite some and not others. The simplest case is for the string to distort into a simple curve given by n=1. (The word ‘sine’ is derived from the Latin sinus, meaning curve). This mode of vibration is called the fundamental. The characteristic equation is: 𝜆2 + 𝜋2 𝑇 𝜌𝐿2 =0 The frequency of vibration is: 1 𝑇 𝑓 = 2𝐿 √𝜌 Hz A similar approach was used to find the buckling load of a column. A wide range of dynamic problems, involving vibration modes and critical dynamic loads, are addressed in a similar manner. As a simple extension to the above problem we could use the result from engineer’s bending theory to calculate the frequency of vibration of a beam: 178 Cybernetics for Systems Engineers – DRAFT F 𝜌 𝑑2 𝑦 𝑑𝑡 2 + 𝐸𝐼 𝑑4 𝑦 𝑑𝑥 4 =0 In cases like this, where deflection depends on time and position, we should use the partial 𝜕𝑦 derivative notation (𝜕𝑥 ), using the Russian letter ‘da’. Since the problem is separable in x and t, it doesn’t really make much difference which notation we use. Although computer codes exist for this class of problem, the user should be in a position to estimate the answer roughly before using the code. A fool with a tool is still a fool, and contrary to widespread belief, the quality of the analysis depends on the quality of the analyst, not on the quality of the code used in the analysis. 11.3 Directional Stability of a Road Vehicle When cornering , the body of a road vehicle does not necessarily point in the direction of motion. Instead, the tyres distort as they rotate resulting in an angle between the plane of the wheel hubs and the direction of motion. This is known as a slip angle, although there is no actual slippage of the tyre on the road. The distortion generates forces which keep the vehicle on the road. The tyre forces depend on the slip angle and on the load on the tyre, but for simplicity we shall ignore the dependency on load, and assume the tyre forces are proportional to the slip angle only. These are generated in an axis set which is fixed with respect to the vehicle, which is not an inertial frame of reference (we would not feel centrifugal accelerations on corners if it were). The equations of motion must be derived in this moving axis set. If the body heading is ψ, the fixed axes components of velocity may be transformed into body axes using a 2D axis transformation matrix Ψ (psi): u B u f Where u is the velocity and the subscript B means body and f means fixed axes. Differentiating: u u B u f f The force acting in fixed axes is F f mu f , where m is the vehicle mass, so the body axes force is: FB F f mu f Expressed in terms of the body orientation ψ (note, the transformation matrix is orthogonal): cos sin sin sin , cos cos cos 0 sin 0 Hence, the equation of motion, expressed in body axes is: 0 u B mu B FB m 0 179 Cybernetics for Systems Engineers – DRAFT F Expressed as a pair of scalar equations: X v m Y v u m u Where u,v are the body axes components of velocity and X,Y are the body axes components of force. The first term, on the right hand side of these equations, is the one expected from Newton’s Second Law, the second means that the components of the projection of the velocity on to body axes changes as the body rotates with respect to the velocity vector. Figure 11-2: Road Vehicle Dynamics We assume that the vehicle is moving at constant speed (U), so that the first equation is satisfied identically (the product v is second order in small quantities, and can be neglected). The second equation becomes: v Y U m The moment equation is: C N Where C is the moment of inertia about a vertical axis and N is the yawing moment. 180 Cybernetics for Systems Engineers – DRAFT F If the front axle is ‘a’ metres ahead of the centre of gravity, the slip at the front axle is (since v and are small quantities). f v a U The tyre force is assumed proportional to this, so the front axle generates a force: v a Y f k U The moment is: v a N f ka U Similarly, the rear axle forces and moments are: v b Yr k U v b N r kb U The total side force is: 2v a b Y k U The moment is: a b v a 2 b 2 N k U Hence the equations of motion become: a2 b2 ab k k v CU CU 2 k a b v k U v mU mU These are in state space form, so the characteristic equation is given by: 𝑎 2 +𝑏2 𝑎−𝑏 𝜆 + 𝑘 ( 𝑈𝐶 ) 𝑘 ( 𝑈𝐶 ) | |=0 𝑘(𝑎−𝑏) 2𝑘 (𝑈 + 𝑚𝑈 ) 𝜆 + 𝑚𝑈 i.e: 181 Cybernetics for Systems Engineers – DRAFT F k a2 b2 2k k2 a b2 k a b 0 2 mU C mCU CU 2 If the constant term is negative, there will be at least one positive root. This will occur if a>b and k a b U m a b 2 2 So if the centre of gravity lies closer to the front axle than the rear, the vehicle will be directionally stable. If this condition is not met, instability will be encountered at some speed. The onset of instability may be delayed using a long wheel base, stiff tyres and lightweight construction. Tyre forces also depend on the vertical load applied, so that the lateral forces arising from cornering result in different tyre forces on the left hand side compared to the right, and this must be taken into account when designing suspension systems. However, this simplified analysis does illustrate the importance of the condition of the rear tyres for directional stability. 11.4 Hunting of a Railway Wheelset A railway vehicle differs from a road vehicle in that directional stability is provided by means of a pair of rails rather than the adhesion of the tyres. Actually, railway tyres distort in the region of contact, but since they are usually made of steel, the degree of distortion is hardly discernable. The distortion of the region of contact gives rise to the forces which ultimately determine the directional stability. The kinematic stability of a wheel set is achieved by virtue of the wheel profile. Railway wheels have flanges for safety, but these should rarely come into contact with the rail. The treads are machined with a slight taper, so that lateral displacement changes the diameter of the wheel where it contacts the rail, increasing it on the opposite side to the disturbance, reducing it on the other. The two wheels are rigidly connected together, so they have the same angular velocity. The speed of the hub of the outer wheel increases, that of the inner wheel reduces, and the wheelset steers back towards the centre. If pure rolling contact existed between wheel and rail, the wheel set would continue through the central position to a displacement equal to the initial disturbance on the other side, and subsequently continue to oscillate from side to side. In fact, the motion damps out below a certain speed and is amplified above the critical speed. 182 Cybernetics for Systems Engineers – DRAFT F Figure 11-3: Coning Action of Railway Wheelset The interaction between the wheel and rail is similar to that between a pneumatic tyre and a road surface, the amount of distortion is small, but the high Young’s modulus for steel ensures that the forces involved are high. Figure 11-4 : Contact Points of Yawed Axle 183 Cybernetics for Systems Engineers – DRAFT F As can be seen from the above diagram, the position of the point of contact moves outwards on the treads as the axle yaws, but this is of second order in the yaw angle, and can be neglected. If the speed of the wheelset is U, the angular velocity of the axle is about: U r0 Where r0 is the radius of the wheel for straight running. If the wheelset is displaced to one side by an amount y, the peripheral speed of the wheel in contact with the rail is given by: u p r0 ky Where k is the wheel taper. If the angular velocity of the axle is ωz, the velocity of the point of contact is: uc U d z 2 Where d is the track gauge. These two speeds differ slightly, resulting in elastic distortion of the wheel and rail in the region of contact. In the railway context, this is called ‘creep’, although it is practically identical to the phenomenon of ‘slip’ in road vehicle wheels. The force arising from creep is assumed proportional to this difference in speeds. F u p u c The lateral force arises from the difference in inclination of the normal reactions on the rails: y U z W 2k y m d Where W is the wheelset load and m its effective mass. As the wheelset displaces laterally it also rotates about an axis pointing in the direction of motion. The angle the wheelset makes with respect to the horizontal is: 2ky d If the lateral acceleration is y , the angular acceleration is: 2k y d So the inertial resistance is: L A 184 Cybernetics for Systems Engineers – DRAFT F Where A is the moment of inertia of the wheelset about the direction of motion. This results from a lateral force Y acting at the points of contact, so the moment is: L r0Y Or Y 2k Ay r0 d From which it follows that the effective mass term is: m m0 2k A r0 d Where m0 is the actual mass of the wheelset. The yaw equation becomes: N C z d 2Uk y d z 2 r0 Where σ (sigma) is the creep coefficient. The creep force is proportional to the difference in speed of the wheel hub due to the tread contacting the rail and the axle yaw. The characteristic equation becomes: W k m 2d Ukd r0 C 2 U d 2 0 2C Expanding: W kd U 2 kd d 2 2 W kC 0 C3 2 m 2 d m 4 r 0 For an iron tyre on a steel rail, the creep force generated per unit speed mismatch is enormous, so σ is expected to be very large, and the coefficients not containing the creep coefficient may be ignored. The characteristic equation reduces to: 𝜆2 + 2𝑘 𝑊 ( 𝑑 4𝑚 − 𝑈2 ) 𝑟0 =0 Evidently, the motion becomes unstable at a critical speed given by: 185 Cybernetics for Systems Engineers – DRAFT F U2 Wr0 4m The critical speed depends on the ratio of load to rolling mass. A more complete analysis includes lateral creep, which has here been treated as a rigid constraint. Despite achieving speeds well in excess of 180mph, modern suspension design results in trains that are practically immune to hunting. This was achieved by analysing the root cause of the problem and changing the system parameters, rather than proposing any form of closed loop control. 186 Cybernetics for Systems Engineers – DRAFT F 12 Linearisation 12.1 Introduction It has been mentioned several times that linear homogenous equations usually represent an approximation to the true plant behaviour. This is not an excuse to dismiss the entire edifice of mathematical systems theory, and science in general. That we use approximations is of no concern, provided our theory furnishes us with methods which quantify how wrong our plant equations can become before there is cause for concern. The usual trick of linearization is to identify states which vary slightly over the timescales of interest, and assume all system disturbances are small quantities, so that second and higher order terms may be neglected. In essence, we start with the system in a condition defined by a design case established from earlier analysis. The issue is whether the system will remain close to the design condition if perturbed. This is best illustrated with an example. 12.2 Stability of a Circular Orbit A particle is orbiting around a central attractor in which the law of attraction is an inverse rn relationship. For what values of n is the orbit stable? Since the force attracts to the centre, there is no force acting tangentially, so the angular momentum is constant: ur h Where u is the tangential velocity. The system is in equilibrium such that the centrifugal acceleration is equal to the force per unit mass of the attractor: u2 1 n r r Or, in terms of the angular momentum: h2 k n 3 r r If the orbit increases by an amount Δr, force acting is F h2 k 3 m r r r r n Where F is the force and m the paricle mass. Δr is small compared with r, so the increment in force is given by: 187 Cybernetics for Systems Engineers – DRAFT F Δ𝐹~ 𝑑𝐹 ∆𝑟 𝑑𝑟 = 𝑚 (−3 ℎ2 𝑟4 +𝑛 𝑘 𝑟 𝑛+1 ) ∆𝑟 The equilibrium condition requires: 𝑘 = ℎ2 𝑟 𝑛−3 The equation of motion governing Δr is: 𝑑 2 (∆𝑟) 𝑑𝑡 2 ℎ2 = (𝑛 − 3) 𝑟4 ∆𝑟 The characteristic equation becomes: 2 n 3 h2 0 r4 Evidently, n must be less than 3 for a stable orbit. For the case n=2, the period of the perturbation is: 𝑇= 2𝜋𝑟 2 ℎ = 2𝜋𝑟 2 𝜔𝑟 2 Where ω is the angular velocity of the radius vector from the centre of attraction to the particle. It follows that the period of small perturbations is the same as the period of the orbit. In other words, the orbit changes from circular to slightly elliptical. 12.3 Rigid Body Motion When predicting the motion of free-flying vehicles, such as aircraft or rockets, we have the same difficulty as with the stability of road vehicles; the forces which determine the motion are generated with respect to the vehicle body axes, but Newton’s Second Law requires the motion to be referred to an inertial (constant velocity with respect to the fixed stars) frame of reference. It is necessary to work in the non-inertial frame of the body of the vehicle. In the 2D case, we resolved from fixed to moving axes using a 2×2 axis transformation matrix, the rows of which were the components of lines of unit length along the body x and y axes. These lines are called unit vectors. The 2D case is represented as the trigonometric functions of an angular rotation, and so there is a strong temptation to define the 3D axes in terms of angles. In three dimensions, the orientation of the axes of rotation aren’t fixed, whilst in the 2D case the axis is always parallel to the z-axis, and this leads to confusion. The human eye is limited to interpretation of 2D images, and angles are strictly associated with planes rather than spaces. As the works of M C Esher illustrate, our 2D perception is easily fooled when we try to represent 3D objects on a 2D medium. Of course, the same self-conceit which cannot admit ignorance, and is commonly manifest as an absence of integrity, will ignore this advice. Let us admit our physiological limitations, and avoid this particular path to delusion and self-deception. 188 Cybernetics for Systems Engineers – DRAFT F Figure 12-1 : Illustration of Ambiguity of 2D Pictures of 3D Objects Architect’s drawings and engineering blueprints are produced using orthogonal projection for a very good reason. We cannot extend the angle description to the 3D case because angles are inherently twodimensional entities. We need first to define the plane in which the angle is measured. However, defining the orientation as an orthogonal matrix whose rows are the unit vectors along the three moving axes is simpler, and a direct extension of the 2D approach. 12.3.1 Nomenclature Since aircraft and spacecraft have the least constraints on their motion, we shall derive the general equations of motion using an aeronautical convention. The axis set chosen has the x-axis pointing from the centre of gravity towards the nose. The y-axis points to starboard (to the right, when looking along the x-axis). In order to complete a right-handed orthogonal axis set, the z-axis points downwards. This is a body axis set. The velocity components along the x,y and z axes are denoted u,v and w respectively. The forces are X,Y and Z, the angular velocities around the x, y and z axis are p, q and r for roll, pitch and yaw respectively. The moments about the axes are L, M and N. Consistent with the notation for the moments, the x-axis is the unit vector, (l1, l2, l3), the y-axis is the unit vector; (m1, m2, m3), and the z-axis is (n1, n2, n3). The subscript denotes the fixed axis to which the component of the moving axis is referred. A vector in body axes with components aB, bB, cB, becomes, in fixed axes: 𝑚1 𝑛1 𝑙1 𝑎 𝑚 𝑛 (𝑏) = 𝑎𝐵 (𝑙2 ) + 𝑏𝐵 ( 2 ) + 𝑐𝐵 ( 2 ) 𝑚3 𝑛3 𝑐 𝑙3 189 Cybernetics for Systems Engineers – DRAFT F This may be written in matrix form: 𝑙1 𝑎 (𝑏) = (𝑙2 𝑐 𝑙3 𝑚1 𝑚2 𝑚3 𝑛1 𝑎𝐵 𝑛2 ) (𝑏𝐵 ) 𝑛3 𝑐𝐵 Or more compactly: 𝑎 = 𝑇𝑎𝐵 Where it is understood that ‘a’ is now a vector, and T is a matrix. Resolving from fixed axes to body axes: 𝑎𝐵 = 𝑎𝑙1 + 𝑏𝑙2 + 𝑐𝑙3 With similar expressions for bB and cB. Or, in matrix form: 𝑎𝐵 = 𝑇 𝑇 𝑎 In particular the fixed axes velocity is resolved into body axes: 𝑢𝐵 = 𝑇 𝑇 𝑢 12.3.2 Translation Equation In order to find the inertial force in body axes, the expression for the body axes velocity must be differentiated with respect to time: 𝑢̇ 𝐵 = 𝑇 𝑇 𝑢̇ + 𝑇̇ 𝑇 𝑢 By Newton’s Second Law: 1 𝑢̇ = 𝑚 𝑋 Where m is the mass and X is the fixed axes force vector, so TT X is the body axes force vector. There remains the derivative of T to consider. This term means that the body is rotating with respect to the velocity vector, so the components of velocity along each body axis will change in time even if the magnitude remains constant. Consider the x axis. In a short interval of time δt, it will yaw by an amount rδt in the current y-axis direction, and pitch by an amount –qδt in the z-axis direction, so the change in the x-axis is: 190 Cybernetics for Systems Engineers – DRAFT F Figure 12-2 : Rates of Change of Body Axes 𝑚1 𝑛1 𝛿𝑙1 𝑚 𝑛 (𝛿𝑙2 ) = ( 2 ) 𝑟𝛿𝑡 − ( 2 ) 𝑞𝛿𝑡 𝑚3 𝑛1 𝛿𝑙3 In the limit, as the time interval tends to zero, we have: 𝑙1̇ 𝑚1 𝑛1 ̇ 𝑚 𝑛 (𝑙2 ) = ( 2 ) 𝑟 − ( 2 ) 𝑞 𝑚3 𝑛3 𝑙3̇ Applying the same reasoning to the y and z axes: 𝑇̇ = 𝑇Ω Where: 0 Ω=( 𝑟 −𝑞 Note: −𝑟 0 𝑝 𝑞 −𝑝) 0 Ω𝑇 = −Ω Taking the transpose: 𝑇̇ 𝑇 = −Ω𝑇 𝑇 The translational equation of motion in body axes becomes: 191 Cybernetics for Systems Engineers – DRAFT F 𝑢̇ 𝐵 = 1 𝑋 𝑚 𝐵 − Ω𝑇 𝑇 𝑢 = 1 𝑋 𝑚 − Ω𝑢𝐵 Re-arranging: 𝑋𝐵 = 𝑚(𝑢̇ 𝐵 + Ω𝑢𝐵 ) Or, explicitly in scalar form: 𝑋 = 𝑚(𝑢̇ − 𝑣𝑟 + 𝑤𝑞) 𝑌 = 𝑚(𝑣̇ − 𝑤𝑝 + 𝑢𝑟) 𝑍 = 𝑚(𝑤̇ − 𝑢𝑞 + 𝑣𝑝) 12.3.3 Rotation Equations In order to avoid unnecessary first moment terms, we refer the rotation of the body to the centre of mass. A general particle of the body of mass δm is located at position (x,y,z) in body axes. Since the motion is referred to the centre of gravity, during any translational acceleration, the inertial moment generated by this element of mass is exactly balanced by the other elements of mass, so there is no net moment due to acceleration of the centre of gravity. The position of the particle in fixed axes is: 𝑥 = 𝑇𝑥𝐵 The velocity with respect to fixed axes is: 𝑥̇ = 𝑇̇𝑥𝐵 = 𝑇Ω𝑥𝐵 The acceleration is: 𝑥̈ = (𝑇̇Ω + 𝑇Ω̇)𝑥𝐵 The inertial force is: 𝛿𝑚 𝑥̈ = 𝛿𝑚 𝑇(Ω2 + Ω̇)𝑥𝐵 𝛿𝑋𝐵 = 𝛿𝑚(Ω2 + Ω̇)𝑥𝐵 We can derive the moment of a force from the co-ordinates of the particle in body axes and the components of force acting: 0 𝐿 (𝑀) = ( 𝑧 −𝑦 𝑁 Or: −𝑧 0 𝑥 𝑦 𝑋 −𝑥 ) (𝑌 ) 0 𝑍 𝐿𝐵 = 𝑅 𝑋𝐵 The moment equation becomes: 𝛿𝐿𝐵 = 𝛿𝑚 𝑅(Ω2 + Ω̇)𝑥𝐵 192 Cybernetics for Systems Engineers – DRAFT F Evaluating the right hand side: −(𝑞2 + 𝑟 2 )𝑥 − (𝑟̇ − 𝑝𝑞)𝑦 + (𝑞̇ + 𝑝𝑟)𝑧 (Ω + Ω̇)𝑥𝐵 = ( (𝑟̇ + 𝑝𝑞)𝑥 − (𝑝2 + 𝑟 2 )𝑦 − (𝑝̇ − 𝑞𝑟)𝑧 ) −(𝑞̇ − 𝑝𝑟)𝑥 + (𝑝̇ + 𝑞𝑟)𝑦 − (𝑝2 + 𝑞 2 )𝑧 2 The result is the complicated expression: 𝛿𝐿 = 𝑝̇ (𝑦 2 + 𝑧 2 ) − 𝑞𝑟(𝑥 2 + 𝑧 2 ) + 𝑞𝑟(𝑥 2 + 𝑦 2 ) − (𝑞̇ − 𝑝𝑟)𝑥𝑦 − (𝑟̇ − 𝑝𝑞)𝑥𝑧 + (𝑟 2 − 𝑞 2 )𝑧𝑦 𝛿𝑚 ( 𝑞̇ (𝑧 2 + 𝑥 2 ) − 𝑟𝑝(𝑦 2 + 𝑥 2 ) + 𝑟𝑝(𝑦 2 + 𝑧 2 ) − (𝑟̇ − 𝑞𝑝)𝑦𝑧 − (𝑝̇ − 𝑞𝑟)𝑦𝑥 + (𝑝2 − 𝑟 2 )𝑥𝑧 ) 𝑟̇ (𝑥 2 + 𝑦 2 ) − 𝑝𝑞(𝑧 2 + 𝑦 2 ) + 𝑝𝑞(𝑧 2 + 𝑥 2 ) − (𝑝̇ − 𝑟𝑞)𝑧𝑥 − (𝑞̇ − 𝑟𝑝)𝑧𝑦 + (𝑞 2 − 𝑝2 )𝑦𝑥 Summing over the whole body, we find that the rotational motion depends on six inertia parameters rather than simply the one (mass) for translational motion, or indeed the single moment of inertia for the 2D case. The three dimensional motion of a rigid body is potentially complex for bodies of arbitrary and irregular shape. The inertia terms are called the moments of inertia and the products of inertia. ∫(𝑦 2 + 𝑧 2 )𝑑𝑚 = 𝐴, the moment of inertia about the roll axis ∫(𝑧 2 + 𝑥 2 )𝑑𝑚 = 𝐵, the moment of inertia about the pitch axes ∫(𝑥 2 + 𝑦 2 )𝑑𝑚 = 𝐶 , the moment of inertia about the yaw axis The remaining terms are called products of inertia. In aircraft stability, it is customary to refer the x-axis to the velocity vector so that the product: ∫ 𝑥𝑧 𝑑𝑚 = 𝐸 Is non-zero if the body of the aircraft is at a significant angle of attack in the design cases of interest (e.g. approach, landing and take-off). 193 Cybernetics for Systems Engineers – DRAFT F It is always possible to choose a set of body axes for which the products of inertia are zero. These are called principal axes. Generally, they are associated with the axes of symmetry, and may be found by inspection. The moment equations with respect to principal axes are also known as Euler’s equations: 𝐿 = 𝐴𝑝̇ − (𝐵 − 𝐶)𝑞𝑟 𝑀 = 𝐵𝑞̇ − (𝐶 − 𝐴)𝑟𝑝 𝑁 = 𝐶𝑟̇ − (𝐴 − 𝐵)𝑝𝑞 The equations of motion contain cross product terms and are not linear. As an aside, the equations of motion for rotation are derived from translational accelerations of the individual particles of the body. If any effects arise from the actual rotation of the particles themselves, they will not be predictable from Euler’s equations, as some who should know better seem to think. 12.4 Perturbation forms 12.4.1 Cartesian Rocket – Steady Flight A typical rocket configuration is notable for its symmetry. Often, the plan view is identical to the side view, so that the pitch and yaw moments of inertia are identical, (i.e. B=C). Also, rockets tend to be slender, so that the roll moment of inertia is much smaller than pitch and yaw (i.e. A<<B). A fin-stabilised rocket is expected to roll slowly, so that all components of angular velocity may be considered small quantities. The moment equations become: 𝐿 = 𝐴𝑝̇ , 𝑀 = 𝐵𝑞̇ = 𝐶𝑞̇ , 𝑁 = 𝐶𝑟̇ = 𝐵𝑟̇ 194 Cybernetics for Systems Engineers – DRAFT F The translational equations are similar, except that the speed of the rocket is usually significant. The upper case U denotes the steady state speed, with lower case referring to small changes. The translational equations become: 𝑋 = 𝑚𝑢̇ 𝑌 = 𝑚(𝑣̇ + 𝑈𝑟) 𝑍 = 𝑚(𝑤̇ − 𝑈𝑞) The x-wise equation should be expected from Newton’s Second Law. The lateral equations introduce a term arising from the change of orientation of the body with respect to the velocity vector. If 𝑣̇ = 𝑤̇ = 0, the body is at a fixed orientation with respect to the velocity (i.e. the angle of attack is constant). The velocity vector must be turning in space at the same rate as the body, which can only happen if a centripetal (if you are observing from an inertial frame outside), or centrifugal (if you are on board) acceleration, is acting. It is important to understand that v and w are projections of the velocity vector on to body axes, they are related to the incidence (α) and sideslip (β) angles: 𝑤 𝑤 𝛼 = − tan−1 ( 𝑈 ) ~ − 𝑈 𝑣 𝑣 𝛽 = tan−1 (𝑈) ~ 𝑈 The longitudinal and two lateral equations of motion are decoupled as far as inertial forces are concerned, so the motion is governed by three independent pairs of equations, and the3D problem reduces to three 2D problems. 12.4.2 Cartesian Rocket – High Angle of Attack High angle of attack is usually associated with high lift, such as would be encountered by a guided missile in a violent manoeuvre. The manoeuvre itself would be associated with high centrifugal acceleration and consequently high pitch rate. We need to know that the vehicle is stable under these conditions and not likely to increase its incidence still further, causing it to break up. The high pitch rate couples the roll and yaw, so that the yaw equation becomes: 𝑁 = 𝐶(𝑟̇ + 𝑄𝑝) Where Q is now the steady state pitch rate. Let the speed of the vehicle be VR, the steady state longitudinal and vertical components of body axes velocity become: 𝑈 = 𝑉𝑅 cos 𝛼 𝑊 = −𝑉𝑅 sin 𝛼 Where α is the steady state angle of attack. 195 Cybernetics for Systems Engineers – DRAFT F We shall assume the steady state forces are accounted for in the X,Y and Z terms of the translational equations. These become: 𝑋 = 𝑚(𝑢̇ + 𝑊𝑞 + 𝑄𝑤) 𝑌 = 𝑚(𝑣̇ − 𝑊𝑝 + 𝑈𝑟) 𝑍 = 𝑚(𝑤̇ − 𝑈𝑞 − 𝑢𝑄) The inertial cross coupling with the yaw channel requires the roll rate to be kept near zero. This tends to be difficult to achieve as fin control authority tends to fall off with increasing incidence, so that most of the available control is required to trim at the high angle of attack. Very little residual control is left over for roll. Note that the X and Z equations are now strongly coupled, so the system is governed by a higher order differential equation than was the case at low incidence. Non-linear, and potentially chaotic aerodynamic effects are also encountered at high angles of attack, rendering the design of autopilots difficult. It is curious how ‘systems thinking’ always seems to be limited to the trivial aspects of systems organisation, with scant attention paid either to system behaviour, or the potential problems of achieving desired results. Thus the relatively mundane but easily understood problems of magazine handling appear to take precedence over the more pressing, but harder to resolve, issue of hitting the target. 12.4.3 Spinning Projectile Another case of interest to the symmetrical body is that of a spinning projectile. As the body axes are spinning rapidly, it is difficult to interpret the results obtained from working in body axes. When observing an axi-symmetric body spinning, the eye ignores the spinning motion altogether, as any surface feature is moving so fast as to be a blur. What we see is the motion in a non-spinning axis set, and as we are seeking insight, rather than an abstract answer, we must resolve the equations of motion into this non-spinning set. We propose a non-spinning axis set such that the body axes are rotated at an angle φ with respect to it. We resolve the body axes quantities into this axis set: Forces: 𝑋𝐵 = 𝑋 𝑌𝐵 = 𝑌 cos 𝜙 + 𝑍 sin 𝜙 𝑍𝐵 = −𝑌 sin 𝜙 + 𝑍 cos 𝜙 Velocity: 𝑢𝐵 = 𝑢 𝑣𝐵 = 𝑣 cos 𝜙 + 𝑤 sin 𝜙 196 Cybernetics for Systems Engineers – DRAFT F 𝑤𝐵 = −𝑣 sin 𝜙 + 𝑤 cos 𝜙 Differentiating: 𝑣𝐵̇ = 𝑣̇ cos 𝜙 + 𝑤̇ sin 𝜙 − 𝑣𝑝 sin 𝜙 + 𝑤𝑝 cos 𝜙 𝑤𝐵̇ = −𝑣̇ sin 𝜙 + 𝑤̇ cos 𝜙 − 𝑣𝑝 cos 𝜙 − 𝑤𝑝 sin 𝜙 (since: 𝜙̇ = 𝑝) The convention for angular velocities and moments is they have direction given by their axis direction, which makes them consistent with other vector quantities. Moments: 𝐿𝐵 = 𝐿 𝑀𝐵 = 𝑀 cos 𝜙 + 𝑁 sin 𝜙 𝑁𝐵 = −𝑀 sin 𝜙 + 𝑁 cos 𝜙 Angular velocities: 𝑝𝐵 = 𝑝 𝑞𝐵 = 𝑞 cos 𝜙 + 𝑟 sin 𝜙 𝑟𝐵 = −𝑞 sin 𝜙 + 𝑟 cos 𝜙 Differentiating: 𝑞𝐵̇ = 𝑞̇ cos 𝜙 + 𝑟̇ sin 𝜙 − 𝑞𝑝 sin 𝜙 + 𝑟𝑝 cos 𝜙 𝑟𝐵̇ = −𝑞̇ sin 𝜙 + ṙ cos 𝜙 − 𝑞𝑝 cos 𝜙 − 𝑟𝑝 sin 𝜙 Substituting back into the body axes equations of motion, we have: 𝑌 = 𝑚(𝑣̇ + 𝑈𝑟) 𝑍 = 𝑚(𝑤̇ − 𝑈𝑞) 𝑀 = 𝐵𝑞̇ + 𝐴𝑟𝑝 𝑁 = 𝐵𝑟̇ − 𝐴𝑝𝑞 The first two equations remove the effect of roll from the projection term, which is not surprising. The moment equations introduce a gyroscopic term. 12.4.4 Spacecraft Since spacecraft operate in vacuum, there is no advantage to long slender configurations, so the principal moments of inertia can be nearly equal, and Euler’s equation reduces to its simplest form: 𝐿 = 𝐴𝑝̇ 197 Cybernetics for Systems Engineers – DRAFT F 𝑀 = 𝐵𝑞̇ ~𝐴𝑞̇ 𝑁 = 𝐶𝑟̇ ~𝐴𝑟̇ Many satellites deploy a boom to stabilise orientation using tidal forces to keep it aligned with the gravitational field, in which case the equations for a single axis of symmetry would apply. 12.4.5 Aircraft A missile typically has one set of lifting surfaces for yaw, and another, usually identical set, for pitch. This is a very heavy configuration for an aircraft, which is designed for efficiency, rather than an ability to hit a target. Aircraft are noted for having a single principal lifting surface, supplemented with control surfaces. The measurement of aerodynamic forces in a wind tunnel is most conveniently performed using a sting balance which is aligned with the airflow, so that they are referred to the velocity direction, and not the body. The consequence of not using a principal axis set is the introduction of products of inertia, in particular, the product of inertia: 𝐸 = ∫ 𝑥𝑧 𝑑𝑚 Is non-zero. This does not affect the translation equations, but the moment equations become: 𝐿 = 𝐴𝑝̇ − (𝐵 − 𝐶)𝑞𝑟 − 𝐸(𝑟̇ − 𝑝𝑞) 𝑀 = 𝐵𝑞̇ − (𝐶 − 𝐴)𝑟𝑝 + 𝐸(𝑝2 − 𝑟 2 ) 𝑁 = 𝐶𝑟̇ − (𝐴 − 𝐵)𝑝𝑞 − 𝐸(𝑝̇ − 𝑟𝑞) For an aircraft configuration B≠C, so the small perturbation equation becomes: 𝐿 = 𝐴𝑝̇ − 𝐸𝑟̇ 𝑀 = 𝐵𝑞̇ 𝑁 = 𝐶𝑟̇ − 𝐸𝑝̇ The wind axis set introduces inertial coupling between roll and yaw. This is of no major concern because the aerodynamic rolling and yawing moments are also strongly cross-coupled. 12.4.6 Wider Application The rigid body equations of motion apply equally well to road and rail vehicles, ships, submarines and airships, although some simplifying assumptions will usually need to be made, depending on the specifics of the design case of interest. Vehicles that rely on Archimedes principle for support must impart accelerations to the surrounding fluid when they accelerate, giving rise to ‘virtual mass’ terms which are included in the external forces and moments. 198 Cybernetics for Systems Engineers – DRAFT F 12.5 Partial Derivatives When considering derivatives we have so far only concerned ourselves with functions of a single variable. The increment in the dependent variable in response to an increment in the independent variable is unambiguous. However, if we are dealing with a function of two or more independent variables, the increment can only be meaningful with respect to each independent variable, whilst keeping the others constant. For example a 2D surface may be represented as height as a function of x and y coordinates: ℎ = ℎ(𝑥, 𝑦) We can increment x, keeping y constant to get the slope in the x direction, finding the increment in h and taking it to the limit, results in a differential coefficient, but since its value depends on the value of y which determines the section through the curve, it is written using Russian (Cyrillic) lower case Д (‘da’) i.e: 𝛿ℎ lim ( ) 𝛿𝑥→0 𝛿𝑥 𝑦 = 𝜕ℎ 𝜕𝑥 This is called the partial derivative of h with respect to x. The forces and moments on a flying body potentially depend on all the state variables and their derivatives, so could be functions of as many as 12 variables. It is most unlikely that these functions will be linear, so we will always be linearising about the current design case. As an example, the drag force is expected to be a function of speed and incidence: 𝑋 = 𝑋(𝑢. 𝑤) The increment in drag force due to a small change in speed is: 𝜕𝑋 ∆𝑋 = 𝜕𝑢 ∆𝑢 Having just introduced the partial derivative notation, we shall immediately discard it. In aeronautics we use the shorthand: 𝜕𝑋 𝑋𝑢 = 𝜕𝑢 The deltas are also omitted as it is understood that we are considering perturbations around a design point. In general, we write for any force or moment F and any state x: 𝜕𝐹 𝐹𝑥 = 𝜕𝑥 This is called a ‘stability derivative’. The perturbation X force, is given by adding the contributions due to changes in all relevant states: 199 Cybernetics for Systems Engineers – DRAFT F 𝑋 = 𝑋𝑢 𝑢 + 𝑋𝑤 𝑤 It requires understanding of the physics of the flight vehicle to decide which of the derivatives are important. In particular, it is important to understand how the stability derivatives relate to the geometry of the vehicle. All too often, the aerodynamicists estimate values for the stability derivatives which the control people use to design autopilots. It is usually nobody’s responsibility to see if the configuration can be tweaked to render the autopilot design simpler, or even unnecessary. 12.5.1 The Symmetrical Rocket Re-visited The longitudinal equation of motion is usually considered decoupled from the body motion of the rocket, so the x-wise motion is ignored. In the pitch plane, there is the normal force Z and the pitching moment M. The symmetry of the problem implies that these determined by the pitch plane quantities w (incidence) and q (pitch rate). We have, for the pitch plane: 𝑍 = 𝑍𝑤 𝑤 + 𝑍𝑞 𝑞 𝑀 = 𝑀𝑤 𝑤 + 𝑀𝑞 𝑞 The Zw is the lift generated whenever a body or body-wing combination is rotated to an angle of attack with respect to the air flow. The lift is in the negative z direction, so Zw is expected to be large and negative. Za is the tail force arising from the increase in incidence at the tail due to the pitch rate, this is typically a very small term and is usually neglected. The Mw term characterises the tendency for a stable vehicle to point in the direction of motion, an increase in incidence (i.e. rotating nose up) causes an increase in tail lift, causing the nose to pitch down, provided the centre of lift is aft of the centre of gravity. This is a negative incremental moment in response to s positive incidence disturbance. Mw must be negative for a stable vehicle. This is the reason why arrows and darts have flights at the rear. The Mq term arises from the increase in tail lift due to the pitch rate causing a change in incidence at the tail. It is always negative. The yaw plane has analogous stability derivatives, Yv, Yr, Nv, and Nr. The same comments apply, except that with the axis convention chosen, a positive sideslip (v) must cause the nose to swing to starboard to cancel it, so that Nv must be positive for directional stability. The significant roll stability derivatives are Lv and Lp. . A cruciform rocket tends to be stable either in the + orientation with respect to the plane of incidence, or in the X orientation. A roll disturbance will cause the plane of incidence to rotate relative to body axes, generating sideslip. For roll stability, the rolling moment resulting from this sideslip must tend to rotate the body back until the velocity vector again lies in the body plane of symmetry. A positive sideslip causes a negative roll for stability, so Lv is negative. In rocket applications, Lv is usually ignored, as it amounts to an external excitation of the roll motion. There is no coupling of roll into yaw which would require us to consider roll and yaw in a single equation. 200 Cybernetics for Systems Engineers – DRAFT F The Lp stability derivative is the roll damping term, it is always negative. Including the inertia terms derived earlier, the lateral equation of motion of the rocket becomes: 𝐵𝑞̇ = 𝑀𝑞 𝑞 + 𝑀𝑤 𝑤 𝑚𝑤̇ = 𝑈𝑞 + 𝑍𝑤 𝑤 The characteristic equation becomes: 𝑍𝑤 𝑚 𝜆2 − ( + 𝑀𝑞 𝐵 𝑀𝑤 𝑈 𝐵 )𝜆 − ( − 𝑍𝑤 𝑀𝑞 ) 𝑚 𝐵 =0 Since Zw and Mq are both negative, the coefficient of λ is positive. The constant term is dominated by Mw, which must be negative. The undamped natural frequency of this mode 𝑍 𝑀𝑞 𝐵 𝜔0 = √ 𝑚𝑤 − 𝑀𝑤 𝑈 𝐵 is called the weathercock frequency, as it characterises the tendency of a weather vane to point into wind. 12.5.2 Stability of a Spinning Shell In order to analyse the spinning body, the equation of motion is resolved from body axes to nonspinning axes. We shall assume that the only significant stability derivatives are Zw, Yv, Mw and Nv. Furthermore, from the symmetry of the problem: 𝑌𝑣 = 𝑍𝑤 𝑁𝑣 = −𝑀𝑤 The forces and moments are: 𝑌 = 𝑌𝐵 cos 𝜙 − 𝑍𝐵 sin 𝜙 = 𝑌𝑣 𝑣𝐵 cos 𝜙 − 𝑌𝑣 𝑤𝐵 sin 𝜙 = 𝑌𝑣 𝑣 𝑍 = 𝑌𝐵 sin 𝜙 + 𝑍𝐵 cos 𝜙 = 𝑌𝑣 𝑤 𝑀 = 𝑀𝐵 cos 𝜙 − 𝑁𝐵 sin 𝜙 = −𝑁𝑣 𝑤 𝑁 = 𝑀𝐵 sin 𝜙 + 𝑁𝐵 cos 𝜙 = 𝑁𝑣 𝑣 There are no stabilising fins, so Nv is expected to be negative. Substituting the forces into the equation of motion: 𝑌𝑣 𝑣 = 𝑚(𝑣̇ + 𝑈𝑟) 𝑌𝑣 𝑤 = 𝑚(𝑤̇ − 𝑈𝑞) −𝑁𝑣 𝑤 = 𝐵𝑞̇ + 𝐴𝑝𝑟 𝑁𝑣 𝑣 = 𝐵𝑟̇ − 𝐴𝑝𝑞 201 Cybernetics for Systems Engineers – DRAFT F The symbol A is used to denote the entire system matrix and the roll moment of inertia, it should be obvious from the context which is meant. The characteristic equation is: 𝑌 (𝜆 − 𝑚𝑣 ) | | 0 0 (𝜆 − − 𝑁𝑣 𝐵 0 𝑈 0 𝑌𝑣 ) 𝑚 0 𝜆 𝑁𝑣 𝐵 (𝐵) 𝑝 𝐴 0 −𝑈 𝐴 | − (𝐵) 𝑝| =0 𝜆 This is rather tedious to evaluate. Since we are looking for relationships between parameters which determine stability boundaries, a numerical solution is a waste of everybody’s time. The reference standard model (so beloved of non-participants, to the extent that it must be developed before anybody knows what to put in it} serves only to produce performance predictions at the end of development, for use in acceptance. Actual understanding of the system requires the relationships between the components and subsystems, and their influence on overall system function to be determined. Such insight is provided much more effectively by approximate analytical methods of known bounds of validity, than by letting the Mongolian hordes loose on handle-cranking a time series numerical model, no matter how accurate a representation of the real would system it may be. Knowing what the answer is (42?) does not furnish us with the knowledge needed to change it, should it be unsatisfactory. 12.6 A Few More Tricks If we find ourselves evaluating determinants of order much higher than four, we probably haven’t employed our practical experience of the problem domain adequately. However, if all else fails, the coefficients of the characteristic equation may be evaluated using the following procedure: The coefficient of λn is unity. The coefficient of λn-1 is equal to –trace(A), where the trace of a matrix is the sum of the diagonal elements. The coefficient of λn-i is (-1)i multiplied by the sum of the minors of order i which can be expanded from the diagonal elements. That is all the determinants made up of the rows and columns containing the lead diagonal elements taken i at a time. No matter how I try to explain it, it always comes out as gibberish. However, to show it is actually quite easy, we shall expand the system matrix for the spinning shell example. Let the characteristic equation be: 𝜆𝑛 + 𝑐𝑛−1 𝜆𝑛−1 + 𝑐𝑛−2 𝜆𝑛−2 ⋯ 𝑐0 = 0 The trace is: 𝑌 𝑐3 = −𝑡𝑟𝑎𝑐𝑒(𝐴) = −(𝑎11 + 𝑎22 + 𝑎33 + 𝑎44 ) = −2 𝑚𝑣 202 Cybernetics for Systems Engineers – DRAFT F The coefficient of λ2 is 𝑎11 𝑐2 = (|𝑎 21 𝑌 2 𝑚 𝑐2 = ( 𝑣 ) + 𝑎11 𝑎12 𝑎22 | + |𝑎31 𝑁𝑣 𝑈 𝐵 𝑎13 𝑎11 𝑎33 | + |𝑎41 +0+0+ 𝑁𝑣 𝑈 𝐵 𝑎22 𝑎14 𝑎44 | + |𝑎32 𝑎23 𝑎22 𝑎33 | + |𝑎42 𝑎33 𝑎24 𝑎44 | + |𝑎43 𝑎34 𝑎44 |) 𝐴 2 𝐵 + ( ) 𝑝2 Note that the number of determinants of order i is the number of different ways i objects can be selected from n; i.e: 𝑛! (𝑛−𝑖)!𝑖! The first coefficient contains 4 terms, the second 6, the third 4 and the constant term is the determinant of the entire matrix. Since we know how many determinants we should have for each coefficient, it is easy to check that we haven’t missed one. 𝑎11 𝑎 𝑐1 = − (| 21 𝑎31 𝑌 𝑎12 𝑎22 𝑎32 𝑎13 𝑎11 𝑎23 | + |𝑎21 𝑎33 𝑎41 𝐴 2 𝑁𝑣 𝑈 ] 𝐵 𝑐1 = −2 (𝑚𝑣 ) [(𝐵) 𝑝2 + 𝑎12 𝑎22 𝑎42 𝑎11 𝑎14 𝑎24 | + |𝑎31 𝑎41 𝑎44 𝑎13 𝑎33 𝑎43 𝑎14 𝑎22 𝑎34 | + |𝑎32 𝑎44 𝑎42 𝑎23 𝑎33 𝑎43 𝑎24 𝑎34 |) 𝑎44 The constant term is the determinant: 𝑌𝑣 𝐴 2 ) ] 𝑝2 𝑚𝐵 𝑐0 = |𝐴| = [( 𝑁𝑣 𝑈 2 ) 𝐵 +( Since the system matrix is usually fairly sparse, a large number of the determinants implicit in this approach evaluate to zero. This method is usually quicker and less error prone compared with evaluating |Iλ-A|, and collecting up terms. Having derived the characteristic equation, we can use our knowledge of the expected magnitudes of the system parameters to factorise it approximately. More generally, however, we need to be able to evaluate stability directly from the coefficients of the characteristic equation. We transform the system matrix into the diagonal form of the eigenvalue matrix in order to determine the eigenvalues, so it is obvious that the two matrices have the same characteristic equation. Applying the above algorithm to a diagonal matrix, it is evident that the trace is the sum of the eigenvalues, and the determinant is the product of the eigenvalues. So for a second order system: 𝑐1 = −(𝜆1 + 𝜆2 ) 𝑐0 = 𝜆1 𝜆2 Rather than solve explicitly for the eigenvalues, we note that if the eigenvalues are a complex conjugate pair, or both real, and have the same sign, the constant term is positive. If they are real and have different signs, one of the eigenvalues must be positive. 203 Cybernetics for Systems Engineers – DRAFT F If c0 is positive, the real parts have the same sign, so if c1 is also positive, both eigenvalues must have negative real parts. We can extend this reasoning to higher order systems. Taking a cubic, the coefficients, expressed in terms of the eigenvalues, become: 𝑐2 = −(𝜆1 + 𝜆2 + 𝜆3 ) 𝑐1 = 𝜆1 𝜆2 + 𝜆2 𝜆3 + 𝜆3 𝜆1 𝑐0 = −𝜆1 𝜆2 𝜆3 We notice that for c0 to be positive, either all the eigenvalues are negative or two of the three are positive. Denote the negative eigenvalue –α and the other two β and γ (gamma). The symbols denote strictly positive quantities. 𝑐2 𝑐1 = (𝛼 − (𝛽 + 𝛾))(𝛽𝛾 − 𝛼(𝛽 + 𝛾)) = 𝑐0 − (𝛼 2 + 𝛽𝛾 − 𝛼𝛽 − 𝛼𝛾)(𝛽 + 𝛾) Or: 𝑐2 𝑐1 − 𝑐0 = −(𝛼 2 + 𝑐1 )(𝛽 + 𝛾) If two of the eigenvalues are positive, this will be negative, so if all the coefficients are positive, and this test function is positive, all eigenvalues will have negative real parts. The test function may be written in determinant form: 𝑐2 𝐻2 = | 1 𝑐0 𝑐1 | The top row consists of the even order coefficients, whilst the second row consists of the odd order coefficients. For an nth order characteristic equation, the presence of unstable modes can be checked using a series of test functions derived from an n×n matrix of the form: 𝑐𝑛−1 1 𝐻 = || 0 ⋮ 0 𝑐𝑛−3 𝑐𝑛−2 𝑐𝑛−1 ⋮ 0 𝑐𝑛−5 ⋯ 0 𝑐𝑛−4 ⋯ 0 𝑐𝑛−3 ⋯ ⋯|| ⋮ ⋮ ⋮ ⋯ 1 ⋯ The top row consists of alternate coefficients starting from that for λn-1, padded out with zeros. The second row consists of alternate coefficients starting from that for λn, i.e. unity. The next two rows consist of the top two rows indented by one column, this is repeated until the matrix has n rows and n columns. A matrix of this form is called a Hurwitz matrix. The test functions are determinants of the sub-matrices formed from the top-left elements of the matrix; i.e: 204 Cybernetics for Systems Engineers – DRAFT F 𝐻1 = 𝑐𝑛−1 𝑐𝑛−1 𝐻2 = | 1 𝑐𝑛−3 𝑐𝑛−2 | 𝑐𝑛−1 𝐻3 = | 1 0 𝑐𝑛−3 𝑐𝑛−2 𝑐𝑛−1 𝑐𝑛−5 𝑐𝑛−4 | 𝑐𝑛−3 ⋮ 𝐻𝑛 = |𝐻| If all the test functions are positive, all the roots of the characteristic equation have negative real parts. 12.6.1 Spinning Shell Conditions for Stability The Hurwitz matrix may now be applied to find the conditions under which the spinning shell becomes stable. The Hurwitz matrix is: 𝑌𝑣 −2 ( ) 𝑚 𝑌𝑣 𝐴𝑝 2 𝑁𝑣 𝑈 −2 ( ) [( ) + ] 𝑚 𝐵 𝐵 1 𝑌𝑣 2 𝑁𝑣 𝑈 𝐴𝑝 2 [( ) + 2 +( ) ] 𝑚 𝐵 𝐵 0 2 𝑌𝑣 𝐴 𝑁𝑣 𝑈 [[( ) ] 𝑝2 + ( ) ] 𝑚𝐵 𝐵 𝐻= 0 0 0 2 𝑌𝑣 −2 ( ) 𝑚 1 ( 0 𝑌𝑣 𝐴𝑝 2 𝑁𝑣 𝑈 −2 ( ) [( ) + ] 𝑚 𝐵 𝐵 0 𝑌𝑣 2 𝑁𝑣 𝑈 𝐴𝑝 2 [( ) + 2 +( ) ] 𝑚 𝐵 𝐵 𝑌𝑣 𝐴 2 2 𝑁𝑣 𝑈 2 [[( ) ]𝑝 + ( ) ] 𝑚𝐵 𝐵 The test functions are: 𝑌 𝐻1 = −2 (𝑚𝑣 ) Since Yv is always negative, this will be positive. 𝑌 −2 ( 𝑣 ) 𝑚 𝐻2 = | 1 𝑌 𝐴𝑝 2 −2 ( 𝑣 ) [( ) + 𝑌 𝑚 2 [( 𝑣 ) + 2 𝑚 𝑁𝑣 𝑈 ] 𝐵 𝐵 𝑁𝑣 𝑈 𝐴𝑝 2 𝐵 +( ) ] |= 𝑁𝑣 𝑈 𝐵 𝑌 + ( 𝑣) 2 𝑚 𝐵 Since the shell is assumed aerodynamically unstable, Nv is negative, so stability requires: 𝑌 2 𝑚 ( 𝑣) > | 205 𝑁𝑣 𝑈 | 𝐵 ) Cybernetics for Systems Engineers – DRAFT F 𝑌 −2 ( 𝑣 ) 𝑚 | 𝐻3 = 1 𝑌 𝐴𝑝 2 𝑚 𝐵 −2 ( 𝑣 ) [( ) + 𝑌 2 [( 𝑣 ) + 2 𝑁𝑣 𝑈 𝑚 | 𝐵 𝑁𝑣 𝑈 ] 0 𝐴𝑝 2 𝑌𝑣 𝐴 2 ) ] 𝑝2 𝑚𝐵 𝐵 + ( ) ] [[( 𝐵 𝑌𝑣 0 +( 𝑌 𝐴𝑝 2 𝑚 𝐵 𝑁𝑣 𝑈 2 ) ] 𝐵 −2 ( 𝑣) [( ) + −2 ( ) 𝑚 𝑁𝑣 𝑈 𝐵 | | ] i.e. 𝑌 2 𝑁𝑣 𝑈 𝑌𝑣 2 [(𝑚) 𝑚 𝐵 𝐻3 = 4 ( 𝑣 ) 𝐴𝑝 2 +( ) ] 𝐵 Since Nv is negative, the shell is unstable, because all the other terms in this expression are positive. There is no need to evaluate the fourth test function, if H3 were positive, H4 will also be positive if the constant term c0 is also positive. The instability arises from the lift term, which for a body alone is small. The airflow does work in increasing the angle of attack, but gyroscopic action is a purely inertial effect, associated with conservative forces, and unlike tail fins cannot do work on the airflow. There does not appear to be any means of transferring energy from the shell back into the airflow, so without some means of dissipating this energy, the pitch and yaw kinetic energy must be increased throughout the flight. The yaw damping, Nr seems the only significant candidate. In any case, shells are notable for being very dense compared with aircraft, so the Yv/m term is expected to be small. If it is taken as zero, all the test functions become zero. If we ignore the effect of aerodynamic lift, on the basis that the shell is dense, the characteristic equation retains only the even terms: 𝜆4 + [2 𝑁𝑣𝑈 𝐵 𝐴𝑝 2 𝑁𝑣 𝑈 2 ) 𝐵 + ( 𝐵 ) ] 𝜆2 + ( =0 This can be solved explicitly for λ, yielding a condition for negative roots: 𝑝2 > 4 |𝑁𝑣 𝑈| 𝐵 (𝐴 ) 𝐴 Note that the analysis of the ‘simple’ system – the spinning shell which consists of only a body of revolution is more complex than that of the rocket with its two sets of lifting surfaces. This is to be expected from the Principle of Requisite Entropy. Contrary to current common misapprehension, simple systems exhibit complex behaviours. Thus, by concentrating one’s effort on the wellbehaved system elements , all we are doing is deferring the evil day when the fundamental problems have to be addressed. Leaving this to the implementation stage is to court disaster. 12.6.2 The Aeroplane The stability of an aircraft is expected to be more complex than that of a missile, but less complex than a spinning shell. Economy in weight dictates only a single set of lifting surfaces, which must provide both vertical and lateral stability. Most designs require a tail fin to provide adequate directional stability, although the ‘butterfly taii’, using steep dihedral on the tail plane to eliminate the fin, has been tried. 206 Cybernetics for Systems Engineers – DRAFT F The measurement of stability derivatives is most easily performed in wind axes with the aircraft trimmed to the appropriate angle of attack. Since principal axes are not used, the inertial moments contain a product of inertia, which introduces additional terms into the roll and yaw equations. The classical aeroplane does not exhibit the degree of symmetry of a rocket or missile, but unless the planform is asymmetric, such as in a slewed wing design, the equations may be treated as decoupled between pitch and yaw/roll. The pitch plane moment equation is dominated by the ‘static’ stability term Mw, which characterises the tendency of the aircraft to point into wind. This is negative for a statically stable aircraft and minimum values are specified by the certifying authorities for this derivative. The pitch damping term Mq is also significant, there is also a further moment term arising from the rate of change of incidence. This arises because it takes a finite amount of time for the change in downwash from the main wing to be experienced at the tail. Instantaneous changes in incidence angle cause lift to increase on both main plane and tail plane. The tail lift is then reduced as the increased downwash reaches it. The net effect is an increment in the steady state value of Mw, so it has the same sign as Mw. This derivative is denoted 𝑀𝑤̇ . Unlike the rocket or shell examples, the longitudinal equation of motion is not decoupled from the motion, so there is an Mu term, caused by a small increase or reduction in dynamic pressure. The pitch plane force per unit incidence Zw must be large and negative if the aircraft is to fly at all. The longitudinal equations include Zu which is the change in lift due mainly to the change in dynamic pressure which accompanies a speed change. The Zq term arises because the tail acts a bit like a paddle, generating additional lift when the aircraft rotates in pitch. A further consequence of including the x-wise force is the contribution of gravity. This is usually neglected from missile, shell or rocket stability, as it varies over a much longer timescale than the weathercock mode, characterising rocket stability. The pitch plane equations of motion are: 𝑤 𝑚𝑢̇ = 𝑋𝑢 𝑢 + 𝑋𝑤 𝑤 − 𝑚𝑔 (𝜃 + 𝑈 ) Xu is the change in drag force, mainly due to the change in dynamic pressure, Xw is the change in liftdependent drag (induced drag). The angle θ is the perturbation in the pitch attitude of the body, so the incidence angle must be subtracted from it to yield the perturbation in climb or dive angle. 𝑚(𝑤̇ − 𝑈𝑞) = 𝑍𝑢 𝑢 + 𝑍𝑤 𝑤 + 𝑍𝑞 𝑞 𝐵𝑞̇ = 𝑀𝑢 𝑢 + 𝑀𝑤̇ 𝑤̇ + 𝑀𝑤 𝑤 + 𝑀𝑞 𝑞 𝜃̇ = 𝑞 The yaw/roll equations introduce the product of inertia E, and since aircraft usually have high aspect ratio wings (aspect ratio is: span squared divided by area) compared with rockets or missiles, the 207 Cybernetics for Systems Engineers – DRAFT F cross coupling between roll and yaw is much greater. The rolling moment due to yaw rate (Lr) is significant. This is because the outer section of the port wing move faster than the outer section of the starboard wing with positive yaw rate (nose to starboard) causing a tendency for the starboard wing to drop. This is a positive roll, so Lr is expected to be positive. Np, the yawing moment due to roll rate arises from increase in sideslip at the fin, which isn’t usually balanced by a ventral fin. The roll rate also increases the incidence of the starboard wing, but reduces it on the port. The difference in induced drag of the wings generates an additional yawing moment. As with the pitch plane, gravity is present in the side force equation, proportional to the perturbation in body roll angle (φ). The lateral motion equations become: 𝑚(𝑣̇ + 𝑈𝑟) = 𝑌𝑣 𝑣 + 𝑌𝑟 𝑟 + 𝑚𝑔𝜙 𝐴𝑝̇ − 𝐸𝑟̇ = 𝐿𝑣 𝑣 + 𝐿𝑟 𝑟 + 𝐿𝑝 𝑝 𝐶𝑟̇ − 𝐸𝑝̇ = 𝑁𝑣 𝑣 + 𝑁𝑟 𝑟 + 𝑁𝑝 𝑝 𝜙̇ = 𝑝 Some manipulation is needed to transform the equations into standard state-space form. These equations are studied extensively in specialist books on aircraft stability (e.g. Babister), and there is insufficient space here to even scratch the surface of such a complex subject. Suffice it to say that understanding of the relative magnitudes of the various stability derivatives provides a basis for factorising the two fourth-order characteristic equations approximately, into two quadratic terms for the pitch motion and one quadratic and two simple factors in the case of the lateral motion. The modes associated with these factors are well known to engineers and pilots and have been given names. In the pitch plane, the high frequency quadratic factor is akin to the weathercock mode of the fin-stabilised rocket, and is known as the short period pitch oscillation. The low frequency mode governs the trajectory of the aircraft in the absence of pilot input, and is called the phugoid. This is an oscillation in height and speed, with the two states in quadrature. The lateral motion consists of three modes. The fastest mode tends to be roll subsidence which is simply a decaying roll rate. The lateral equivalent of the weathercock stability is called ‘Dutch roll’ which consists of coupled, damped oscillations in yaw and roll, such that the wingtips describe elliptical paths with respect to the centre of gravity. The low frequency mode is called the spiral mode, which like the phugoid, characterises the trajectory the aircraft would follow in the absence of pilot activity. The effect of the stability derivatives on these modes and the relationship between the stability derivatives and the aircraft and flow field geometry, is used to design an aircraft which is naturally pleasant to fly, as opposed to the ‘white-knuckles special’ which would arise from treating them as purely abstract numbers. 208 Cybernetics for Systems Engineers – DRAFT F The process of approximate factorisation will be illustrated using a simpler example in the following sections. Whereas the rocket, with its rapid response and high degree of symmetry is represented by identical second order equations in pitch and yaw, the classical aeroplane requires different quartic equations for longitudinal and lateral motion. The penalty for removing a set of lifting surfaces, is a more complex description of the dynamics. 12.6.3 The Brennan Monorail At the turn of the Twentieth Century, Louis Brennan, the inventor of the world’s first guided weapon, turned his attention to railways. Never one to adopt the uninspired, pedestrian approaches of his contemporaries, when it came to solving the hunting oscillation which was limiting the speed of trains, he decided the best solution was to discard one of the rails, and run the vehicle on a single rail. This approach presented the obvious problem that the vehicle would topple over, and this problem would indeed deter a lesser inventor. To overcome the problem, Brennan proposed generating stabilising moments by forcing a gyroscope to precess. As the following sections will show, this is not as crazy as it might first appear. Its bad press has more to do with the lack of integrity of experts in conventional wisdom, who were not prepared to be ousted by an upstart amateur like Brennan, than with any technical difficulties. The gyroscope equations were derived when we considered the motion of a spinning shell. It is difficult to keep track of a particle on a rapidly spinning disc, but if we plot its path in space when the 209 Cybernetics for Systems Engineers – DRAFT F plane of the disc is also rotating (i.e. precessing) we see that the curvature of the path changes direction during a single spin. The resulting centripetal force appears as a moment around an axis perpendicular to both the spin axis and the precession axis. Summed over the whole disc, this becomes a continuous moment. As with the works of M C Esher, it is our inability to perceive 3D geometry adequately that renders gyroscopic action counter-intuitive, and a source of anxiety for many who should know better. 12.6.4 Equation of Motion The motion of the vehicle in roll, in the absence of a gyroscope is characterised by: 𝐴𝜙̈ = 𝑊ℎ𝜙 Where A is the moment of inertia of the vehicle about the rail, W its weight, h the height of the cg above the rail, and φ the roll angle. The characteristic equation is: 𝑊ℎ ) 𝐴 𝜆2 − ( =0 It should come as no surprise that this has one positive root. Now we shall add a gimbal frame free to pivot about an axis perpendicular to the roll axis (it may be vertical or horizontal). We have: 𝐽𝜃̈ + 𝐻𝜙̇ = 𝑀𝑐 𝐴𝜙̈ − 𝐻𝜃̇ = 𝑊ℎ𝜙 Where J is the moment of inertia of the gimbal/gyro assembly about the precession axis and H is the product of the gyro moment of inertia about its spin axis multiplied by the spin rate. This will be called the gyro angular momentum. Θ is the gimbal deflection angle and Mc is the controlling moment. With no feedback to the gimbal, the characteristic equation becomes; (𝐽𝐴𝜆2 + (𝐻 2 − 𝑊ℎ))𝜆2 = 0 This is an improvement as the modes will be neutrally stable, provided H2 > Wh. In practice, we should expect friction in the gimbal, which we shall represent as viscous friction proportional to the gimbal angular velocity: 𝐽𝜃̈ + 𝐻𝜙̇ = −𝑀𝜔 𝜃̇ Where Mω is the friction moment per unit angular velocity. The characteristic equation taking friction losses into account is not very encouraging: 𝐽𝐴𝜆4 + 𝑀𝜔 𝐴𝜆3 + (𝐻 2 − 𝑊ℎ)𝜆2 − 𝑀𝜔 𝑊ℎ𝜆 = 0 The friction renders the system unstable. Before addressing the issue of stability, it is evident that the characteristic equation requires a constant term, this is introduced by feedback of the gimbal deflection. Without this term, there is 210 Cybernetics for Systems Engineers – DRAFT F nothing to stop the gimbal from gradually drifting on to the stops, even if the effect of friction were overcome. 𝐽𝜃̈ + 𝐻𝜙̇ = −𝑀𝜔 𝜃̇ + 𝑀𝜃 𝜃 Where Mθ is a gimbal moment due to gimbal deflection. The characteristic equation is: 𝐽𝐴𝜆4 + 𝑀𝜔 𝐴𝜆3 + (𝐻 2 − 𝑊ℎ − 𝑀𝜃 )𝜆2 − 𝑀𝜔 𝑊ℎ𝜆 + 𝑀𝜃 𝑊ℎ = 0 In order for the constant term to be positive, Mθ must be positive. The feedback must be positive. The mechanism implementing this must tend to push the gimbal further away from the equilibrium position, rather than pull it back. This can be implemented as a spring mechanism similar to the toggle mechanism of a bathroom light switch. 12.6.5 Approximate Factorisation If power is lost to the gyro motor, it will start to slow down. In order to ensure there is sufficient time to bring the vehicle to a halt, and to sure it up, before instability sets in, the gyro angular momentum must be very large. It follows, that for any practical system, the stability quartic may be split into two approximate quadratic factors: 𝜆2 + 𝑀𝜔 𝜆 𝐽 𝜆2 − 𝑀𝜔 𝑊ℎ 𝜆 𝐻 𝐻 𝐻2 + ( 𝐽𝐴 − + 𝑊ℎ 𝐴 𝑀𝜃 𝑊ℎ 𝐻 𝐻 − 𝑀𝜃 ) 𝐽 =0 =0 The high frequency mode is called the ‘nutation’, and the low frequency mode the ‘precession’. The nutation is evidently stable, but the precession is dynamically unstable. The vehicle will tend to right itself but will oscillate from side to side with ever increasing amplitude. The root cause of the problem is the fact that unlike a statically stable system, the equilibrium position is with the centre of gravity at its maximum height. In other words, the equilibrium position is a maximum energy, not minimum energy, state. Loss of energy due to friction will result in the gradual lowering of the cg height. It is impossible to rig up a collection of springs and dampers or other passive devices to recover this energy, so the balancing system must employ an active servo of some kind. This differs from all previous examples in that artificial feedback is essential for stability, the system cannot be corrected by merely tweaking the values of parameters. Positive feedback of gimbal position appears to impart static stability, implying positive feedback of gimbal rate might impart dynamic stability. At first sight, positive rate feedback would render the nutation unstable, making matters worse, because the nutation is a very rapid motion. 12.6.6 Brennan’s Solution However, we could exploit the fact that the two modes must be widely separated in frequency for robustness to loss of gyro power. If the servo used to provide the artificial feedback were too slow in its operation to influence the nutation, it could be used to provide positive rate feedback to the precession alone. 211 Cybernetics for Systems Engineers – DRAFT F Consider a servo designed to generate an output torque in response to an input torque demand. It will tend to work by adjusting the output torque until it matches the demand. The simplest representation of a servo of this kind is a first order equation: 1 𝑀̇ 𝑠 = 𝜏 (𝑀𝐷 − 𝑀𝑠 ) Where τ (tau) is the response time of the servo, usually called the time constant, Ms is the servo moment and MD the demanded moment. The moment demand is a signal proportional to the gimbal rate. How this is actually implemented is irrelevant at this stage. Brennan used a pneumatic servo, August Scherl an hydraulic servo. An electric servo would be another possibility. The feedback might even be implemented manually. The equations of motion now include the servo response: 𝐴𝜙̈ − 𝐻𝜃̇ = 𝑊ℎ𝜙 𝐽𝜃̈ + 𝐻𝜙̇ = −𝑀𝜔 𝜃̇ + 𝑀𝜃 𝜃 + 𝑀𝑠 1 𝑀̇ 𝑠 = 𝜏 (𝑘𝜃̇ − 𝑀𝑠 ) Where k is the feedback gain. The characteristic equation becomes: 1 𝑘 (𝜆 + 𝜏 ) (𝐴𝐽𝜆4 + 𝑀𝜔 𝜆3 + (𝐻 2 − 𝑊ℎ − 𝑀𝜃 )𝜆2 − 𝑀𝜔 𝑊ℎ𝜆 + 𝑀𝜃 𝑊ℎ) − 𝜏 𝜆(𝐴𝜆2 − 𝑊ℎ) = 0 If we design the servo to have time constant: 1 𝜏 𝑊ℎ 𝐴 =√ Then: 𝐴𝜆2 − 𝑊ℎ = 𝐴 (𝜆 − √ 𝑊ℎ ) (𝜆 𝐴 +√ 𝑊ℎ ) 𝐴 1 1 = 𝐴 (𝜆 − 𝜏 ) (𝜆 + 𝜏 ) The servo mode is present in both terms, so the characteristic equation governing nutation and precession reduces to a quartic: 𝑘𝐴 𝐴𝐽𝜆4 + 𝑀𝜔 𝜆3 + (𝐻 2 − 𝑊ℎ − 𝑀𝜃 )𝜆2 + ( 𝜏2 − 𝑀𝜔 𝑊ℎ) 𝜆 + 𝑀𝜃 𝑊ℎ = 0 Applying the approximate factorisation, it is evident that a slow servo will stabilise both the nutation and the precession. 12.6.7 Cornering The gyroscopic moments arise from precession with respect to an inertial frame of reference, but the chassis of the monorail is an accelerating frame of reference on corners. The angular velocity of the vehicle introduces an additional gimbal angular velocity, in addition to that applied by the balancing system. With a single gyro, this additional angular velocity results in instability on corners. 212 Cybernetics for Systems Engineers – DRAFT F Brennan solved this problem by using counter-rotating gyros mounted in gimbals which precess in opposite directions. Any torques arising from the vehicle body motion would then have equal and opposite effects on the two gyros, and would cancel out. 12.6.8 Side Loads The stability analysis shows that the vehicle is stable. Side forces are resisted by the vehicle leaning into them until the rolling moment due to the side force is balanced by that due to the weight. It follows that the vehicle negotiates corners by leaning into them until no net side force is experienced on board. 12.7 Concluding Comments Hopefully, it should be clear that deriving the plant equations is seldom the trivial problem implied by the assumption that they are always ‘given’. It is important to analyse the dynamic behaviour of the basic plant to gain insights into the causes of instability, and the potential for selecting plant parameters which yield stable behaviour. The analysis yields stability bounds which help provide understanding of the critical design issues. This analysis provides a sound foundation for the design of a closed loop controller to modify the plant behaviour, should it be deemed necessary. The results take the form of analytical relationships between the plant parameters which define stability boundaries. These are infinitely more useful than numerical results which are only possible for specific values of plant parameters. Whilst the numerical results tell us exactly how stable a particular option might be, they provide very little assistance in deciding on a set of plant parameters which will result in stable, let alone optimum, behaviour. The converse is also true, the Hurwitz matrix and approximate factorisation of the characteristic equation do not form a sensible basis for numerical approaches, they are strictly intended only for manual analysis. Also, deriving and solving the characteristic equation is a very error prone method of calculating the eigenvalues numerically. Most reliable methods employ matrix decomposition approaches such as the QR algorithm. 213 Cybernetics for Systems Engineers – DRAFT F 13 Modifying the System Behaviour 13.1 Introduction Stability analysis forms part of the system characterisation which must precede any other activity. It is undertaken to investigate whether the proposed system will be well-behaved, and provide the stability boundaries which usually define the upper limits of performance which may be achievable. More often than not, however, it is brought in later to explain why a badly-specified system exhibits unpredictable behaviour. Such behaviour is so often attributed to ‘emergent’ properties, but in the vast majority of cases it arises from inadequate analysis. We need to identify the dynamic modes of the environment, rather than assume that there is a simple monotonic relationship between CO2 emissions and global warming. The rise of CO2 emissions in the 19th and 20th Centuries may well have excited an environmental mode which currently manifests itself as global warming. There is nothing to say that reducing them now will have the slightest effect. Depending on the dynamics of the atmosphere, it might even make matters worse. We find the circumstances which result in unsatisfactory behaviour, and if we can’t solve the problem by avoiding the conditions which result in instability, or we can’t adjust the system parameters to achieve the desired behaviour, we must consider using artificial feedback. One of the first applications of artificial feedback was probably to the Whitehead torpedo in the 19th Century. As anybody who has ever tried to sail a toy boat will know, it is practically impossible to get it to follow a straight path without some form of remote control. In order to solve this problem, the torpedo contained a gyroscope which served to detect the deviation of the torpedo heading from the aim direction. The gyroscope was sufficiently massive so that the controlling moments it applied to the tail fins did not upset it appreciably. The yawing motion causes displacement of the gyro in pitch, so that the pitch orientation of the gyro was expected to change but the yaw would be unaffected. Pitch motion of the torpedo was controlled by sensing depth. Feedback was purely mechanical, with the gyro acting both as sensor and actuator, but this is irrelevant from a systems perspective. The torpedo appeared to endow small boats with the capability to take on mighty battleships. However, the introduction of cordite in the 1890s, increasing the accurate range of heavy guns to well beyond that of torpedoes, soon dashed the hopes of the ‘Nouveau école’, who hoped to oust Britain’s naval supremacy with fleets of torpedo boats. As is usually the case, military doctrine was dictated by the contemporary military technology. Admiral Jellicoe, at the Battle of Jutland didn’t employ Nelson’s tactics because they were incompatible with steam powered, armoured battleships having accurate gun ranges in excess of 214 Cybernetics for Systems Engineers – DRAFT F fifteen kilometres. The absurdity of trying to derive technological requirements from military doctrine is self-evident, although this is what actually underpins modern procurement decisions. The torpedo only came into its own when combined with the stealth of the submarine or the speed of an aircraft. Again, these are technological developments, which drove military doctrine. They did not emerge as a result of military aspiration. Indeed, the military described them as ‘underhand and unfair’. The military attitude to novelty, even war-winning novelty, has always been obstructive. Brennan’s monorail could not be made dynamically stable by passive elements alone, so a dashpot arrangement controlled the valves admitting compressed air into a pneumatic servo. A spring arrangement was then included in this sensor to provide positive feedback of both gimbal angle and gimbal rate. The best known early application of artificial feed back was the automatic pilot for aircraft. We have seen that the long term modes of an aircraft consist of the phugoid and the spiral instability, which characterise the trajectory in the absence of the pilot. Applying the torpedo guidance philosophy, only this time using the gyro as a sensor, with separate servos, these modes could be suppressed and in still air at least, the aircraft could fly straight and level unattended. The sensors which early technology could deliver tended to require very high standards of workmanship, and tended to be expensive. Consequently systems were designed to make use of as few sensors as possible. The latter part of the Twentieth Century saw a drastic reduction in the cost of sensors, making multi-sensor systems feasible at low cost. Servos, on the other hand tend to remain expensive, and unless there are strong overriding reasons, most individual controllers will be limited to a single servo. We shall set aside the state-space description used in the basic plant stability analysis for the moment and consider a form of plant description which is more appropriate for a system composed of multiple components connected together in any possible configuration. 13.2 The Laplace Transform We recall that the dynamics of a component are represented as the coefficients of a differential equation, the output from the component is the set of states for which the equations are solved and the input is a function of time. We have shown, that for the constant coefficient plant equation, the output is related to the input (f(t))via the impulse response: ∞ 𝑦(𝑡) ≈ ∫0 𝑓(𝑡1 )ℎ(𝑡 − 𝑡1 )𝑑𝑡1 So this integral must be evaluated for every time sample of the output function. This process is called convolution. This becomes the input to the next system component, whose output is then found from another convolution. Evidently, this is not a very convenient way to derive the system behaviour. Ideally, we should prefer a system representation in which the output is more simply related to the input and the component response. The impulse response is found by assuming the solution is of the form: 215 Cybernetics for Systems Engineers – DRAFT F 𝑦 = 𝑦0 𝑒 𝑠𝑡 Although up until now we have used the symbol λ in the exponent, the use of the symbol s is for consistency with the notation convention associated with this solution method. The non-homogenous linear differential equation becomes: 𝑒 𝑠𝑡 (𝑠 𝑛 + 𝑐𝑛−1 𝑠 𝑛−1 ⋯ + 𝑐1 𝑠 + 𝑐0 )𝑦0 = 𝑓(𝑡) For a linear equation, s is independent of time, so the solution can be found as: 𝑡 𝑡 ∫0 𝑦0 𝑑𝜏 = ∫0 𝑦(𝜏)𝑒 −𝑠𝜏 𝑑𝜏 = (𝑠𝑛 +𝑐 𝑡 1 𝑛−1 𝑠 𝑛−1 ⋯ ) ∫0 𝑓(𝜏)𝑒 −𝑠𝜏 𝑑𝜏 The upper time limit is arbitrary, so to avoid the inconvenience of terms in t, it will be taken as infinity. The resulting integral: ∞ 𝐹(𝑠) = ∫0 𝑓(𝜏) 𝑒 −𝑠𝜏 𝑑𝜏 Is called the Laplace Transform of f(t). The solution of the non-homogenous equation may therefore be written in terms of s, the Laplace operator, rather than time: 𝑌(𝑠) = 𝐺(𝑠)𝐹(𝑠) Where G(s) is the reciprocal of the polynomial in s representing the homogenous equation of the plant in isolation. It is called the transfer function of the plant. Now, if the output from plant G(s) is fed as input into plant H(s), there will be a new output Z(s): 𝑍(𝑠) = 𝐻(𝑠)𝑌(𝑠) = 𝐻(𝑠)𝐺(𝑠)𝐹(𝑠) The transfer function of the combined plant is the product of the transfer functions of the components. This is precisely the formulation we require if we are to consider dynamic components connected together in a network. The components are defined in the transform domain, they are combined using the rules of elementary algebra, and if desired, the time series response to an input may be obtained from an inverse Laplace Transform. 13.3 Transforms of Common Functions 13.3.1 Derivatives and Integrals We justified the substitution: 𝑑𝑦(𝑡) 𝑑𝑡 → 𝑠𝑌(𝑠) on the basis of the characteristic equation. It may be shown to be a general result from the definition of the Laplace transform: ∞ 𝑑𝑓(𝑡) −𝑠𝑡 𝑒 𝑑𝑡 𝑑𝑡 𝐷(𝑠) = ∫0 Integrating by parts: 216 Cybernetics for Systems Engineers – DRAFT F ∞ −𝑠𝑡 𝐷(𝑠) = [𝑓(𝑡)𝑒 −𝑠𝑡 ]∞ 𝑑𝑡 0 + 𝑠 ∫0 𝑓(𝑡)𝑒 The term is square brackets is -f(0), which if the system is started from rest, is taken as zero. It follows that: 𝐷(𝑠) = 𝑠𝐹(𝑠) It can be shown by repeated integration by parts, and the assumption that all higher derivatives are initially zero that: ∞ 𝑑𝑛𝑓 𝐷 𝑛 (𝑠) = ∫0 𝑑𝑡 𝑛 𝑒 −𝑠𝑡 𝑑𝑡 = 𝑠 𝑛 𝐹(𝑠) Quite often the D symbol is used in this sense to denote differentiation, allowing differential equations to be treated as algebraic equations. This is equivalent to using the Laplace operator, so is only mentioned because it is used elsewhere. Equivalently: ∞ 𝑡 1 ∞ 𝑡 1 ∞ 𝐼(𝑠) = ∫0 (∫0 𝑓(𝜏)𝑑𝜏) 𝑒 −𝑠𝑡 𝑑𝑡 = − 𝑠 [(∫0 𝑓(𝜏) 𝑑𝜏) 𝑒 −𝑠𝑡 ] + 𝑠 ∫0 𝑓(𝑡)𝑒 −𝑠𝑡 𝑑𝑡 0 The integral in the first term is taken between the limits t=0 and t=∞, so is identically zero. It follows that, consistent with the use of s in place of a derivative, we may write 1/s in place of an integral. 1 𝐼(𝑠) = 𝑠 𝐹(𝑠) The Laplace operator obeys the rules of elementary algebra, with multiplication and division interpreted as the equivalents of differentiation and integration in the time domain. 13.3.2 Exponential Forms The importance of exponential functions to the mathematical description of self-adaptive systems should be self-evident. The Laplace transform is: ∞ ∞ 1 ∞ 𝐹(𝑠) = ∫0 𝑒 𝑎𝑡 𝑒 −𝑠𝑡 𝑑𝑡 = ∫0 𝑒 −(𝑠−𝑎)𝑡 𝑑𝑡 = − (𝑠−𝑎) [𝑒 −(𝑠−𝑎)𝑡 ]0 We can only ensure a finite result if s>a. Positive values of a (time constant) not only imply a divergent time solution, they bring into question the validity of the Laplace transform itself. Either way, an (s-a) factor in the transfer function denominator is likely to be cause of concern. With this proviso, the Laplace transform of the exponential function becomes: 1 𝐿(𝑒 𝑎𝑡 ) = (𝑠−𝑎) Second order factors are associated with sine and cosine functions. These are conveniently expressed as complex exponentials: sin 𝜔𝑡 = 𝑒 𝑗𝜔𝑡 −𝑒 −𝑗𝜔𝑡 2𝑗 Where j is the square root of -1. 217 Cybernetics for Systems Engineers – DRAFT F The Laplace transform becomes: ∞ ∞ 1 ∫0 sin 𝜔𝑡 𝑒 −𝑠𝑡 𝑑𝑡 = 2𝑗 ∫0 (𝑒 −(𝑠−𝑗𝜔)𝑡 − 𝑒 −(𝑠+𝑗𝜔)𝑡 )𝑑𝑡 Using the result for the real-valued exponential: 𝜔 𝐿(sin 𝜔𝑡) = (𝑠2 +𝜔2 ) Similarly, for the cosine: 𝐿(cos 𝜔𝑡) = (𝑠2 𝑠 +𝜔2 ) 13.3.3 Polynomials The input to the system may be any arbitrary time function, which could be fitted by trigonometric functions or alternatively, by polynomials. ∞ 1 𝑛 ∞ 𝑛−1 −𝑠𝑡 𝐿(𝑡 𝑛 ) = ∫0 𝑡 𝑛 𝑒 −𝑠𝑡 𝑑𝑡 = − 𝑠 [𝑡 𝑛 𝑒 −𝑠𝑡 ]∞ 𝑒 𝑑𝑡 0 + 𝑠 ∫0 𝑡 This yields the reduction formula 𝑛 𝑠 : 𝐿(𝑡 𝑛 ) = 𝐿(𝑡 𝑛−1 ) If n=1: 𝐿(𝑡) = ∫0 𝑒 −𝑠𝑡 𝑑𝑡 = 2 𝑠 𝑠 1 ∞ 1 Repeated application of the reduction formula yields: 𝐿(𝑡 𝑛 ) = 𝑛! 𝑠𝑛+1 13.4 Impluse Function and the Z Transform The exponential, sine and cosine are probably the most important signals used to excite systems in order to discover their properties. The next most important is the impulse: ∞ 𝐿(𝛿(0)) = ∫0 𝛿(0)𝑒 −𝑠𝑡 𝑑𝑡 The delta function is very handy for sleight of hand derivations, which leave the reader more browbeaten than convinced. We shall consider a constant amplitude signal of duration Δt. The Laplace transform for a signal of amplitude a and duration Δt is: ∆𝑡 1 ∫0 𝑎𝑒 −𝑠𝑡 𝑑𝑡 = − 𝑠 [𝑒 −𝑠𝑡 ]∆𝑡 0 Expanding the exponential function as a series, and assuming our error tolerances are such that a first order expansion is adequate: ∆𝑡 ∫0 𝑎𝑒 −𝑠𝑡 𝑑𝑡 = 𝑎∆𝑡 This depends on the actual value of the time interval. 218 Cybernetics for Systems Engineers – DRAFT F How one deals with this difficulty depends on the application. In finding the impulse response of linear systems, we found that an impulse of size 1/(aΔt), is equivalent to releasing the system from rest with unit deflection. The most important application of the impulse is in a sampled data system, such as the input to a digital processing system (not necessarily a computer; a set of shift registers or delay lines could be used, depending on application and ingenuity of the designer). The input is sampled, with an interval between samples of T seconds. A sample of duration at least Δt is required to trigger the input circuits. So if the signal level is y(t), the impulse triggering the input is y(t)Δt. However, we are not initially interested in such low level detail as to how a particular hardware solution conducts the sampling. At the functional level before any hardware solution is considered, we assume that the details of the sampling will be taken care of, such that it is legitimate to define the sampling as unit impulses: ∞ ∫0 𝑦(𝑡) 𝛿(𝑡1 )𝑑𝑡 = 𝑦(𝑡1 ) Any dimensional inconsistency introduced by this perfect sampler is assumed compensated for in the physical system. We get away with apparent dimensional inconsistency because it is the information attached to processes that concern us. The physical quantities concern us only in so far as the saturations they introduce when limits are exceeded and minimum measurement intervals which are possible with affordable sensors. Now a signal which is sampled at intervals of T seconds has the Laplace Transform: ∞ ∫ 𝑦(𝑡)(𝛿(0) + 𝛿(𝑇) + 𝛿(2𝑇) ⋯ 𝛿(𝑛𝑇))𝑒 −𝑠𝑡 𝑑𝑡 0 = 𝑦(0) + 𝑦(𝑇)𝑒 −𝑠𝑇 + 𝑦(2𝑇)𝑒 −2𝑠𝑇 ⋯ + 𝑦(𝑛𝑇)𝑒 −𝑠𝑛𝑇 The convention, when dealing with sampled data is to use the notation: 𝑧 = 𝑒 −𝑠𝑇 So that: 𝑌(𝑧) = 𝑦(𝑡)(1 + 𝑧 + 𝑧 2 ⋯ + 𝑧 𝑛 ) Since z is itself a Laplace transform in disguise, it obeys all the rules of elementary algebra, and is called the Z transform. Most modern systems are expected to implement the control processing digitally as algorithms, abdicating the credit for control system design, in the minds of managers, to software development processes and coders. 13.5 Laplace Transform Solution of Non-Homogenous Equations A linear differential equation with input is no longer homogenous. We shall take an example which will prove important later on. 219 Cybernetics for Systems Engineers – DRAFT F Consider a first order linear system driven by a sinusoidal input: 𝑑𝑦 𝑑𝑡 + 𝑎𝑦 = 𝑏 sin 𝜔𝑡 Taking Laplace transforms: (𝑠 + 𝑎)𝑌(𝑠) = 𝑏 (𝑠2 1 𝜔 +𝜔2 ) 𝜔 𝐴𝑠+𝐵 𝐶 𝑌(𝑠) = 𝑏 ((𝑠+𝑎) (𝑠2 +𝜔2 )) = 𝑏 ((𝑠2 +𝜔2 ) + (𝑠+𝑎)) This process of splitting the denominator in to simple factors whose inverse functions are known, is employed extensively in Laplace transform methods, it is known as partial fraction factorisation. Multiplying both sides by (s+a)(s2+ω2) yields three equations in the unknown coefficients: 𝐴+𝐶 =0 𝑎𝐴 + 𝐵 = 0 𝑎𝐵 + 𝜔2 𝐶 = 𝜔 Solving: 𝜔 𝐶 = (𝑎2 +𝜔2 ) 𝑎𝜔 𝐵 = (𝑎2 +𝜔2 ) 𝐴 = − (𝑎2 𝜔 +𝜔2 ) Taking the inverse transforms: 𝑦(𝑡) = 𝑏 ( 𝑎 sin 𝜔𝑡−𝜔𝑐𝑜𝑠𝜔𝑡 (𝑎 2 +𝜔2 ) 𝜔 + (𝑎2 +𝜔2 ) 𝑒 −𝑎𝑡 ) The second term is governed by the transient response, which for a stable system will decay to zero. In the longer term, the output will be: 𝑦(𝑡) = 𝑏 ( Let: sin 𝜙 = 𝜔 √𝑎 2 +𝜔2 𝑎 sin 𝜔𝑡−𝜔 cos 𝜔𝑡 ) (𝑎 2 +𝜔2 ) , then cos 𝜙 = 𝑦(𝑡) = 𝑏 𝑎 √𝑎 2 +𝜔2 sin(𝜔𝑡−𝜙) √𝑎 2 +𝜔2 If the input is imagined as the angle of the shaft of an alternating current generator, the angle φ is the angle of the output relative to it. It is called the phase angle. The phase is given by: 𝜔 𝜙 = − tan−1 𝑎 220 Cybernetics for Systems Engineers – DRAFT F It characterises the delay caused by the system. Also, we notice that the amplitude of the output signal is reduced by a factor; 1⁄√𝑎2 + 𝜔 2. This quantity is called the gain. Applying the same approach to an nth order system: (∑𝑛0 𝑐𝑖 𝑠 𝑖 )𝑦(𝑡) = 𝑏 sin 𝜔𝑡 Applying the Laplace transform, we have: 𝜔 1 𝑌(𝑠) = 𝑏 ((𝑠2 +𝜔2 ) ∑𝑛 𝑐 𝑠𝑖) 0 𝑖 The fraction decomposition yields: 𝐴𝑠+𝐵 𝑌(𝑠) = 𝑏 ((𝑠2 +𝜔2 ) + ∑𝑛−1 𝐷𝑖 𝑠 𝑖 0 ) 𝑛 ∑0 𝑐𝑖 𝑠𝑖 The second term is characterised by the transient, which will die out if the component is stable, leaving a sin(ωt) and a cos(ωt) term. The effect of a system of any order is to introduce a phase lag and a gain. This will be considered in greater detail when we discuss frequency domain methods. 221 Cybernetics for Systems Engineers – DRAFT F 14 Impulse Response Methods – Evan’s Root Locus 14.1 Introduction Historically, the design of control systems originated simultaneously in two different fields; that of aeronautics and that of electronics. The former was an extension of the stability analysis relating behaviour to aircraft configuration and flow field characteristics, whilst the latter was an extension of AC circuit theory. The equivalence of the methods used was not appreciated because of the differences in the physical hardware, a failing which is even more common nowadays, particularly amongst those who delude themselves into believing their approach to reflect ‘systems thinking’. The aeronautical approach is largely one of assessing the behaviour in response to a sudden input, such as the pilot suddenly pulling back on the joystick. The transients were long enough to be observed directly, and pilots could describe their experiences, assuming they survived the test flight. If we consider a system having transfer function from control input to the desired output given by G(s), we know from the Principle of Requisite Entropy that we bring the actual output close to the desired value by adjusting the input until the error between the two is zero. In a linear system the steady state output is proportional to the input, so we fulfil the requisite entropy requirement by applying the error directly to the control input. We make use of the fact that Laplace transforms obey the rules of elementary algebra, to derive the transfer function of the system when we close the loop. The closed loop transfer function is related to the open loop transfer function according to: 𝑘𝐺 (𝑠) 0 (𝑠) 0 𝐺𝑐 (𝑠) = 1+𝑘𝐺 222 Cybernetics for Systems Engineers – DRAFT F Where k is a multiplier of the error signal, called the feedback gain. Note, that if the gain becomes very large, the transfer function tends to unity. In other words, the system becomes completely insensitive to the open loop plant behaviour. Specification of components for the closed loop system cannot be based on open loop characteristics for this reason. In general, open loop systems require components to be manufactured to the tightest tolerances, requiring the highest standard of workmanship. Closing the loop obviates the requirement for extreme precision. Practically all of control theory is concerned with finding the maximum feedback which is achievable in practice before the system goes berserk. Usually, the feedback consists of more than a gain. An artificial dynamic system, called a compensator, is usually included to modify the plant dynamic response to permit high values of feedback gain. 14.2 Example: Missile Autopilot My apologies to any pacifists who might be reading, but the missile autopilot is so much simpler than that of an aircraft, so it has been chosen for the sake of clarity, rather than indicative of warmongering. The autopilot associated with the short period pitch mode of an aircraft is very similar. The natural stability of a missile has already been considered. The equations of now need to be modified to take account of control inputs. The aeronautical convention for control deflection is ξ (ksi), η (eta) and ζ (zeta) for aileron, elevator and rudder deflections, respectively. The control deflections generate forces and moments, characterised by control derivatives, by analogy with the stability derivatives. The control derivatives are the incremental changes in force or moment due to incremental changes in the corresponding control deflection, under the current flight condition. The roll control derivative is Lξ, the rolling moment due to aileron deflection. Aileron deflection does not usually generate significant pitch or yaw forces or moments, but might increase drag slightly. The pitch and yaw control moment derivatives are Mη and Nζ respectively. The convention used here is for positive control deflections to cause a positive moment about the associated axis. On a tail controlled missile the deflection is positive with the trailing edges deflected upwards, on a canard controlled missile (control surfaces ahead of centre of gravity), they are deflected downwards. In yaw the deflection is to the right for tail control, to the left for canard control. This is not the usual definition, which is based on the direction of rotation of the control surface about the hinge line. Since the control surfaces generate lift in order to apply moments to the body, the derivatives Zη and Yζ may also be significant. For a tail controlled missile Zη is positive, for canard control it is negative, with the signs reversed for Yζ. Note: the trajectory modes of a missile; the equivalents to aircraft phugoid and spiral modes are dominated by the steering law used in the guidance, and are consequently irrelevant to the autopilot. 223 Cybernetics for Systems Engineers – DRAFT F The equations of motion in yaw (pitch is identical) are: 𝑚𝑣̇ = 𝑌𝑣 𝑣 − 𝑚𝑈𝑟 + 𝑌𝜁 𝜁 𝐶𝑟̇ = 𝑁𝑣 𝑣 + 𝑁𝑟 𝑟 + 𝑁𝜁 𝜁 Taking the Laplace transforms: 𝑌 𝑌 𝜁 (𝑠 − 𝑣 ) 𝑈 𝑣(𝑠) 𝑚 𝑚 ( )( ) = (𝑁𝜁 ) 𝜁(𝑠) 𝑁 𝑁 − 𝑣 (𝑠 − 𝑟 ) 𝑟(𝑠) 𝐶 𝐶 𝐶 Solving yields transfer functions relating sideslip to fin deflection and yaw rate to fin deflection: 𝑌𝜁 𝑣 𝜁 𝑟 𝜁 = 𝑌𝜁 𝑁𝑟 𝑁𝜁 𝑈 𝑠−( + ) 𝑚 𝑚 𝐶 𝐶 𝑌𝑣 𝑁𝑟 𝑁𝑣 𝑈 𝑌𝑣 𝑁𝑟 2 𝑠 −( + )𝑠+( + ) 𝑚 𝐶 𝐶 𝑚 𝐶 = 𝑁𝜁 𝑌 𝑁𝜁 𝑌𝜁 𝑁𝑣 𝑠−( 𝑣 − ) 𝐶 𝑚 𝐶 𝑚 𝐶 𝑌 𝑁 𝑁 𝑈 𝑌 𝑁 𝑣 𝑟 𝑣 𝑠2 −( + )𝑠+( + 𝑣 𝑟) 𝑚 𝐶 𝐶 𝑚 𝐶 When the transfer functions are written in terms of s, it is obvious that the states are functions of s and not time, so the explicit (s) has been omitted. Considering the proper function of the autopilot (a concept conspicuous by its absence from most control texts), it is unlikely that the output from the system should be either sideslip or yaw rate. The guidance command takes the form of a steering command, requiring the missile to generate centripetal acceleration. The lateral acceleration is, from the translational equation of motion: 𝑓𝑦 (𝑡) = 𝑣̇ + 𝑈𝑟 Taking the Laplace transform: 𝑓𝑦 (𝑠) = 𝑠𝑣(𝑠) + 𝑈𝑟(𝑠) After some manipulation, this yields: 𝑓𝑦 𝜁 𝑌𝜁 = 𝑌 𝑁𝜁 𝑌𝜁 𝑁𝑣 𝑠−( 𝑣 − )𝑈 𝑚 𝐶 𝑚 𝐶 𝑚 𝐶 𝑌𝑣 𝑁𝑟 𝑁𝑣 𝑈 𝑌𝑣 𝑁𝑟 2 𝑠 −( + )𝑠+( + ) 𝑚 𝐶 𝐶 𝑚 𝐶 𝑚 𝑠2 − 𝑌𝜁 𝑁𝑟 This is a more typical transfer function in that it consists of both a numerator polynomial and a denominator polynomial. The denominator is recognisable as the characteristic equation of the open loop plant. It is more easily derived using Laplace transforms, because the derivatives of the quantities involved do not introduce additional state variables, so the order of matrices employed is much lower. The order of the numerator polynomial is the same as that of the denominator. This means that there is a direct link between the input control deflection and the output lateral acceleration. Physically, this is due to the direct lift of the control surface. Deflecting the control surface 224 Cybernetics for Systems Engineers – DRAFT F generates lift immediately; but it takes a finite amount of time for the body to rotate to an angle of attack. In a tail controlled missile, this lift is in the opposite sense to the wing/body lift, whilst for a canard control it acts in the same sense. However, the canard angle of attack is greater than that of the body when trimmed to the manoeuvre angle of attack, whilst for tail control it is less, so the tail-controlled missile is expected to trim at a higher angle of attack before the control surfaces become ineffective. If the missile is statically unstable (i.e. Nv is negative) , this situation is reversed. There is a potential benefit in making the missile statically unstable and canard controlled, but stability would have to be achieved by feedback control. There are many other issues which must be addressed before it can be decided whether a canard design is appropriate in a particular application. The wider analysis requires full understanding of the dynamics of all options, if it is to be any more than hot air and power point. The control deflection demand generated by the feedback cannot be realised immediately, we should expect some delay between the demand and achieved acceleration, so we can propose an exponential decay servo response: 𝜁 𝜁𝐷 1 = 1+𝜏𝑠 Where the subscript D denotes a demanded quantity. The dynamics of the servo/missile combination is represented by simply multiplying the two transfer functions. 𝑓𝑦 𝜁𝐷 𝑌𝜁 = 𝑌𝜁 𝑁 𝑌 𝑁𝜁 𝑌𝜁 𝑁 𝑟 𝑣 𝑣 2 1 𝑚 𝑠 − 𝑚 𝐶 𝑠−( 𝑚 𝐶 − 𝑚 𝐶 )𝑈 1+𝜏𝑠 𝑠2 −(𝑌𝑣+𝑁𝑟 )𝑠+(𝑁𝑣𝑈+𝑌𝑣𝑁𝑟 ) 𝑚 𝐶 𝐶 𝑚 𝐶 The dynamics of sensors, human beings in the loop, or any other dynamic process may be included in a similar manner. 225 Cybernetics for Systems Engineers – DRAFT F 14.3 General Transfer Function Figure 14-1: Open Loop Poles and Zeros The missile example has demonstrated the point that transfer functions representing the dynamic behaviour of real world objects consist in general of a ratio of two polynomials in s with the numerator order no higher than that of the denominator. The denominator polynomial is the characteristic equation of the component in isolation. Since it is generally found by concatenating dynamic processes, the open loop transfer function is usually factorised, so that the roots of the numerator and of the denominator may be found by inspection. The numerator roots are called the zeros, and the denominator roots are called the poles. A general open loop transfer function will take the form: 𝑍(𝑠) 𝐺𝑜 (𝑠) = 𝑃(𝑠) Closing the loop, we have: 𝐺𝑐 (𝑠) = 𝑘𝐺𝑜 (𝑠) 1+𝑘𝐺𝑜 (𝑠) = 𝑘𝑍(𝑠) 𝑃(𝑠)+𝑘𝑍(𝑠) Evidently, when the feedback gain is small, the closed loop characteristic equation is nearly the same as the open loop characteristic equation, so the closed loop poles are close to the open loop poles. As the gain tends to infinity, the closed loop poles tend to the open loop zeros, and as the number of poles usually exceeds the number of zeros, some of the closed loop poles tend to infinity as the gain increases. Now we could calculate the closed loop poles for a range of values of feedback gain using one of the analysis packages which are readily available, for a price. This would give us an accurate plot of the 226 Cybernetics for Systems Engineers – DRAFT F migration of poles, but without a certain degree of geometric intuition which is gained from plotting this root locus manually, it is impossible to decide how to design compensators to distort it into a more desirable shape. A fool with a tool remains a fool. 14.4 Rules for Constructing the Root Locus The poles and zeros of the open loop system are known from the outset. These are plotted on an Argand diagram (Cartesian coordinates in which the real component is plotted in the x direction and the imaginary component in the y direction). These will either be real or appear as complex conjugate pairs (i.e. same value of real part with the imaginary parts having opposite signs). Figure 14-2 : Root Locus Along The Real Axis 14.4.1 Real Axis Starting from the rightmost pole or zero join the poles and zeros in pairs along the real axis. The pairs may be both poles, both zeros or a pole and a zero, it doesn’t matter. Multiple poles or multiple zeros which coincide on the plot are treated as line segments of zero length for the purposes of deciding where the gaps in the real axis plot occur. Coincident pole zero pairs are not plotted, unless they occur in the right hand half plane. 227 Cybernetics for Systems Engineers – DRAFT F Figure 14-3 : Asymptotes Directions and Intersections with Real Axis 14.4.2 Asymptotes An asymptote is a straight line which the curve tends towards as s tends to infinity. The number of asymptotes is equal to the difference between the number of open loop zeros and number of poles. The asymptotes correspond to the condition |s|→∞, for which the closed loop characteristic equation becomes: 𝑠 𝑝−𝑧 + 𝑘 = 0 s is a complex number which may be written in polar form: 𝑠 = |𝑠|(cos 𝜃 + 𝑗 sin 𝜃) = |𝑠| ( 𝑒 𝑗𝜃 +𝑒 −𝑗𝜃 2 +𝑗 𝑒 𝑗𝜃 −𝑒 −𝑗𝜃 ) 2𝑗 = |𝑠|𝑒 𝑗𝜃 The direction, θ is called the argument. Evidently, the argument of –k is π radians (180⁰) . The directions of the asymptotes with respect to the positive real axis are therefore iφ, with i=1 to p-z, and φ given by: 180 𝜙 = 𝑝−𝑧 The intersection of the asymptotes with the real axis is given by the centre of gravity of the open loop poles and zeros taking the poles as positive and zeros as negative. 228 Cybernetics for Systems Engineers – DRAFT F Figure 14-4 : Angles of Departure from Poles and Arrival at Zeros 14.4.3 Exit and Arrival Directions The factor (s+z) is the vector from the point s to the point –z on the Argand diagram. The argument of this number is the direction of the open loop zero at –z from the current pole. In order to satisfy the equation: 𝑍(𝑠) 1 + 𝑘 𝑃(𝑠) The sum of the directions from the poles minus the sum of the directions from the zeros must be π radians (180⁰). The direction of the locus as it leaves a complex pole, or the angle of arrival at a complex zero, is: 𝜃𝑖 = 180 − ∑ 𝜃𝑝 + ∑ 𝜃𝑧 Where the subscript z indicates the line drawn from the node of interest to the zeros and p denotes the directions to the poles. The node of interest, whether pole or zero is not included in the sums on the right hand side. All directions are with respect to lines drawn parallel to the real axis. 14.4.4 Break Away Points Between pairs of zeros, and pairs of poles, the root locus branches out. At the branch out point two branches of the locus intersect the real axis, so they correspond to a condition where the closed loop characteristic equation has a repeated root. This corresponds to a turning point in Go(s), i.e: 𝑑𝐺𝑜 𝑑𝑠 Or, = 𝑃(𝑠) 𝑃 𝑑𝑍 𝑑𝑃 −𝑍 𝑑𝑠 𝑑𝑠 𝑃2 𝑑𝑍(𝑠) 𝑑𝑠 =0 − 𝑍(𝑠) 𝑑𝑃(𝑠) 𝑑𝑠 =0 This may seem a laborious exercise, but we know from the outset, that the solution must be real and lie between the two nodes, so start with the mid-point, and try a few values. 229 Cybernetics for Systems Engineers – DRAFT F 14.4.5 Value of Gain It is usually necessary to estimate the gain at positions of interest along the root locus, such as in the vicinity of the imaginary axis. If the point of interest is s1, then: 𝑘=− 𝑃(𝑠1 ) 𝑍(𝑠1 ) Figure 14-5 : Rules Applied to Produce a Sketch of the Root Locus 14.4.6 Comment on Rules In this age of expensive analysis packages, there is a temptation on the part of technically naive managers to imagine that high quality work is impossible without the best software money can buy. Admittedly, it is easier to manage a software development than it is research, development or studies work, but we achieve little in pretending that all tasks are software developments, as appears to be the current wisdom. It is the quality of the analysts that determines the quality of the results, not the quality of the software employed. No analyst worth his salt would accept the results from a model without conducting his own tests, relevant to the task in hand. Nor would he ever accept the claims of salesmen or other parties motivated by self-interest, at face value. How times have changed. Many of the systems still in service were designed with little more than slide rules and graph paper, yet despite the advances in analysis software and revolution in sensor technology, the actual results produced are hardly commensurate. Asking a modern analyst to demean himself with graph paper, ruler and protractor is like asking him to write with a quill pen. Unfortunately, we cannot acquire the geometric intuition which constitutes understanding of how the open loop nodes affect the closed loop behaviour, by handle cranking a code. We can simulate rain showers in the computer, but they will never make us wet. 230 Cybernetics for Systems Engineers – DRAFT F One should indeed check the manual sketch of the root locus with a mechanical calculation of the closed loop roots for a range of feedback gains. Indeed, once the design is completed, it probably is as well to reproduce it using the computer, as results from a machine, complete with spurious precision, carry more weight than equally correct manual plots on graph paper. 14.5 Autopilot Revisited The root locus is a numerical approach to system analysis and design, so the derivation of algebraic conditions defining constraints on the system parameters cannot be accommodated. In order to overcome this limitation, we try to minimise the reliance on magic numbers, and try to characterise the numerical quantities employed in terms which characterise the system. We note that the missile weathercock mode is characterised by an undamped natural frequency: 𝜔0 = √| 𝑁𝑣 𝑈 𝐶 + 𝑌𝑣 𝑁𝑟 | 𝑚 𝐶 We expect that nodes in the region of this frequency will have greatest effect on the corresponding mode. It would seem reasonable to scale time by this quantity so that experience gained with one missile autopilot design stands a chance of being transferable to other missiles. We can usually derive a scaling term of this kind from the proper function of the controller which is being designed. 14.5.1 Numerical Example The following values are have been derived for a typical missile configuration , although for obvious reasons, they do not represent any specific missile. The mass is taken as 50kg, with longitudinal moment of inertia 15kg m2, and the speed is 680ms-1. The stability derivatives are: 𝑍𝑤 = 𝑌𝑣 = −255.0 𝑁𝑚−1 𝑠 𝑀𝑤 = −𝑁𝑣 = −26.8 𝑁𝑠 𝑀𝑞 = 𝑁𝑟 = −5.0 𝑁𝑚 𝑟𝑎𝑑−1 𝑠 𝑍𝜂 = −𝑌𝜁 = 1350 𝑁𝑟𝑎𝑑−1 𝑀𝜂 = 𝑁𝜁 = 9500 𝑁𝑚 𝑟𝑎𝑑−1 The negative Mw immediately tells us that the missile is statically stable. The positive Zη tells us it is tail-controlled. By reversing the signs of appropriate derivatives, we can investigate canard control and statically unstable missiles. The following analysis is artificial in the sense that it is assumed the acceleration of the centre of gravity can be measured. In practice, the centre of gravity moves during flight, and the centre of gravity is typically located within a rocket motor, which is not the easiest place to mount a sensor. However, we shall ignore hardware related problems, until we have completed at least one iteration on the basis of perfect sensors. Sensor performance flows down from autopilot requirements. 231 Cybernetics for Systems Engineers – DRAFT F In fact, we are ahead of ourselves in tackling auto pilot design as a first stage in the system design process. We would normally define the target set and the environment in which the targets would be encountered, and the platforms from which the weapon is to be fired. These would define requirements for the guidance loop from which missile agility, manoeuvrability and range requirements would be determined, together with launch condition constraints . Magazine handling or logistics might then dictate size and weight, but only if these constraints do not impinge on primary function. Modern ‘systems thinking’ appears to drive system design from secondary or tertiary requirements, with no attempt at prioritising which issues need to be addressed with greatest urgency. For the immediate purposes of illustrating root locus, we shall assume the missile design emerged from the wider system requirements. The weathercock frequency for this missile is readily calculated as about 35 rad/sec (5.5 Hz). The transfer function relating lateral acceleration to control deflection is: 𝑓𝑦 𝜂 Let = −27𝑠2 −9.0𝑠+2.16×106 𝑠2 +5.43𝑠+1225.0 𝑠 𝜎=𝜔 0 𝑓𝑦 𝜂 = −27 (𝜎 2 +0.33𝜎−65) 𝜎 2 +0.16𝜎+1 The zeros areboth real, and very nearly symmetrical about the imaginary axis at a distance of 8 characteristic frequency units. The poles form a complex conjugate pair symmetrically placed above and below the real axis. The coordinates are about -0.1±j1.0. As there are as many zeros as poles, there are no asymptotes. The break away point is at -0.1+j0. As the gain increases, the poles move down towards the break away points. This reduces the autopilot natural frequency (slows down its response), but increases the damping. Increasing the gain beyond the break away point causes one pole to move along the negative real axis to the negative zero. The other pole moves out towards the right half plane zero, introducing an unstable mode. 232 Cybernetics for Systems Engineers – DRAFT F Figure 14-6: Lateral Acceleration Response As the airframe response probably required effort and ingenuity on the part of the aerodynamicists to achieve, it does not seem sensible to add complexity in order to make the missile more sluggish. However the damping should be closer to 0.7 for a satisfactory response. In fact the situation is worse than this. The assumption has been made implicitly that the servo response is so rapid that its associated pole lies to the left of the left half plane zero. In practice, this is seldom, if ever, possible. 14.5.2 Initial Selection of Servo Pole The placement of nodes far to the left in the Argand diagram ensures the root locus will only enter the right hand half plane for very high values of gain. There is nothing in the root locus analysis which prohibits this, so at first sight we should be able to add zeros and poles in the compensator to distort the root locus to achieve any response we desire. In practice this is not the case. Although we must design the system to be stable, it isn’t stability which dictates where we can place the compensator or servo nodes. Considering the servo; it is evident that speeding up its response has implications regarding power, weight, size and cost. The servo time constant must be chosen to be as long as we can get away with. Also, the feedback signal is contaminated with noise, and generally the shorter the response time (higher the bandwidth), of the compensator, the more noise will be fed into the system. Introducing zeros, which is the usual strategy for improving stability, effectively differentiates the signal, which serves to amplify the noise. These effects are taken into account in an initial iteration by applying a few rules of thumb: In order to minimise the interaction between the servo mode and the airframe response, the servo bandwidth should be at least three times that of the weathercock mode. 233 Cybernetics for Systems Engineers – DRAFT F In order to avoid excessive noise, no compensator may have bandwidth greater than ten times the weathercock mode. The effect of servo saturation and system noise will be considered presently. Once an initial design has been roughed out, these effects will be analysed explicitly, using methods described in later chapters. The selection of a servo, or any other system element, independently of the remainder of the system, is a recipe for disaster. However, current ‘systems thinking’ appears to imagine that all the components can be specified on cost, maintainability and other grounds irrelevant to primary function, in the forlorn hope that they can be assembled into a working system. Dream on. Figure 14-7 : Effect of Servo Pole Placing the servo pole at s=-3+j0 introduces an asymptote on the negative real axis starting at the left half plane zero. The branch on the real axis runs from the servo pole to the right half plane zero. The two branches starting from the open loop weathercock mode poles must eventually reach a break away point beyond the left half plane zero. Before this happens,there is a significant excursion into the right hand half plane. When these two branches finally move back into the left hand half plane, we should expect the servo mode to be dangerously close to the right. If the system without the servo was unsatisfactory, we should usually expect additional lags to make matters worse. Remember that the fundamental assumption of linear feedback is the output of the plant is a good estimate of the input, and vice versa. As we increase the feedback gain, the less tolerant the system becomes to errors arising because the open loop plant does not have time to reach equilibrium. Estimating the future output state by differentiation (adding zeros) delays the onset of instability, whilst integration (adding poles) aggravates the problem. These principles apply for any null-seeking regulatory system, regardless of the details of its operation, and not just to linear feedback systems. Any self-adaptive process which can regulate 234 Cybernetics for Systems Engineers – DRAFT F itself in an environment whose disturbance set is open, must be limited by delay. There seems to be an odd perception that because the system is complex, it is somehow immune to these effects, why else are they totally ignored? The immediate effects of the servo pole on the root locus may be found from the angle of departure from the weathercock poles. One zero is roughly at 180⁰, whilst the other is at 0⁰ the other pole of the complex pair is at -90⁰, whilst the servo pole is at 180⁰+tan-11/3. The angle of departure from the pole with the positive real part is Θ=180-(-90+180+tan-1 1/3)+180=90-tan-1 1/3=72⁰ The servo pole repels the weathercock mode from the branch on the real axis, so that the system becomes unstable at very small values of feedback gain. 14.5.3 Adding Compensation Since the root locus starts at the open loop poles and finishes on the open loop zeros, the poles can be said to repel the locus, whilst the zeros tend to attract it. We can use this geometric intuition to decide where zeros must be placed to achieve the desired behaviour. To begin with, we noticed that the servo mode messed things up rather badly, so placing a zero coincident with it will prevent it from interfering with the closed loop transfer function. We cannot construct a compensator having more zeros than poles, so we try a compensator of the form: 𝜎+3 𝐶(𝜎) = 𝜎+10 This effectively shifts the servo pole well over to the left. The consequences of doing this in terms of noise and saturation will be considered in later iterations. Placing an artificial zero over a plant pole is called pole/zero cancellation. We notice that the right half plane zero attracts a branch of the root locus as the gain increases, and life would be simpler if it could be cancelled as well. Unfortunately, introducing an unstable pole to cancel this zero will not work. Pole zero cancellation is more accurately called pole/zero concealment. The dynamics represented by the pole or zero which is cancelled still exists in the system, they just can’t be influenced by the control input or observed from the measurement. Introducing an unstable, cancelled pole introduces an unstable mode which we can neither see, nor influence, with catastrophic consequences. The right half-plane zero puts an upper bound on the achievable closed loop bandwidth. In order to pull the weathercock mode poles further over to the left, we must introduce a pair of complex conjugate zeros into the left half plane. The obvious choice is to cancel the open loop poles and introduce new poles in a more favourable position. For example, consider the compensator: 𝐶(𝜎) = 235 𝜎 2 +0.16𝜎+1 𝜎 2 +1.4𝜎+1 Cybernetics for Systems Engineers – DRAFT F Figure 14-8 : Using a Series Compensator This cancels the weathercock poles and effectively shifts them to a more favourable position on the Argand diagram, with damping ratio increased from the original 0.08 to a more satisfactory 0.7. This appears to solve the airframe limitations without requiring the loop to be closed. The resulting controller involves pre-compensating, with no feedback at all. This solution has in fact been used to avoid the expense of sensors, but the resulting system is no longer error seeking, so is inherently lacking in robustness to uncertainty in the plant and disturbances in the loop. The pre-compensator can only deal with disturbances at the input. If we know that the disturbances, due to say, air turbulence, are likely to introduce less noise into the system than a sensor, it may not be worth closing the loop. Without knowledge of the form of disturbances occurring within the loop, we must seek the highest loop gain we can get away with. Zeros must be introduced further over to the left in order to attract the weathercock poles. These require two additional poles well over to the left. In order to avoid introducing excessive noise, the characteristic frequency associated with these compensator zeros cannot be much higher than 3. If the compensator zeros are too far over to the left, the root locus originating at the weathercock poles will cease to be attracted to them, and the root locus will be attracted to the pair of zeros on the real axis. The right hand zero puts a limit on the position of the compensator zeros, and ultimately on the achievable closed loop bandwidth. In practice, it is not possible to design a satisfactory controller having closed loop bandwidth greater than half that of the right half plane zero. 236 Cybernetics for Systems Engineers – DRAFT F In this case, the right half plane zero is located well over to the right, so this particular upper bound will not apply. However, if the tail moment arm were reduced and, fin size increased to compensate, this zero would move closer to the origin, and would severely limit the response time which could be achieved. Figure 14-9 : Introduction of Compensator Zeros In the absence of the right half plane zero, we could make the gain arbitrarily large, because the only asymptotes lie in the negative real direction. In practice, all the sensor responses and the structural modes, which have been ignored, would influence the behaviour at very high values of gain. The right half plane zero is a consequence of the reverse action of the fin. This is not experienced in canard control configurations, but these too have their own sense of humour. 14.5.4 Concluding Comments The example indicates that the knowledge needed to design satisfactory control systems originates from outside the control domain, and depends on a broader understanding of the factors which limit the behaviour. Part of the problem arises because we have chosen an issue which cannot be taken in isolation from the remainder of the system. Our starting point is to specify what targets must be engaged, from what launch platforms and in what environment. The end user is over-stepping his competence in specifying details of weapon behaviour. It is up to the designers, who are the competent authorities, to flow down this information to actual system constraints. We started at an inappropriate point, not because the methods are only applicable to ‘low level’ system issues, but because, in the missile context, we have not yet covered the material needed for the higher level analysis. One further point is that accelerometer feedback by itself is almost never used. The impossibility of positioning the accelerometers at the centre of gravity is one reason, but in many cases the natural 237 Cybernetics for Systems Engineers – DRAFT F airframe weathercock response has adequate bandwidth, but is lightly damped. Feedback of angular velocity via a rate gyro is probably a more urgent requirement. The yaw rate response is characterised by a single zero, and this lies in the left half plane, because the moment applied by the control surface does not change direction during the transient response, like the lateral acceleration. 𝑌 The position of the zero is very nearly − 𝑚𝑣 , and characterises the lag between the body rate and the turn rate of the trajectory. Alternatively, it characterises the time taken for the angle of attack to reach its steady state value. For this reason, it is often referred to as the ‘incidence lag’. 14.6 Gyro Monorail Although more obscure, the monorail is a more self-contained problem than the autopilot. Instead of requirements flowing down from the need to hit a specified target, the need to stand upright is a pretty fundamental system requirement. Unlike a statically stable missile, which will fly after a fashion with no active feedback, the monorail must employ feedback control to stand up. In order to apply root locus methods, we require some numerical values representative of the problem. These are derived in the following paragraphs. 14.6.1 Representative Parameters We shall consider a 20 tonne vehicle, with centre of gravity 2.5m above the rail at equilibrium. The moment of inertia about the rail is therefore about 1.25×105kg m2. Denoting the gyro angular momentum H, it must be large compared with the unstabilised toppling time constant: 𝐻2 𝑊ℎ ~100 𝐴 𝐽𝐴 J is dominated by the gyro, and is equal to roughly half the polar moment of inertia about the spin axis (denoted I): 𝐻 = 𝐼𝜔 Where ω is the gyro spin rate in radians/sec. In order to avoid instability on corners, there must be two counter-rotating gyros, linked to precess in opposite directions. Let H1 be the angular momentum of a single gyro: (2𝐻1 )2 2𝐽 Or: =2 𝐻12 𝐽 =4 𝐼 2 𝜔2 𝐼 = 100𝑊ℎ 𝐼𝜔2 = 25𝑊ℎ The polar moment of inertia is given by: 1 𝐼 = 2 𝑚𝑟 2 Where m is the gyro mass and r is the radius. This yields an expression for the gyro mass: 238 Cybernetics for Systems Engineers – DRAFT F 𝑊ℎ 𝑚 = 50 (𝜔𝑟)2 Now ωr is the peripheral speed. Experience with gas turbine discs indicates that peripheral speeds of 400ms-1 do not present any significant azard, so we shall assume half this value. The gyro mass then works out at: 𝑚 = 50 × 20000×9.81×2.5 2002 = 125𝑘𝑔 Or 250kg for both. This does not seem prohibitive, as is often claimed. Treating the gyro as a disc with thickness to diameter of 10:1, made of steel, with density 7870kg m-3, the radius is 0.29m. The polar moment of inertia is 5.3kgm2, and the lateral moment of inertia is 2.7kgm2. The spin rate is 690 rad/sec (6600 rpm), resulting in a gyro angular momentum, for both gyros of 20700 kgm2s-1. The undamped natural frequency of the nutation mode is: 𝐻2 𝜔𝑛 = √ 𝐽𝐴 = 7.4 rad/sec The precession must be widely separated from the nutation if Brennan’s solution of applying positive gimbal rate feedback is to work. The precession frequency should be about 1/10th of the nutation frequency; i.e: 0.75 rad/sec, to minimise the risk of interference between the modes. The gimbal deflection feedback is: 𝐻2 𝑀𝜃 = 𝜔𝑝 2 𝑊ℎ From which Mθ=491 Ns/radian. The nutation mode will be damped so as to achieve a damping ratio of 0.7. This implies the damping moment should be: 𝑀𝜔 = −2𝜁𝑛 𝜔𝑛 𝐽 Where ζ in this context means damping ratio. This yields the result that Mω, regardless how it is implemented, must have a value -28 Ns/rad/sec. The assumption is made that these two feedbacks are implemented mechanically. This is only because we are currently dealing with single input/single output methods. 14.6.2 Root Locus The plant equations, with the gimbal deflection feedback and viscous damping become: 𝐽𝜃̈ + 𝑀𝜔 𝜃̇ − 𝑀𝜃 𝜃 + 𝐻𝜙̇ − 𝑀𝑠 = 0 𝐴𝜙̈ − 𝑊ℎ𝜙 − 𝐻𝜃̇ = 0 𝜏𝑀̇ 𝑠 + 𝑀𝑠 = 𝑀𝐷 Where τ is the servo time constant and MD is the input signal to the servo. 239 Cybernetics for Systems Engineers – DRAFT F From a systems perspective, there is something intrinsically weird about controlling the gimbal angular velocity in order to regulate the roll angle. For the slower mode (the precession), we should expect the gyroscopic terms to dominate the roll equation of motion, which becomes: −𝐻𝜃̇ ~𝑊ℎ𝜙 So that, as far as the precession mode is concerned, the gimbal angular velocity is roughly proportional to the roll perturbation, so conceptually, the loop is regulating the roll angle. The transfer function relating gimbal angular velocity to servo demand is: 𝜃̇ 𝑀𝐷 1 𝑠(𝐴𝑠2 −𝑊ℎ) 3 2 2 𝜔 𝑠 +(𝐻 −𝑊ℎ−𝑀𝜃 )𝑠 −𝑀𝜔 𝑊ℎ𝑠+𝑀𝜃 𝑊ℎ) = (𝜏𝑠+1) (𝐴𝐽𝑠4 +𝑀 𝑊ℎ , 𝐴 The zeros are located at ±√ so that by using a servo time constant equal to this, we can cancel the servo pole with the left half-plane zero. That is why the fifth order characteristic equation governing nutation and precession is reduced to a quartic. There remains a right half plane zero, and a zero at the origin. There are two more poles than zeros, so the asymptotes are perpendicular to the real axis. The precession poles form a complex conjugate pair in the right hand half plane. If the precession frequency is low compared with the frequency of the right hand zero, the branches from these poles will be attracted to the branch on the real axis between the origin and the right half plane zero. Since all these nodes lie in the right hand half plane, the potential for a stable system does not look good. The actual behaviour in the region of the nodes governing precession requires numerical values. The time constant on which the system parameters appear to be based is the toppling time constant of the unstabilised vehicle. Using this as a scale factor, the right half plane zero is located at 1+j0. The nutation poles are given by: 𝜎 2 + 5.25𝜎 + 14 = 0 And the precession poles are: 𝜎 2 − 0.032𝜎 + 0.14 = 0 The servo pole is cancelled with the left half plane zero, so there is only a single open loop zero at s=+1.0. The nutation is stable with poles at -2.625±2.66, i.e. at 45°to the real axis. For the precession mode close to the origin these two directions are equal and opposite and cancel each other out. 240 Cybernetics for Systems Engineers – DRAFT F Figure 14-10 : Brennan Monorail - Precession Mode, Root Locus Sketch The precession poles are located at 0.016±j0.37. The angle of departure is given by: 0.37 )) 0.984 𝜃 = 180 + (90) − (tan−1 ( This is -110⁰. The break away point is given by: 𝑃(𝑠) 𝑑𝑍(𝑠) 𝑑𝑠 − 𝑍(𝑠) 𝑑𝑃(𝑠) 𝑑𝑠 =0 In this case, we ignore the contribution of the nutation, so the equation becomes: (𝜎 2 − 0.032𝜎 + 0.14) − (2𝜎 − 0.032)(𝜎 − 1) = 0 The break away point is either 2+j0 or -0.054+j0. The only root which lies on the real axis locus is 0.054+j0. The same root locus has been calculated explicitly in the following diagram. 241 Cybernetics for Systems Engineers – DRAFT F Figure 14-11: Brennan Monorail - Precession Mode, Numerical Solution The sketch is quicker to produce, compared with programming the calculation and plotting the result on a computer, and practice at drawing root loci yields intuitive understanding of the loop dynamics which could never be obtained from the distraction of writing code. 14.7 Comments on Root Locus Examples The missile autopilot and gyro monorail were chosen deliberately to illustrate the point that practical systems can be very badly behaved. Unstable open loop behaviour, right half plane zeros and the requirement for positive feedback, are not just features contrived to illustrate arcane theoretical niceties. Sufficient information has been presented to enable the interested reader to plot the root loci for unstable missiles and canard controlled missiles, and to consider explicit feedback of roll angle in the case of the monorail balancing system. The remaining examples are better behaved, and consequently not so instructive, but differ from the preceding examples in that without feedback there would be no system function at all. They represent the common situation when the system has no natural feedback. 14.8 The Line of Sight Tracker It is often necessary to be able to maintain a line of sight between a ground station and a moving object. A classical case is the tracking of a satellite so as to ensure a high-gain communications antenna is kept pointing at it. Similar techniques could be used to maintain the sight line of an astronomical telescope, (optical or radio) to a fixed point in the sky as the Earth rotates. 242 Cybernetics for Systems Engineers – DRAFT F All such systems have a sensor operating in the band concerned which has the capability of measuring any error between the target and the actual pointing direction. How this is achieved in a specific system is of no immediate concern. All that need be known is the sensor produces a signal proportional to the angular error. If this angular error is fed into a motor of some kind which applies a torque to rotate the sensor in space. The equation of motion is: 𝐼𝜃̈ = 𝑇 Where T is the motor torque and I the moment of inertia of the sensor. Feedback consists of the error between the target direction and the current pointing direction (otherwise known as the bore sight error): 𝜖 = 𝜃𝐷 − 𝜃 The root locus starts at two poles at the origin which then migrate in opposite directions along the imaginary axis. Simple feedback results in an undamped oscillation which increase in frequency. In fact, the servo pole will incline the asymptotes at 60⁰ to the positive real axis, so increasing the gain actually causes the loop to exhibit a divergent oscillation. Evidently, we need some zeros to pull the locus into the left hand half plane. A realisable compensator requires the number of zeros to be equal to, or less than, the number of poles, so at least one compensator pole must be present. This will be located well over to the left. The servo poles would be expected to put a stability boundary on the system behaviour, but this is rather a ‘tail wag dog’ approach to system design. Servo behaviour ought to be an output from the design process not an input. There are occasions where the design problem consists of making best use of existing components, such as servos, but even in these cases it is worthwhile finding the performance shortfall in introducing this artificial constraint. In many cases it turns out to be a false economy. The sensor is presumably designed to track a target of known brightness at a specific range. In a military situation, the target will be doing all it can to minimise the signal which appears at the detector, failing that it will try to increase the noise, to render the angular measurement as inaccurate as possible for as long as possible. A radio telescope must see as far as it possibly can into space. For this reason, in most situations, when steering a sensor, we should expect the performance to be limited by noise in the loop. Noise precludes the use of very high frequency poles, so will tend to reduce the practical system bandwidth. Indeed, additional high frequency poles may be introduced to prevent the system noise getting to the servos. If targets can come close to the tracker, as can be the case with a missile tracker or gun director, there will be a further constraint imposed by the need for the servos to accommodate the maximum angular acceleration as the target flies past. 243 Cybernetics for Systems Engineers – DRAFT F Performance is expected to be limited by noise or maximum sightline acceleration, or both. Once these have been characterised it is possible to examine the limitations imposed by stability. Usually, stability imposes a more severe performance limitation than would be expected from a naive examination of the open loop component behaviours. 14.9 Tracking and the Final Value Theorem The sensor pointing loop differs from the other two cases in having no natural feedback, the loop behaviour is dictated entirely by the artificial feedback. The implicit, and mistaken, assumption in much finite state machine representations is that all systems are of this form. It also differs, particularly from the monorail, in that the primary function is to follow a moving target. The input demand, rather than zero, is a function of time. Apart from stability, the quality of the system is determined by how accurately the output follows the input when the latter varies in time. It does not follow that because a system is a good regulator that it is a good tracker. The quality of a tracker is assessed by the transfer function from the input demand to the error signal. The transfer function is: 𝜀 𝑦𝑖𝑛 1 1+𝑘𝐺0(𝑠) = = 𝑆(𝑠) If the input is a polynomial in time of order n: 𝑛! 𝑦 𝑦𝑖𝑛 (𝑠) = 𝑠𝑛+1 If the tracking error is to be maintained at a constant value: 𝜀 𝜖(𝑠) = 𝑠 The steady state corresponds to s=0, so the steady state tracking error is: 𝜖 = (𝑠𝑆(𝑠)𝑦𝑖𝑛 (𝑠))𝑠→0 It follows that in order to track an input signal defined by a polynomial in time of order n with zero error: 𝑆(𝑠) ( 𝑠𝑛 ) 𝑠=0 =0 The numerator of S(s) is the denominator of the open loop transfer function (i.e. P(s)). If the input is a constant signal, S(0) must equal zero for zero tracking error. This result is known as the final value theorem. Consider the missile autopilot example. The open loop transfer function is: 𝑓𝑦 𝜁 244 𝑌𝜁 = 𝑌 𝑁𝜁 𝑌𝜁 𝑁𝑣 𝑠−( 𝑣 − )𝑈 𝑚 𝐶 𝑚 𝐶 𝑚 𝐶 𝑌 𝑁 𝑁 𝑈 𝑌 𝑁 𝑠2 −( 𝑣+ 𝑟 )𝑠+( 𝑣 + 𝑣 𝑟 ) 𝑚 𝐶 𝐶 𝑚 𝐶 𝑚 𝑠2 − 𝑌𝜁 𝑁𝑟 Cybernetics for Systems Engineers – DRAFT F It follows that the error transfer function is: 𝑆(𝑠) = 𝑌 𝑁 𝑁 𝑈 𝑌 𝑁 (𝑠2 −( 𝑣+ 𝑟 )𝑠+( 𝑣 + 𝑣 𝑟 )) 𝑚 𝐶 𝐶 𝑚 𝐶 𝑌𝜁 𝑌𝜁 𝑁 𝑌 𝑁 𝑁 𝑈 𝑌 𝑁 𝑌 𝑁𝜁 𝑌𝜁 𝑁𝑣 𝑟 (𝑠2 −( 𝑣 + 𝑟 )𝑠+( 𝑣 + 𝑣 𝑟 ))+𝑘( 𝑠2 − 𝑠−( 𝑣 − )𝑈) 𝑚 𝐶 𝐶 𝑚 𝐶 𝑚 𝑚 𝐶 𝑚 𝐶 𝑚 𝐶 From which: 𝑆(0) = 𝑁 𝑈 𝑌 𝑁 (( 𝑣 + 𝑣 𝑟 )) 𝐶 𝑚 𝐶 𝑁𝑣 𝑈 𝑌𝑣 𝑁𝑟 𝑌 𝑁𝜁 𝑌𝜁 𝑁𝑣 (( + ))+𝑘( 𝑣 − )𝑈 𝐶 𝑚 𝐶 𝑚 𝐶 𝑚 𝐶 This only tends to zero if the feedback gain tends to infinity, so the achieved lateral acceleration follows the demand with a finite error. In order to overcome this problem the denominator of the open loop transfer function must contain an integrator. The feedback consists at least of a proportional term, plus an integral term: 𝑘 = 𝑘𝑝 + 𝐾𝐼 𝑠 Integral of error feedback is commonly used, not only to improve the tracking behaviour, but also to avoid the need for very high feedback gains. The sensor pointing loop already contains two integrators, so should be expected to track a constantly accelerating sight line with constant error. Tracking loops are characterised by the order of polynomial signal they can track with zero error. This is the ‘Type’ of the tracking loop. A loop of type n is one which can track a polynomial input of order n-1 with zero error. 14.10 The Butterworth Pole Pattern We have seen that for a second order system with no zeros (otherwise known as minimum phase), a reasonable pole pattern is given by the equation: 𝑠 2 + 1.4𝜔0 𝑠 + 𝜔02 = 0 The roots of this equation lie on the intersections of a circle of radius ω0, with two lines inclines at ±45⁰ to the negative real axis. This distribution is called a second order Butterworth pole pattern. This would usually characterise a mode of the system. In general, we should study the dynamic behaviour to understand the system modes, and we should avoid coupling modes together by ensuring there is a separation in undamped natural frequencies. However, if all else fails, and the problem is ‘given’ as one of pure control, with no broader knowledge of the plant, the poles may be placed in higher order Butterworth patterns given as solutions of the following equations: 𝑠 3 + 2𝜔0 𝑠 2 + 2𝜔02 𝑠 + 𝜔03 = 0 𝑠 4 + 2.613𝜔0 𝑠 3 + 3.4𝜔02 𝑠 2 + 2.613𝜔03 𝑠 + 𝜔04 = 0 245 Cybernetics for Systems Engineers – DRAFT F The poles lie on a circle of radius ω0. A Butterworth pole pattern is unlikely to be of much use to the Brennan monorail, which relies on the wide separation of the modes. Nor is there any benefit in coupling the short period pitch oscillation with the phugoid mode of an aircraft. In general, we would no more tune functionally different modes to the same frequency than we would tune all the strings of a guitar to the same frequency. 14.11 Status of Root Locus Methods Understanding the root locus provides useful insights into how the stability of the plant is affected by the addition of compensation. It is one of many techniques which would be employed in order to gain understanding of the plant, and like all the other methods presented in this book, it is far from a panacea. The prohibited regions on the Argand diagram are defined by considerations of, for example, noise and saturation, which lie outside the scope of an analysis of stability. The methods remain important because stability is the most important single attribute of the system. Stability is strictly a system attribute which is not attributable to any single component, but to the behaviour of all components in concert. In that sense, it could be considered an ‘emergent’ property. It is important also to realise that the poles and zeros of the plant can wander appreciablly as we consider the range of values the plant parameters can feasibly adopt. In principle it should be possible to ascribe probability distributions to the plant parameters and from them determine the probabilities of closed loop poles migrating into the right half plane. Currently, such an explicit approach to characterising the system robustness has not been attempted. Instead, traditional rough rules of thumb have been extended into multi-input-multioutput system analysis with a degree of precision which is hardly consistent with the fundamental crudeness of the traditional robustness measures. Most plots describing control systems should probably be drawn with a marker pen or thick crayon to avoid any hint of spurious precision. Remember, if the plots were precise, we probably wouldn’t need feedback control in the first place. 246 Cybernetics for Systems Engineers – DRAFT F 15 Representation of Compensators 15.1 Introduction Despite the title this section does not go into the specifics of any particular implementation, but presents a more complete mathematical representation which can be used for analysis and modelling purposes. In an era of widespread digital computing, it is unlikely that compensators will be implemented as continuous time (analogue) circuitry, although initial conceptual design may be undertaken on the assumption that signals are continuous. In essence, we start from an infinite sampling rate before we decide, after adequate analysis, what the sampling rate should be. As in all system analysis we seek a constraint which is dictated by the system proper function, even when processor speed has been specified beforehand. This is because we wish to know what the shortfall in potential performance may be as a consequence of adopting hardware which was decided on anything but system considerations. Even if we do settle on a digital solution, the fact remains that reality is not a sampled data process, so in analysing control loops we must represent the real world as if it were sampled data, or the compensator as if it were continuous. In practice, we do both. 15.2 Modelling of Continuous Systems Although the basic design work will be undertaken using manual analysis, sooner or later we are going to require a model of the aspects of the system which are of current interest. This may take the form of a time-stepping model, but as we shall see, the model which most obviously and directly represents the system is seldom appropriate for furnishing the insight needed to understand it. As the sage put it; to err is human, but to really screw it up, get a computer. As managers appear to be unaware of any form of computer model beyond the high-fidelity reference standard model appropriate for acceptance, but useless for both analysis and design, it is advisable to have some idea as to how compensators might be represented using a high level language, such as C. A continuous time system is typically modelled using some form of numerical integration algorithm. The Fourth Order Runge-Kutta algorithm is quite popular. As the name implies, it assumes that over the duration of the time step, the system time evolution is adequately described as a quartic polynomial in time. It is of paramount importance that all state variables are updated simultaneously, and not piecemeal. Failure to do so introduces the risk of ‘sorting errors’ by which variables which have already been updated are used as the starting point to update other variables. The result is a spurious divergence in the solution. The best defence against typical coding errors of this form is a sound knowledge of the expected system behaviour derived from analysis of the problem space. In order to ensure all states are updated simultaneously, the integration algorithm is usually designed to operate on a pair of arrays, the first containing the current values of the state variables, 247 Cybernetics for Systems Engineers – DRAFT F the second containing the derivatives. A typical implementation in the C language is presented below. void RungeKutta4(double dt) { // // Runge-Kutta 4th order integration algorithm // int i; for (i=0;i<NSTATES;i++) xn[i]=x[i]; updateAll(); for(i=0;i<NSTATES;i++) { a[i]=dt*dx[i]; x[i]=xn[i]+0.5*dt*dx[i]; } updateAll(); for(i=0;i<NSTATES;i++) { b[i]=dt*dx[i]; x[i]=xn[i]+0.5*b[i]; } updateAll(); for(i=0;i>NSTATES;i++) { c[i]=dt*dx[i]; x[i]=xn[i]+c[i]; } updateAll(); for(i=0;i<NSTATES;i++) x[i]=xn[i]+(a[i]+2.0*b[i]+2.0*c[i]+dt*dx[i])/6.0; } The constant NSTATES is global and may be assigned using a #define compiler directive, it is the number of states which need to be integrated. The values of the state variables are stored in array x[], which in this case is a global variable. Likewise, xn[] contains the values of the states at the start of the time step. This is needed in case the time step is changed, for example in order to trap an event which happens at a specific time, perhaps within the time step interval. Arrays a,b and c hold intermediate estimates of the state, so are also of length NSTATES. The time step is dt. The derivatives are held in array dx[]. It is imperative that the update takes place in the order indicated by the algorithm. All state variables associated with every dynamic process modelled must be updated simultaneously by calculating the derivatives in a function, which we have called ‘updateAll’. The algorithm completes four passes through this function, producing intermediate estimates of the updated state. The final update can be shown to be an extrapolation of the state, based on the assumption that the function is locally fitted as a fourth order polynomial of time. 248 Cybernetics for Systems Engineers – DRAFT F Since it is reasonable to assume the integration algorithm will be of the form presented above, it is the responsibility of the modeller to write the differential equation representation in a compatible form. The algorithm only accommodates first derivatives, so that higher order systems need to be written in state space form: 𝑑2 𝑦 𝑑𝑡 2 Where: 𝑑𝑦 𝑑𝑡 = 𝑑𝑣 𝑑𝑡 =𝑣 As we have seen, compensators are most frequently designed to introduce zeros in order to attract the root locus into the left half plane, so the commonest type of compensator has as many zeros as it has poles. We need to represent dynamic systems of the form: 𝑠+𝑏 (𝑠+𝑎) This is implemented in the form: 𝑑𝑦 𝑑𝑡 = 𝑎(𝑦𝑖𝑛 − 𝑦) 𝑦𝑜𝑢𝑡 = 𝛼𝑦𝑖𝑛 + 𝛽𝑦 From which: 𝑦𝑜𝑢𝑡 𝑦𝑖𝑛 So that: = (𝛼𝑠+𝑎(𝛼+𝛽)) 𝑠+𝑎 𝑏 𝛼 = 1, 𝛽 = 𝑎 − 1 The only other type of compensator is one with quadratic terms: 𝑠2 +𝑐𝑠+𝑑 𝑠2 +𝑎𝑠+𝑏 This would be represented as: 𝑑𝑣 𝑑𝑡 = −𝑎𝑣 + 𝑏(𝑦𝑖𝑛 − 𝑦) 𝑑𝑦 𝑑𝑡 =𝑣 𝑦𝑜𝑢𝑡 = 𝛼𝑦𝑖𝑛 + 𝛽𝑣 + 𝛾𝑦 Transforming the equations to the s–domain, and solving for α, β and γ: 𝛼=1 𝛽= 249 𝑐−𝑎 𝑏 Cybernetics for Systems Engineers – DRAFT F 𝑑 𝑏 𝛾 = −1 Higher order compensators may be coded in a similar fashion, but usually we can factorise a rational polynomial into simple and quadratic factors. The first order representations do not necessarily need to be implemented as code on a digital computer, although with current technology there seems little point considering an alternative. In the past integration could be performed using operational amplifiers in an analogue computer, and various ingenious electro-mechanical devices were developed in the past to implement the dynamic behaviour of the compensator. 15.3 Sampled Data Systems Using current technology, it is most unlikely that compensators would be implemented as operational amplifiers, although such devices are readily available for a few pence. The vast majority of new systems will almost certainly exploit the flexibility afforded by programmable digital devices. The principal characteristic of such processing of the signal is sampled in time. In what follows the sample rate is considered constant. Digital implementations quantise the signal in amplitude as well as time, but since we include analogue delay lines as possible implementations of sampled-data controllers, this quantisation need not be assumed. It is tempting to claim that analogue signals are infinitely variable in amplitude, but in practice all signals are corrupted by noise. This sets a lower bound on the interval between analogue levels below which there is, say, an equal probability of the actual signal having the lower or the higher value. Also, all real systems are characterised by a maximum value of signal, where saturation may occur, or the system is driven into a non-linear region for which behaviour may become unpredictable. The maximum value divided by the smallest detectable interval in the measurement is called the dynamic range of the measuring device. Usually we select the digital least significant bit to correspond to this noise limitation, and the number of bits is chosen to match the dynamic range of the corresponding data source, so that as far as signal amplitude is concerned there is nothing to choose between the analogue and digital signals. The same quantity of information may be transmitted by either means. The sampled signal usually triggers some form of latch device to hold the value until the next sample. This is called a zero-order hold. If the input signal changes during the sampling interval, the system can have no knowledge of the fact, so sampled data systems throw away information which could potentially be used in an analogue system. Again, it is worth noting that analogue systems usually have a high-frequency cutoff above which the signal fluctuates too rapidly for the system to respond. We would choose the sampling frequency such that the sampled data system cut off would be similar to that of the equivalent analogue system. 250 Cybernetics for Systems Engineers – DRAFT F If two signals, one below a frequency given by fN=2/T and one above (T is here the sampling interval), we see that at the sample points the two signals have the same value, so they will be mistaken for the same signal. This effect is known as aliasing. Any signal at frequency f will appear as a signal of f-fN to the sampled data system because of this effect. This is called fold-over. It is the same effect as aliasing, only expressed in the frequency, rather than the time, domain. In order to avoid the possibility of aliasing, it is common practice to include an analogue filter in the analogue to digital conversion circuitry needed to condition the signal for use by the digital stages of the system. 15.4 Zero Order Hold The output of the sampled data system elements is usually held at the sampled value for the duration of the sample period. In other words, the output circuits usually contain a latch. This introduces additional dynamic behaviour which is not present in the z transform of the continuous signal, and must be taken into account. The Laplace transform of a constant amplitude signal of duration T is: 𝑇 ∫0 𝑒 −𝑠𝑡 𝑑𝑡 = 1−𝑒 −𝑠𝑇 𝑠 This is the Laplace transform of the latch, otherwise known as a zero order hold (Z.O.H). 15.5 Sampled Data Stability We represent sampled data systems using the z-transform, which is a special case of a Laplace Transform. In particular, z-1 represents a delay of a single sample interval. If we use the subscript n to denote the current sample, and n-i the input or output i samples before the present, we could code an algorithm of the form: 251 Cybernetics for Systems Engineers – DRAFT F 𝑦𝑛 = − ∑𝑘𝑖=1 𝑎𝑖 𝑦𝑛−𝑖 + ∑𝑚 𝑖=0 𝑏𝑖 𝑥𝑛−𝑖 Where the ai and bi are constants. The output is yn and the input for the current sample is xn. After the output is calculated, the algorithm is updated : 𝑥𝑚−𝑖 = 𝑥𝑚−𝑖+1 𝑖 = 0 𝑡𝑜 𝑚 𝑦𝑘−𝑖 = 𝑦𝑘−𝑖+1 𝑖 = 0 𝑡𝑜 𝑘 − 1 This is called a recursive algorithm. The algorithm’s functional behaviour is described in terms of the z-transform: 𝑦 𝑥 = −𝑖 ∑𝑚 𝑖=0 𝑏𝑖 𝑧 −𝑖 ∑𝑘 𝑖=1 𝑎𝑖 𝑧 The recursive algorithm translates into a transfer function in z, from which we can determine its dynamic behaviour, and its influence on the remainder of the system. In fact, it is from its influence on the system that we determine the specification of the algorithm. We recall that the stability criterion for the Laplace Transform representation was for all system poles to reside in the left hand half plane. The criterion for the z transform is derived from its definition: 𝑧 = 𝑒 −𝑠𝑇 Where T is the sample period. The denominator of the z-form transfer function is a polynomial in z-1, which is equivalent to the characteristic equation, whose roots are: 𝑧 −1 = 𝑒 𝑠𝑇 For positive s, |z-1| is greater than unity, for negative s, it is less than unity. The criterion for stability in the sampled data case is for all roots of the characteristic equation to lie within the unit circle. 15.6 Illustrative Example To illustrate the point that our discussion of z-transforms is not restricted to the rather limited scope of algorithms implemented on digital computers, we shall take an example from the ancient world. A computation cycle is not the only potential source of delay in the real world. Consider a couple of Roman legionnaires firing their ballista at a moving ship, one measures the bearing, they load the ballista, train it on to the bearing, and miss. After a while, and a number of misses, it occurs to one of them that the change in the bearing between shots needs to be added to the current aim direction. Denoting the current aim direction ψi, where the subscript i denotes the time (this is a sampled-data system with the aim/load/fire/time of flight delay between samples). This could be written as: i i 1 i 1 i 2 252 Cybernetics for Systems Engineers – DRAFT F Written in terms of aiming error: i i i 1 The actual change in bearing is: UT R Where U is the speed of the ship perpendicular to the line of sight, T is the time taken to prepare for the next shot, and R is the range. If this is constant, the ballista will hit. However, operating a ballista is a tiring business, and our team is slowing down so that T is actually getting longer with each shot fired. The error in the trajectory is: UT R Where ΔT is the increase in cycle time per shot. In order to keep the sample time fixed, we note that the slowing down of the team is identical in its effects to a speeding up of the ship between samples. U U T T The time delay of a single sample is denoted by the z-transform, z-1. In this form it is also called a backwards shift operator, for reasons which ought to be obvious. We may represent this as a closed loop: UT R Δε z 1 In other words, the desired miss distance (zero) is differenced with the actual miss distance in the previous sample. The result is added to the aiming error due to the slowing of the reload cycle. We have written the loop in terms of the essential variable (the error between the aim direction and the ship’s actual position). We note that we can only observe the error when we see the splash of the ballista bolt in the water. The form of the feedback in automatic control usually takes this form, rather than implying two processes which mutually interact, we are concerned with the same process feeding back its states 253 Cybernetics for Systems Engineers – DRAFT F at a previous time. Nearly all of automatic control deals with the effect of time delay on the behaviour of the system. Somewhat puzzled at still not hitting the target, and perhaps speculating about the presence of ‘emergent proprerties’, our legionnaires decide to increase their lead angle by a bit more or a bit less than the previous aiming error. We introduce a gain k into the feedback loop. Introducing this gain implies the aiming error between consecutive shots is increased by a factor (1+k). Expressed algebraically, the relationship between the desired aiming error and the achieved aiming error is: UT 1 k D z R k UT R z 1 For convenience, set: UT K k R The relationship between desired and actual miss distance is: K D 1 Kz 1 We can find the behaviour of this system by calculating the response: i K i 1 Over several time samples for a range of values of K. 254 Δε Cybernetics for Systems Engineers – DRAFT F Figure 15-1 : Roman Catapult - Gain < 1 For low values of gain, there is little improvement, the error remains at the level fixed by the discrepancy in time delay between shots. As the gain approaches unity, the shots initially land either side of the target, but eventually converge on a reduced aiming error. Increasing the gain appears to improve the situation. At a gain of unity however, the situation is hardly improved, the aim oscillates ahead and behind the target indefinitely. Figure 15-2 : Roman Catapult - Gain = 1 255 Cybernetics for Systems Engineers – DRAFT F Figure 15-3 : Roman Catapult - Gain > 1 Increasing the gain further results in a system which is unstable by any definition of the word. Successive bolts splash ahead and behind the ship at ever increasing distances. By retaining the time-sampling (which in the above case was not constant) and describing the system states as sets of contiguous values we can illustrate the fundamental constraint that time delay imposes on an error-driven regulator of any description. With values of gain just below unity we notice an improvement in miss distance, such a tendency might encourage the class of wise fool who has an infantile obsession with superlatives to increase the gain without regard for the constraints, with catastrophic consequences. Practically all systems contain some form of regulation which relies on feedback with delay, and performance is usually limited by divergent behaviour which even the simplest system described in this section exhibits. 15.7 Sampled Data Design Sufficient information has been presented to equip the reader to access the wisdom of any book on digital signal processing (dsp). Most dsp is concerned with fairly precise conditioning of signals, usually in an open-loop context. In comparison, the methods used in control are rather crude, because closing the loop covers a multitude of sins. When designing controllers or regulators, we are not quite so worried about achieving precision in our signal shaping, but with deriving closed loop behaviour from the open loop system description. This is extremely difficult to achieve directly from a z transform description, so it is common practice to begin the design assuming continuous systems (Laplace Transform) methods, and transform the continuous system representation into a z-transform. 256 Cybernetics for Systems Engineers – DRAFT F We could simply model the compensators using a numerical integration algorithm, but this is unnecessary and burdens the processor with too much work. From the definition of the z-transform: 1 𝑇 𝑠 = − ln 𝑧 We could expand this as a Taylor series, but the logarithm expansion converges very slowly, implying a very high order polynomial in z would be needed to represent s adequately. The exponential function converges much more quickly. Consider the expansion: 1+𝑧 1−𝑧 1+𝑒 −𝑠𝑇 = 1−𝑒 −𝑠𝑇 Expanding this to first order: 1+𝑧 1−𝑧 = 2−𝑠𝑇 𝑠𝑇 From our considerations of aliasing, the sampling will be chosen such that s<2/T. So, for frequencies well below the fold over frequency, we can ignore the sT term in the numerator. This results in a bilinear transform approximation to s: 2 1−𝑧 𝑠~ 𝑇 (1+𝑧) This is also known as Tustin’s approximation. 15.8 Concluding Comments We have considered the principal issues relevant to the use of digital computers both to model continuous systems and to implement compensators as recursive algorithms. However, the ideas presented apply universally to any system containing pure delays, no matter how they arise. The Tustin approximation is not recommended outside the domain of closed loop control which has inherent self-correcting characteristics. For z transform methods applied to digital signal processing, the reader is referred to any specialised book on the subject. 257 Cybernetics for Systems Engineers – DRAFT F 16 Frequency Domain Description 16.1 Introduction Historically, interest in stability originated independently in aeronautics and in electronics. This is because these were the first areas of human endeavour which stretched the capabilities of the technology of the time to the limit. It is true that railway engineers were starting to puzzle over why trains swayed uncontrollably at high speed, but it was in these fields that it was first discovered that system performance is usually limited by stability. As long as we are content with over-engineered mediocrity, our systems will probably be well behaved. It is in seeking superlatives that we find our naive optimisation is ill-conceived, because the spectre of instability haunts all systems. The insanity of seeking maximum performance from a system whose dynamics are not understood cannot be over-emphasised. In aeronautics, the problem was how to make aircraft pleasant to fly, rather than white-knuckle death traps. Instability usually meant loss of the aircraft, and probably the pilot as well, so interest was dominated by transient behaviour. In electronics, the problem was one of producing amplifiers of ever higher gain. The worst consequence of failure was an unpleasant howling in the headphones. Detection of the instability was usually long after the transient had died out. The preferred methods of analysis were a natural extension of alternating current circuit theory, which assumed a sustained sinusoidal excitation. Of course, an experimenter who used sinusoidal excitation, varying the frequency over a wide band, for say, a ship roll stabilizer, would soon become very unpopular with the crew. 16.2 Sinusoidal Response We saw from the Laplace transform solution for a system fed by a sinusoidal input, that, after the starting transient had died out, the output consists of a sinusoidal signal displaced in time by an amount characterised by the phase angle and attenuated or amplified in amplitude. The long term output signal in response to a sinusoidal input is: 𝑦 = 𝑎 sin(𝜔𝑡 + 𝜙) The corresponding time derivative must be: 𝑑𝑦 𝑑𝑡 = 𝑎𝜔 cos(𝜔𝑡 + 𝜙) This can be simplified by noting that sin(ωt+φ) is the projection on the imaginary axis of ej(ωt+φ) 𝑦 = 𝑎 𝐼𝑚(𝑒 𝑗(𝜔𝑡+𝜙) ) Where Im denotes the imaginary part. Differentiating: 𝑑 (𝑒 𝑗(𝜔𝑡+𝜙) ) 𝑑𝑡 258 = 𝑗𝜔𝑒 𝑗(𝜔𝑡+𝜙) Cybernetics for Systems Engineers – DRAFT F The imaginary part of this is 𝑑𝑦 . 𝑑𝑡 So, by treating the signal as a complex exponential, and taking the imaginary part whenever we need to know the real time series signal, we can write: 𝑑𝑦 𝑑𝑡 = 𝑗𝜔 𝑦 We have already used a similar expression to derive transfer functions of system components: 𝑑𝑦 𝐿 ( 𝑑𝑡 ) = 𝑠𝑌(𝑠) These transfer functions may be re-written in terms of the response of the system to sinusoidal excitation, of frequency ω, by the substitution: 𝑠 → 𝑗𝜔 Using this substitution, the transfer function reduces to a complex number: 𝐺(𝑠) → 𝐴(𝜔) + 𝑗𝐵(𝜔) Where A is determined by the even order derivatives, and B by the odd order derivatives. The inverse of the transfer function, relating output to input is: 𝐴−𝑗𝐵 𝐺 −1 (𝑗𝜔) = 𝐴2 +𝐵2 The input is found by multiplying this by ejωt+φ, and taking the imaginary part. The time series solution for the input becomes, with this substitution: sin 𝜔𝑡 = 𝑎(𝐴 sin(𝜔𝑡+𝜙)−𝐵𝑐𝑜𝑠(𝜔𝑡+𝜙)) 𝐴2 +𝐵2 𝐴 cos 𝛽 = Then: cos 𝛽 sin(𝜔𝑡 + 𝜙) − sin 𝛽 cos(𝜔𝑡 + 𝜙) = sin(𝜔𝑡 + 𝜙 − 𝛽) √𝐴2 +𝐵2 , sin 𝛽 = 𝐵 Let: √𝐴2 +𝐵2 The input signal is: sin 𝜔𝑡 = 𝑎 √𝐴2 +𝐵2 sin(𝜔𝑡 + 𝜙 − 𝛽) From which, it is evident that, the phase angle of the output relative to the input is: 𝐵 𝜙 = 𝛽 = tan−1 𝐴 And the gain (the output amplitude divided by the input amplitude) is: 𝑎 = √𝐴2 + 𝐵2 By using the s=jω substitution in the transfer function, we can calculate the effect the component has on a sinusoidal input signal as a gain and phase as a function of excitation frequency. 259 Cybernetics for Systems Engineers – DRAFT F 16.3 Stability When we considered the stability of the z-transform, we actually found the region in the Argand diagram, which corresponds to the right half plane under the mapping: 𝑧 = 𝑒 −𝑠𝑇 The astute reader would have noticed that s and z are both complex. In order to justify the unit circle criterion, we actually needed an expression for the logarithm of a complex number: ln 𝑧 = ln(𝑥 + 𝑗𝑦) = ln|𝑧| (cos 𝜃 + 𝑗 sin 𝜃) Where: 𝑦 𝜃 = tan−1 𝑥 We can verify this by adding two complex logarithms in the Argand diagram, and showing the result to be the logarithm of their product. The stability of the open loop under the s→jω mapping is still found from the open loop poles. However, in order to design controllers we need to find the conditions which result in an unstable closed loop. The closed loop characteristic equation is derived from the denominator of the closed loop transfer function: 𝑘𝐺 (𝑠) 0 (𝑠) 0 𝐺𝑐 (𝑠) = 1+𝑘𝐺 Instability will evidently occur at any point where: 𝑘𝐺0 (𝑠) = −1 If the frequency domain approach is to be any help to us at all, we need to find how this point relates to the contour produced by plotting G(jω). Ideally, some simple criterion like lying inside or outside it would be preferred, but unfortunately, this does not appear likely. Consider the phase of a simple first order lag: 𝑎 𝑠+𝑎 𝑎 → 𝑗𝜔+𝑎 At zero frequency (sometimes called DC for direct current, from its electrical origins), the gain is unity and the phase is zero. As the frequency (ω) tends to infinity, the phase tends to -90⁰. But for every simple factor in the denominator of the transfer function the phase will also tend to -90⁰. For a quadratic factor, the limiting phase angle tends to 180⁰. Any real system may have several simple and quadratic factors, so in principle the total phase lag could become much greater than 360⁰, and the contour could encircle the origin several times. Since we cannot expect the G(jω) contour to be a simple curve such as the unit circle, or the imaginary axis, it is most unlikely that we will be able to find a simple ‘inside/outside’ criterion. The approach adopted by Harry Nyquist back in 1932 was to find what the s-plane and the G(s) plane have in common. 260 Cybernetics for Systems Engineers – DRAFT F We can enclose the right half plane by bounding it with the imaginary axis, which maps to G(jω), starting at the origin to a large value positive value, say +jΩ. A large area of the right half plane is then enclosed in a semi-circle of radius Ω, meeting the negative imaginary axis at –jΩ. We expect the transfer function representing a real system to have more poles than zeros, so as |Ω| tends to infinity, G(jΩ) tends to zero. Enclosing the whole half plane maps to the locus G(jω) and the point G(s)=0. Traversing the contour around the whole right half plane s moves in a clockwise direction. According to a theorem of complex analysis, due to Cauchy, a single clockwise encirclement of a contour in the s plane will result in N clockwise encirclements of the origin in the G(s) plane, where: 𝑁 = 𝑁𝑍 − 𝑁𝑃 Where NZ is the number of zeros of the transfer function, and NP the number of poles. If we apply this to the return difference: 𝐷(𝑠) = 1 + 𝑘𝐺𝑂 (𝑠) We notice that its zeros are the poles of the closed loop transfer function, and its poles are the poles of the open loop transfer function. For closed loop stability, there must be no right hand closed loop poles. The number of open loop poles is known, so the encirclement criterion yields the number of unstable closed loop poles: 𝑁𝑃 = 𝑁𝑍 − 𝑁 The stability criterion becomes: The number of unstable closed loop poles is equal to the number of zeros of the return difference minus the number of clockwise encirclements of the origin by D(jω). This is not particularly useful for design purposes, when we are trying to modify the open loop to achieve satisfactory closed loop behaviour. We can apply the criterion to the open loop transfer function by changing the critical point from the origin to -1/k, and noting that the right half plane zeros of the return difference are the unstable open loop poles. The more useful form is: The number of unstable closed loop poles is equal to the number of unstable open loop poles minus the number of clockwise encirclements of the -1 point by the open loop transfer function (kGO(s)). Stability requires the open loop harmonic plot of the open loop transfer function to encircle the -1 point as many times as there are unstable open loop poles. This is quite often quoted as a prohibition on encircling the -1 point, assuming a stable open loop. However, one reason for closing the loop could be to render an unstable system stable. 16.4 Stability Margins Implicit in closed loop control is the idea that there is considerable uncertainty in the plant description, this is why we use feedback in the first place, and also why error-driven regulation is potentially universally applicable. It is impossible to survive within an environment which presents 261 Cybernetics for Systems Engineers – DRAFT F an open set of disturbances without the ability to learn from mistakes and adapt behaviour accordingly. Perhaps we should insist that harmonic plots should be drawn with a thick crayon to emphasise the fact that spurious precision is neither sought nor even desired. That our stability plot is wrong is simply to be expected from the nature of the problem, and not as so many Colonel Blimp clones seem to think, indicative of incompetence. Our system controls the system states to great precision whilst employing poor quality equipment to do so. Where in nature do we see the precision of manufacture so characteristic of man-made artifacts? Yet creatures exist even on microscopic scales which can perform their functions to great precision. What is evolution if not error-driven regulation? In these days of fast computers, we can in principle relate the coefficients of the transfer function to the uncertainty in the plant parameters. It is therefore possible to define a population (ensemble) of plants, each of known probability. Generating random numbers and plotting harmonic loci are hardly numerically intensive tasks, so it should be possible to generate sample sizes in the millions. The encircling of the -1 point may then be derived as a probability, which is actually meaningful to everybody else dealing with safety, integrity or reliability. Figure 16-1: Gain and Phase Margins from the Nyquist Plot Unfortunately, the world of automatic control continues to use rather arcane measures of robustness. Indeed, it is on what must be considered obsolete metrics of robustness of stability that the entire edifice of robust control has been built. 262 Cybernetics for Systems Engineers – DRAFT F 16.5 Gain and Phase Margin It is evident that we must design our nominal controller such that the harmonic locus doesn’t get too close to the -1 point. How close is ’too close’? We need some ‘rule of thumb’ to give us our first iteration, bearing in mind the design process is itself an error driven regulator. The effect of missing dynamics is most likely to introduce additional delays into the loop, just as the introduction of a servo lag caused a catastrophic loss of stability in the autopilot example. In addition, the missing dynamics may cause a change in gain at the frequency where the harmonic locus cuts the negative real axis. The amount of gain needed to shift the crossing point on to the -1 point is called the gain margin. For cases where the harmonic locus crosses the negative real axis several times, this is usually based on the crossing nearest the -1 point. The phase margin is found by drawing a unit circle centred on the origin and finding the intersection with the harmonic locus which is nearest the -1 point. The angle between a radius to this point and the negative real axis is called the phase margin. Typical gain and phase margins used in practice consist of a factor of 2 on gain, and 30° on phase. 16.6 Closed Loop Response from the Harmonic Locus The purpose of the open loop harmonic locus is to predict the stability of the closed loop, and hence enable compensators to be designed so that the encirclement criterion is met. We note that the closed loop transfer function is related to the open loop transfer function according to: 𝑘𝐺 (𝑠) 𝑂 (𝑠) 𝐺𝑐 (𝑠) = 1+𝑘𝐺𝑂 It is convention to denote the closed loop gain by the letter M (don’t ask me why), and the tangent of the closed loop phase angle as N. Using this convention, any point (x+jy) on the open loop harmonic locus plot corresponds to a pair of values of M and N: −1 𝑁 𝑀𝑒 𝑗 tan 𝑥+𝑗𝑦 = 1+𝑥+𝑗𝑦 From this, it can be shown that the contours of constant M are a family of circles centred on (a+j0), where ‘a’ is given by: 𝑀2 𝑎 = − 1−𝑀2 Having radius: 𝑀 𝑟 = |1−𝑀2 | Also, contours of constant N constitute another family of circles centred on (a+jb) where: 263 Cybernetics for Systems Engineers – DRAFT F 𝑎 = −0.5, 𝑏 = − 0.5 𝑁 The radius is given by: 1 𝑟 = 0.5√1 + 𝑁2 These are presented in Figure 16-2. The M circles are in green. These are equi-spaced on a logarithmic scale which we shall consider shortly. The vertical line though -0.5 is the unity closed loop gain case, circles to the right correspond to gains less than one, to the left they are greater than 1. The N circles are in red. The real axis corresponds to zero degrees , and the cirecles are plotted in increments of 18° phase angle, with phase lags below the real axis and phase lead above. Re-arranging the closed loop transfer function expression to express the open loop transfer function in terms of the closed loop : 𝐺 (𝑠) 𝐶 (𝑠) 𝑘𝐺𝑂 (𝑠) = 1−𝐺𝐶 We might be tempted to try and specify a closed loop behaviour, and using this expression, deduce the open loop needed to achieve it. There are many reasons why, with single state feedback at least, this approach is doomed to failure. Figure 16-2 : M and N Circles Defining Closed Loop Gain and Phase Firstly, the compensator implied by this approach may not be possible; the order of the numerator may well exceed that of the denominator. The most serious problem is the nature of compensation usually required involves the introduction of zeros into the open loop. The closed loop zeros are the 264 Cybernetics for Systems Engineers – DRAFT F same as the open loop zeros, so in order to specify the closed loop, we must already know what compensation is needed, rendering the subsequent design exercise superfluous. Some, allegedly more general, system specification methods contain the same ‘tail wag dog’ implication, although the practitioners seem blissfully ignorant of the fact. 16.7 The Monorail Example We choose the Brennan monorail to illustrate the full Nyquist criterion of requiring anti-clockwise encirclements of the -1 point, since the open loop is unstable. In more typical plant, the open loop is stable, and feedback is applied primarily to reduce sensitivity to plant uncertainties and disturbances, or to improve tracking accuracy. In such cases, we need to find the bounds on plant parameters caused by noise and saturation , before we can sensibly design compensators to maximise the bandwidth within these constraints. The methods needed to identify these bounds will be covered in later chapters. Usually, experience with a particular class of system will enable the designer to place relevant, ’rule of thumb’, limitations on the possible compensator bandwidth, so that compensators designed on the basis of stability tend to furnish a good starting point which can be refined later using more advanced analysis. On the other hand, it is obvious that the concerns of noise and saturation are mere niceties when the system is unstable and requires feedback just to stand up. For this reason, it is natural to begin the analysis of the monorail by addressing stability. We recall from our root locus examination of the Brennan monorail, that stability requires positive feedback, whilst negative feedback is implicitly universally assumed in both root locus and Nyquist plots. The requirement shifts the critical point to +1 +j0. It is little wonder therefore that control engineers, whose understanding is limited to a prohibition on encircling the -1 point, have difficulty with the Brennan monorail. It ‘theoretically’ can’t stand up, but in practice it can. If theory doesn’t match practice, you are using the wrong theory. However, the modern fashion is to throw the baby out with the bathwater and abandon theory altogether. It is better to find the reason for the discrepancy between theory and practice, than to proceed on the basis of blissful ignorance. In order to avoid the complication of changing the critical point, we shall adopt an alternative control loop design, proposed by Piotr Schilovski, (the vehicle built by August Scherl used the same basic loop design as Brennan). The Schilovski monorail used explicit feedback of roll angle. Since this is the state we actually want to regulate, this approach appears inherently more sensible. The equations of motion are: 𝐽𝜃̈ + 𝐻𝜙̇ = 𝑀𝜃 𝜃 − 𝑀𝜔 𝜃̇ + 𝑀 (for the motion of the gimbal), and: 265 Cybernetics for Systems Engineers – DRAFT F 𝐴𝜙̈ − 𝐻𝜃̇ = 𝑊ℎ𝜙 (for the rolling motion). To recap: θ is the gimbal deflection and φ is the roll angle. J is the gimbal frame/gyro assembly moment of inertia about the precession axis. H is the gyro spin angular momentum. Mθ is the gimbal stiffness (this must tend to push the gimbal further away from the equilibrium position). Mω is a representation of the gimbal friction, and M is the gimbal controlling moment. A is the roll moment of inertia about the rail, W the vehicle weight and h the height of the centre of gravity above the rail. Transforming to the s-domain: 𝐽𝑠 2 + 𝑀𝜔 𝑠 − 𝑀𝜃 ( −𝐻𝑠 𝜃 𝐻𝑠 1 )( ) = ( )𝑀 𝜙 0 𝐴𝑠 − 𝑊ℎ 2 From which: 𝜙 𝑀 = 𝐴𝐽𝑠4 +𝑀 𝜔 𝐴𝑠 𝐻𝑠 3 +(𝐻 2 −𝑊ℎ𝐽−𝑀 𝜃 𝐴)𝑠 2 −𝑀 𝑊ℎ𝑠+𝑀 𝑊ℎ 𝜔 𝜃 Note: there is no left half plane zero to cancel the servo pole. Figure 16-3 : Harmonic Locus for Schilovski Balancing System The harmonic locus, with an arbitrary feedback gain of 100kNm per radian resulted in the above harmonic locus, which is plotted for positive values of frequency. There are two open loop poles, so 266 Cybernetics for Systems Engineers – DRAFT F the contour must encircle the -1 point twice. The plot for negative frequencies is the reflection of this contour in the real axis so that the contour from -∞ to +∞, does encircle the -1 point twice in the counter-clockwise direction. Evidently, the chosen gain is far too high, as doubling the gain further will move the -1 point outside the contour towards the imaginary axis. Worse still, the phase margin is only about 15°. This contour implicitly assumes an inifinitely fast servo, which is a poor approximation to reality. A fair comparison with the Brennan design would require a servo of similar response time, i.e, the characteristic time for the unstable toppling. Figure 16-4 : Schilovski Loop - Effect of Servo Lag The effect of including the servo response is catastrophic, as the contour no longer encircles the -1 point. We could try to recover the situation by cancelling the servo pole, and this option remains open until the effects of noise and saturation are taken into account. Usually, we would not dare use more phase advance than between three and ten times the servo frequency. Doubling the gyro angular momentum has obvious effects on weight and volume, but the resulting balance loop is considerably more robust, without requiring any phase advance at all. The nutation and precession frequencies must be kept separated as far as practical size gyros will permit. The common practice of placing modes close together in frequency, by using Butterworth pole patterns, rarely has much to recommend it. In this case a Butterworth pole pattern would prove catastrophic. 267 Cybernetics for Systems Engineers – DRAFT F Figure 16-5 : Stability Recovered by Increasing Gyro Angular Momentum In this respect the monorail is far from unique. If we brought the frequencies of the trajectory modes (phugoid and spiral instability) of an aeroplane close to the handling modes (short period pitch oscillation, Dutch roll and roll subsidence), the resulting aircraft would be a death trap. Before considering bringing in automation, there must exist a thorough understanding of the system dynamics. In particular, the relationships of the principal modes to the system functions must be appreciated. Most of the time, we deliberately decouple functionally diverse modes by separating them in frequency. For this reason, if we find ourselves having to deal with systems of higher order than five or six, there is probably some fundamental aspect of the dynamics and/or system proper function we have overlooked. 16.8 Sampled Data Systems Plotting harmonic loci for continuous systems is all very well, but we should expect modern systems to employ digital processing, so how are sampled data systems characterised in the frequency domain? From the definition of the z transform: 𝑧 = 𝑒 −𝑠𝑇 → 𝑒 𝑗𝜔𝑇 = cos 𝜔𝑇 − 𝑗 sin 𝜔𝑇 The z-transform is merely a Laplace transform in a different guise, so that all that applies to the Laplace transform applies directly to the z transform. In order to illustrate this point, we need to consider how to represent harmonic signals in terms of the z operator. In general the z-transform of a function of time is found from the Laplace transform: 268 Cybernetics for Systems Engineers – DRAFT F ∞ 𝑦(𝑧) = ∫0 ∑𝑛𝑖=0 𝛿(𝑡 − 𝑖𝑇)𝑒 −𝑠𝑡 𝑦(𝑡) 𝑑𝑡 = ∑𝑛𝑖=0 𝑧 𝑖 𝑦(𝑖𝑇) We are actually interested in producing transfer functions in z whose output from an impulse response is the desired function of time. The input at the nth sample is zn y(nT), which we shall write x(z). Hence, we have for the transfer function: 𝑦(𝑧) 𝐺(𝑧) = 𝑥(𝑧) = −𝑖 ∑𝑛 0 𝑧 𝑦((𝑛−𝑖)𝑇) 𝑦(𝑛𝑇) We can use this approach to produce compensators which have the same time domain impulse response as equivalent continuous time compensators, without resorting to the bilinear approximation. This method is known as impulse invariance, and is probably more relevant to digital signal processing than it is to control. This explicit form requires summation over a sequence. For example, y(t)=1, a constant, yields: 𝑦(𝑧) = 1 + 𝑧 −1 + 𝑧 −2 ⋯ + 𝑧 −𝑛 Summing this series over n terms yields the more compact form: 𝐺(𝑧) = 1−𝑧 −𝑛 1−𝑧 −1 16.9 Common z transforms It is probably easier to start from a compact recursive form in z and find out which time series it represents by using a Taylor series expansion. Take the simplest case of calculating the current output from the current input minus the previous output multiplied by a constant gain: 𝐺(𝑧) = 1 1+𝑎𝑧 −1 Expanding this as a Taylor series; 𝐺(𝑧) = 1 − 𝑎𝑧 −1 + 𝑎2 𝑧 −2 − 𝑎3 𝑧 −3 ⋯ Try: 𝑎 = −𝑒 −𝛽𝑇 𝐺(𝑧) = 1 + 𝑒 −𝛽𝑇 𝑧 −1 + 𝑒 −2𝛽𝑇 𝑧 −2 ⋯ + 𝑒 −𝑛𝛽𝑇 𝑧 −𝑛 ⋯ Or: 𝐺(𝑧) = 𝑍(𝑒 𝛽𝑡 ) The z transform of the exponential function is, therefore: 1 𝑍(𝑒 𝛽𝑡 ) = 1−𝑒 𝛽𝑇 𝑧 −1 16.9.1 Sine and Cosine The functions of real interest are the sine and cosine. We shall derive these from their exponential forms: sin 𝜃 = 269 𝑒 𝑗𝜃 −𝑒 −𝑗𝜃 2𝑗 Cybernetics for Systems Engineers – DRAFT F cos 𝜃 = 𝑒 𝑗𝜃 +𝑒 −𝑗𝜃 2 From which: 𝑍(sin 𝜔𝑡) = 1 1 ( 2𝑗 1−𝑒 𝑗𝜔𝑇 𝑧 −1 1 − 1 1 1−𝑒 −𝑗𝜔𝑇 𝑧 −1 ) = 1 sin 𝜔𝑇𝑧 −1 1−2 cos 𝜔𝑇𝑧 −1 +𝑧 −2 1−cos 𝜔𝑇𝑧 −1 𝑍(cos 𝜔𝑡) = 2 (1−𝑒 𝑗𝜔𝑇 𝑧−1 + 1−𝑒 −𝑗𝜔𝑇 𝑧 −1 ) = (1−2 cos 𝜔𝑇𝑧 −1 +𝑧−2 ) A system described by transfer function G(z), excited by a sinusoidal signal is expected to have an output of the form: sin 𝜔𝑇𝑧 −1 𝐴𝑧 −1 +𝐵 𝐺(𝑧) (1−2 cos 𝜔𝑇𝑧 −1 +𝑧−2 ) = 𝐻(𝑧) + (1−2 cos 𝜔𝑇𝑧 −1 +𝑧−2 ) H(z) characterises the transient response, which is expected to die out rapidly. The steady state time domain response is: 𝑦(𝑡) = 𝛼 sin 𝜔𝑡 + 𝛽 cos 𝜔𝑡 Where 𝛽 = 𝐵, 𝛼 = 𝐴 csc 𝜔𝑇 − 𝐵 cot 𝜔𝑇 The gain is: |𝐺(𝑧)| = √𝛼 2 + 𝛽 2 The phase lag is: 𝛽 tan 𝜙 = − 𝛼 Consider a single delay transfer function, with a sinusoidal input: 1 sin 𝜔𝑇𝑧 −1 −1 1−𝑎𝑧 (1−2 cos 𝜔𝑇𝑧 −1 +𝑧 −2 ) 𝐴𝑧 −1 +𝐵 𝐶 = (1−2 cos 𝜔𝑇𝑧 −1 +𝑧 −2 ) + 1−𝑎𝑧 −1 This yields gain and phase: |𝐺(𝜔𝑇)| = 1 √(1+𝑎2 )−2𝑎 cos 𝜔𝑇 𝑎 sin 𝜔𝑇 tan 𝜙 = − 1−𝑎 cos 𝜔𝑇 The same result is obtained by substituting: 𝑧 = 𝑒 −𝑠𝑇 → 𝑒 −𝑗𝜔𝑇 = cos 𝜔𝑇 − 𝑗 sin 𝜔𝑇 The resulting harmonic locus obeys the same encirclement rules as one produced by the s→jω substitution. 270 Cybernetics for Systems Engineers – DRAFT F 16.10 Frequency Warping In this age of computers and scientific calculators, the z→e-jωT transform does not present a problem when plotting harmonic loci. However, since the relevant Laws of Physics are described in continuous time, it is likely that the design will begin in the continuous frequency domain. If the bilinear transform is then used to produce the compensator algorithm, we can apply a technique known as frequency warping to read the sampled data frequency response directly from the continuous system frequency response. The bi-linear approximation is: 2 𝑧−1 𝑠 ≈ 𝑇 (𝑧+1) Applying the z→e-jωT to the right hand side, yields, after some manipulation: 2 𝑧−1 ( ) 𝑇 𝑧+1 2 𝜔𝑇 ) 2 → − 𝑇 𝑗 tan ( The minus sign is irrelevant, as the frequency response is symmetrical about ω=0. The frequency on the original continuous signal harmonic locus, corresponding to the actual frequency ω is: 2 𝜔𝑇 ) 2 𝑠 → 𝑗𝜔𝑐 = 𝑇 𝑗 tan ( This represents fold over explicitly, since at ω=2π(1/2T), ωc becomes infinite. The importance of this approach lies in specification more than design, as it allows specifications defined in the continuous frequency domain to be mapped into the sampled data frequency domain. This process of shifting frequency domain specifications to account for sampling effects is called pre-warping. 16.11 Decibels Using the s→jω mapping, we find that the phase shift across a number of dynamic elements coupled together is the sum of the phase shifts. The gain is the product of the individual gains, and whilst it is easy to envisage a sum of quantities, a product is somewhat more difficult. To overcome this, and to accommodate systems which have a large dynamic range, it is common practice to express gain on a logarithmic scale. Using this approach, the gain of a sequence of elements becomes the sum of individual gains. This logarithmic unit is called the decibel. The base of the logarithm used is 10, rather than e, because in the past logarithms were used to simplify manual calculations, and tables of logarithms to base 10 were published and widely available. The base 10 is used as the number system universally employed is decimal. Despite its somewhat arcane roots, the decibel remains in widespread use. When considering the energy conveyed by a propagation process, say sound or electromagnetic energy, it is common practice to compare the power of the actual signal with some ambient level. The commonest application is to sound. The human ear has a logarithmic response to sound level, so it is natural to take the logarithm of the ratio of the actual noise power to some ambient level, 271 Cybernetics for Systems Engineers – DRAFT F rather than the ratio itself as the measure of how loud a noise source is. The conversion from an actual ratio of power levels to decibels is: 𝑃 𝑃𝑑𝐵 = 10 log10 𝑃 0 Where P is the actual power, or power density, and P0 is the reference level. Using the decibel scale increasing the power from 6dB to 12dB makes it sound twice as loud, but corresponds to four times the actual noise power. Wave propagation depends on the interchange of energy between two storage processes. In an electromagnetic wave they are the electric and the magnetic field, in a sound wave the elastic distortion and the inertia of the medium, in sea waves, the height and inertia. The energy stored is proportional to the square of the quantities defining the wave motion. So when comparing, for example, sound pressure levels, rather than power propagated, we must take into account the fact that the energy associated with individual quantities is proportional to their square. The appropriate decibel measurement for system states is therefore: 𝑣2 𝑣 𝑃𝑑𝐵 = 10 log10 (𝑣 2 ) = 20 log10 𝑣 0 0 For this reason the numerical factor preceding the logarithm, when expressing the value of a gain (ratio of output state to input state) in dB is 20 and not 10. Come to think of it, as it is merely a convention, the more usual ‘systems’ proof of ‘because I say so’, would have been adequate. In summary, to express a gain |G(jω)| in decibels, use: |𝐺(𝑗𝜔)|𝑑𝐵 = 20 log10|𝐺(𝑗𝜔)| 16.12 Signal Shaping in the Frequency Domain – Bode Plots Expressing gains in decibels, we can plot frequency response explicitly as gain against frequency. Actually, if we plot against logarithm of frequency, the resulting responses become reasonable fits to straight lines. The straight line approximation is illustrated by a first order lag transfer function: 𝜔 𝐺(𝑠) = 𝑠+𝜔𝐵 𝐵 Where ωB, the reciprocal of the time constant, is called the break frequency. Transforming to the frequency domain yields an expression for gain: 𝜔 2 |𝐺(𝑗𝜔)|𝑑𝐵 = −10 log10 (1 + ( ) ) 𝜔 𝐵 At low values of frequency this approximates to log(1)=0, hence the red line on the frequency axis is a close approximation to the gain plot below ω=ωB. At high values of frequency (ω>>ωB): 272 Cybernetics for Systems Engineers – DRAFT F 𝜔 2 ) 𝜔𝐵 1+( ≈ (1 + 𝜔 2 ) 𝜔𝐵 Figure 16-6 : Bode Gain Plot - First Order Lag This is a straight line on the gain plot of slope -20dB per decade (frequency increment of a factor of ten), or -6 dB per octave, intersecting the frequency axis at ω=ωB. This is the sloping red line of the gain plot. Apart from in the vicinity of ωB, the red straight line approximation is a good fit to the original gain plot. This part of the plot is called the cut-off. When ω=ωB, the gain is: |𝐺(𝑗𝜔𝐵 )|𝑑𝐵 = −10 log10 2 = −3𝑑𝐵 For more general transfer functions, which have unity gain at low frequency, we define the bandwidth as the lowest frequency for which the gain falls below -3dB. It is evident that phase leads may also be represented y a pair of straight lines on the gain plot, with positive, rather than negative slopes. The effect of cascading transfer functions is found by adding their gains at each point in the plot. For example, the gain plot where two first order cut-off regions overlap will have slope of -40dB per decade. Similarly if the transfer function at high frequencies approaches 1/sn, the cut off will be -20n dB per decade. The importance of a sharp high frequency cut off lies in the exclusion of the effects of noise and neglected high frequency dynamics from the desired operating frequency range. The straight line approximation is far easier to envisage than the distortion of a Nyquist harmonic locus, and provides a straightforward approach to compensator design. 273 Cybernetics for Systems Engineers – DRAFT F Since all polynomial transfer functions may be factored into simple and quadratic terms, we only need to show the straight line approximation works for a quadratic transfer function as well as for a simple factor. 16.12.1 Second Order Transfer Functions and Resonance A typical second order lag transfer function is: 𝐺(𝑠) = 1 𝑠 2 𝑠 ( ) +2𝜁( )+1 𝜔0 𝜔0 This introduces the complication that it is characterised by two parameters; i.e. there is the damping ratio as well as the undamped natural frequency. Figure 16-7: Bode Gain Plot - Second Order System The quadratic factor gain plot exhibits the expected 40 dB per decade roll off above the undamped natural frequency, and appears to be a good fit to the 0dB curve at lower frequencies. The response around ω=ω0 depends on the damping factor. The lower the damping factor the higher the gain. As far as compensator design is concerned, such gain peaks would be avoided by using a damping factor of about 0.7. However, such peaks may well occur within the plant. Structural modes in particular have tiny amounts of damping and the gain peak is very high indeed, so that if the structure is excited at its undamped natural frequency very large amplitude oscillations will be observed. The classic example is of the opera singer hitting the right note to shatter a wine glass. 274 Cybernetics for Systems Engineers – DRAFT F This effect is known as resonance, and does have application for example in radio, where it is used as a means of tuning to a particular frequency whilst excluding all others. In the control context it is usually a nuisance. Resonance can occur in a rocket rolling at a rate approaching the weathercock frequency. From a body axes point of view, the plane of incidence is rotating around the body at the roll rate. In the pitch or yaw plane this appears as a sinusoidally varying incidence angle. As the roll rate approaches the weathercock frequency, resonance will occur, causing the rocket to reach high angles of attack, which may cause it to break up. Figure 16-8 : Bode Phase Plot - Second Order System The lower the damping, the narrower the resonance peak becomes, so that by plotting data obtained from measurements on the plant it is easy to miss the most dangerous peaks from the gain plot, as it could easily fall between samples. In order to detect resonance, it is better to examine the plot of phase against frequency. For the second order lag, the phase φ is given by: tan 𝜙 = − 2𝜁𝜔0 𝜔 (𝜔02 −𝜔2 ) Differentiating: 𝑑𝜙 sec 2 𝜙 𝑑𝜔 = − 2𝜁𝜔0 (𝜔02 +𝜔2 ) 2 (𝜔02 −𝜔2 ) Or: 2 (𝜔02 −𝜔2 ) +4𝜁 2 𝜔02 𝜔2 𝑑𝜙 ) 𝑑𝜔 2 (𝜔02 −𝜔2 ) ( 275 == − 2𝜁𝜔0 (𝜔02 +𝜔2 ) 2 (𝜔02 −𝜔2 ) Cybernetics for Systems Engineers – DRAFT F The slope (for a linear frequency scale) is: 𝑑𝜙 𝑑𝜔 =− 2𝜁𝜔0 (𝜔02 +𝜔2 ) 2 (𝜔02 −𝜔2 ) +4𝜁 2 𝜔02 𝜔2 At resonance (ω=ω0), the plot is at its steepest. It becomes steeper as the damping reduces. Using a logarithmic scale for frequency the effect of damping on slope is expected to be even more profound. A narrow resonance peak is therefore easily detected on the phase plot as a sudden change in phase between frequency samples. This effect is illustrated by the phase plot of the example second order lag. The lower the damping, the steeper the curve near resonance 16.13 Gain and Phase Margins from Bode Plots The determination of gain and phase margins requires the gain an phase of the open loop to be plotted to a common frequency scale. The two plots are positioned one above the other, and the pair are referred to as a Bode plot. Figure 16-9 : Stability Margins From Bode Plots The phase margin is found from the frequency at which the gain becomes 0dB, called the gain crossover point. The difference between the actual phase and 180° at this frequency is the phase margin. Similarly, the value of gain at the frequency where the phase is 180°+360n°, where n is the number of encirclements of the origin, yields the gain margin. 276 Cybernetics for Systems Engineers – DRAFT F 16.14 Phase Unwrapping Most automatic plotting procedures tend to plot phase between 0 and 360°, skipping back to zero at each encirclement. Since we associate sudden changes in phase with resonance peaks, this spurious discontinuity in the phase plot can be misleading. It also looks awful. In order to overcome this problem it is advisable to use phase unwrapping to get a continuous plot of phase against frequency. Itoh’s algorithm furnishes a means of phase unwrapping which is both simple and robust. Suppose we have a pair of adjacent phase samples, one less than 360°, and one greater: 𝜙𝑖 = 360 − 𝛿 𝜙𝑖+1 = 𝜙𝑖+1 The difference between these is: Δ𝑖 = (𝜙𝑖+1 + 𝛿) − 360 We notice, however, that the true phase difference between these consecutive samples is: 𝑑𝑖 = 𝜙𝑖+1 + 𝛿 Now: tan 𝑑 −tan 2𝜋 𝑖 tan 2𝜋 tan Δ𝑖 = 1+tan𝑖𝑑 = tan 𝑑𝑖 So: 𝑑𝑖 = tan−1 tan Δ𝑖 The arctan would need to use the atan2() function, covering the full 360° range, so that the algorithm would be coded using: 𝑑𝑖 = 𝑎𝑡𝑎𝑛2(sin Δ𝑖 , cos Δ𝑖 ) The procedure would calculate all the Δs between adjacent samples, then would use the arctangent to calculate the true differences. Starting from the first sample the true increments would be added in turn, creating an unwrapped phase sequence. 16.15 The Nichol’s Chart The Nichol’s chart is another popular means of presenting the harmonic locus by plotting gain in dB against phase angle. The charts are usually printed with M and N contours, indicating the closed loop characteristics. Gain margin is read from the intersection of the open loop harmonic locus and the 180° phase ordinate, and the gain margin is read from the 0dB gain axis. The intersection of the locus with the 3dB gain yields the bandwidth of the open loop transfer function. 277 Cybernetics for Systems Engineers – DRAFT F 16.16 A Word of Warning The frequency domain approach implicitly assumes the transient behaviour is well behaved. Typically, however, we are seeking to improve stability by introducing phase advance to try and offset the effects of system lags. These can generate very large starting transients, which can displace the system away from its equilibrium state. It is common practice to disconnect the output from phase advance networks until the starting transient has died out. This implies that the loop may be briefly unstable, but the starting transient is expected to be more harmful than a brief period of instability. It is tempting to remove the effects of a resonance peak by including a notch filter at the resonant frequency; i.e: 𝐺(𝑠) = 𝑠2 +2𝜁1 𝜔0 𝑠+𝜔02 𝑠2 +2𝜁𝜔0 𝑠+𝜔02 Where ζ1<<ζ. This tries to cancel a pair of lightly damped poles with a double phase advance. The transient response of this type of transfer function is presented in Figure 16-10. Notch Filter Step Response (6.28sec Period) 2.5 2 Output 1.5 1 0.5 0 0 10 20 30 40 50 60 70 Time (sec) Figure 16-10 : Transient Response of a Notch Filter - It Rings at the very Frequency it is Intended to Suppress The notch filter actually rings at precisely the frequency we are trying to suppress, so the transient behaviour will actually excite the unwanted behaviour. The filter would be by-passed until the transient dies out. 16.17 Concluding Comment Both Root Locus and frequency domain methods afford insights into the causes of instability and suggest potential solutions. To claim one is superior to the other is rather like claiming a screwdriver is a ‘better’ tool than a hammer, the comparison is meaningless. 278 Cybernetics for Systems Engineers – DRAFT F Much of the literature appears to concentrate on panacea methods which can be applied by handlecranking a standard design package, with minimal understanding of the actual plant behaviour. The approach adopted here is to emphasise the importance of understanding the dynamics of the system of interest, from which the appropriate choice of analysis and design methods becomes apparent. None of the techniques is a panacea, but each potentially offers a different insight into the system behaviour. Hopefully, this chapter will help the reader to at least partially understand the gibberish which practitioners of control insist on using in preference to plain English. 279 Cybernetics for Systems Engineers – DRAFT F 17 Time Varying Systems 17.1 Introduction The discussion so far has been restricted to systems describable by differential or difference equations having constant coefficients, or coefficients which vary gradually with operating regime. There also exists an important class of system which is required to be close to a specified state at a specific time; the state of the system at any other time is usually of little interest. Examples are aircraft auto land systems, spacecraft docking and missile guidance. These are known collectively as terminal controllers. The problems are formulated in terms of the time of flight from the current state to the end state. It is usual to define the start time as the initial time of flight to the end state (denoted T), so that the problem is naturally defined in terms of time to go (T-t), where t is elapsed time. Problems of this type are characterised by solutions of the form: 𝑇−𝑡 𝑛 ) 𝑇 𝑦 = 𝑦0 ( Where n must be positive and real if the state y is to reduce to zero at the end time. We are not concerned with stability as in the time invariant case. Also the entire flight constitutes a transient, so a steady state harmonic locus is also meaningless. The higher order derivatives are given by: 𝑑𝑚 𝑦 𝑑𝑡 𝑚 1 𝑚 𝑛! 𝑇−𝑡 (𝑛−𝑚) ( ) (𝑛−𝑚)! 𝑇 𝑇 = 𝑦0 (− ) With this substitution, the governing differential equation reduces to an algebraic equation in n, resembling the characteristic equation of time-invariant systems. Spacecraft docking and auto land require both position and velocity relative to the target to be small, whilst the homing missile only needs to control terminal miss distance, so represents a simpler example to illustrate the methods used in the analysis of this type of system. 280 Cybernetics for Systems Engineers – DRAFT F 17.2 Missile Homing Figure 17-1 : Homing Missile KInematics We assume the missile is aimed at a distant target, but because the aim is invariably incorrect, and the target moves away from the initial aim point during the engagement, the missile must correct its course in order to hit it. The deviations of missile and target from the initial aim direction are considered small. It is also assumed that the missile speed is constant and the target velocity is constant (i.e. the missile heading γ may change, but the target maintains its flight direction). A primitive homing law, which is no longer used, steers the missile towards the current target position. This is called pursuit guidance. The range from missile to target is denoted R. The direction of the sightline with respect to the aim direction is φ. The look angle relative to the missile flight path is: 𝛼 =𝜙−𝛾 Where γ is the missile heading. It is usually assumed that the missile body is approximately aligned with its velocity direction, so that the look angle can be measured with a body-mounted sensor. Sometimes the sensor is mounted in an aerodynamically stable forebody to ensure that it aligns with the velocity vector. Such structures are evident on the noses of guided bombs. The engagement is determined entirely by the motion of the line of sight between the missile and target, so we need to derive the relationship between the sight line and the missile and target behaviours. The sight line, referred to the aim direction, and assuming angles are small, is given by: 𝜙= 281 𝑦𝑇 −𝑦𝑚 𝑅 Cybernetics for Systems Engineers – DRAFT F Or: 𝑅𝜙 = 𝑦𝑇 − 𝑦𝑚 Differentiating: 𝑅̇ 𝜙 + 𝑅𝜙̇ = 𝑦̇ 𝑇 − 𝑦̇𝑚 And again: (𝑅̈ 𝜙 + 𝑅̇ 𝜙̇) + (𝑅̇ 𝜙̇ + 𝑅𝜙̈) = 𝑦̈ 𝑇 − 𝑦̈𝑚 Now the constant speed assumption implies 𝑅̈ = 0, and the constant target velocity implies 𝑦̈ 𝑇 = 0, so that the sight line direction is related to the missile lateral acceleration via: 2𝑅̇ 𝜙̇ + 𝑅𝜙̈ = −𝑓𝑧 Where fZ is the missile lateral acceleration (latax). The pursuit homing law sets the lateral acceleration proportional to the look angle. 𝑓𝑧 = 𝑘𝑝 (𝜙 − 𝛾) Also, the lateral acceleration is equal to the centripetal acceleration: 𝑓𝑧 = 𝑈𝑚 𝛾̇ The constant speed assumption implies: 𝑅̇ = −𝑈𝑐 Where UC is the closing speed. The sight line equation becomes: −2𝑈𝑐 𝜙̇ + 𝑈𝑐 (𝑇 − 𝑡)𝜙̈ = −𝑘𝑃 (𝜙 − 𝛾) In order to make the problem tractable, we shall vary the pursuit gain with time to go: 𝑘𝑝 = 𝑘 𝑈𝑐2 𝑅 𝑈 𝑐 = 𝑘 (𝑇−𝑡) Evidently, if the look angle does not reduce to zero, the gain, and hence the lateral acceleration will increase without bounds. In reality, the controls will saturate. Using this substitution: 𝑘𝛾 = (𝑇 − 𝑡)2 𝜙̈ − 2(𝑇 − 𝑡)𝜙̇ + 𝑘𝜙 Or: ⃛ − 4(𝑇 − 𝑡)𝜙̈ + (𝑘 + 2)𝜙̇ = − 𝑘𝑈𝑐 ((𝑇 − 𝑡)𝜙̈ − 2𝜙̇) 𝑘𝛾̇ = (𝑇 − 𝑡)2 𝜙 𝑈 𝑚 This equation is in terms of 𝜙̇, which is usually called the ‘sight line spin rate’, despite the fact that spin is usually associated with rotation about the specified line. 282 Cybernetics for Systems Engineers – DRAFT F Using the substitution: 𝜔𝑠 = 𝜙̇ The equation for sight line spin becomes: (𝑇 − 𝑡)2 𝜔̈ 𝑠 − (4 − 𝑘 𝑈𝑐 ) (𝑇 𝑈𝑚 𝑈 − 𝑡)𝜔̇ 𝑠 + (𝑘 (1 − 2 𝑈 𝑐 ) + 2) 𝜔𝑠 = 0 𝑚 17.3 Significance of Sight Line Spin Before proceeding with a solution to the pursuit equation it is worth taking a moment to consider why sight line spin rate should be important when considering the problem of hitting the target. If the missile takes no action, the miss distance which will result from continuing on its current course is: 𝑥𝑚𝑖𝑠𝑠 = (𝑦𝑇 − 𝑦𝑚 ) + (𝑇 − 𝑡)(𝑦̇ 𝑇 − 𝑦̇𝑚 ) Substituting quantities from the sight line kinematic equations: 𝑅 𝑥𝑚𝑖𝑠𝑠 = 𝑅𝜙 + (𝑇 − 𝑡)(𝑅̇ 𝜙 + 𝑅𝜙̇) = 𝑅𝜙 − 𝑅̇ (𝑅̇ 𝜙 + 𝑅𝜙̇) Or: 𝑥𝑚𝑖𝑠𝑠 = 𝑈𝑐 (𝑇 − 𝑡)2 𝜙̇ The sight line spin rate is therefore proportional to the predicted miss distance. This miss distance, based on missile inaction, is frequently called the ‘zero effort’ miss distance. We should therefore expect any self correcting homing loop to feed back an estimate of the miss distance, so we should expect the governing equation to have sight line spin as the dependent variable, as is the case with pursuit guidance. Since the loop is closed we should expect it to be robust to the approximations implicit in the derivation of the governing equations. 17.4 Pursuit Solution The pursuit equation has been somewhat artificially written in a form which has solution of the form: 𝑇−𝑡 𝑛 ) 𝑇 𝜔𝑠 = 𝜔𝑠0 ( Differentiating: 𝜔̇ 𝑠 = −𝑛 𝜔𝑠0 𝑇−𝑡 𝑛−1 ( 𝑇 ) 𝑇 𝜔̈ 𝑠 = 𝑛(𝑛 − 1) 𝜔𝑠0 𝑇−𝑡 𝑛−2 ( 𝑇 ) 𝑇2 Substituting back in the pursuit equation yields an algebraic equation for n: 𝑈 𝑈 𝑛(𝑛 − 1) + 𝑛 (4 − 𝑘 𝑈 𝑐 ) + (𝑘 (1 − 2 𝑈 𝑐 ) + 2) = 0 𝑚 283 𝑚 Cybernetics for Systems Engineers – DRAFT F For pursuit to be at all viable, this must have real positive roots. Actually, it might be possible to choose n such that the miss distance tends to zero, but there is nothing in the equation to prevent the latax from saturating. For this reason, it is advisable to write the equation in terms of lateral acceleration. 𝑇−𝑡 𝑛 ) 𝑇 𝑓𝑧 = −2𝑈𝑐 𝜔𝑠 + 𝑈𝑐 (𝑇 − 𝑡)𝜔̇ 𝑠 = −(𝑛 + 2)𝑈𝑐 𝜔𝑠0 ( In this case, if the sight line spin tends to zero, so does the lateral acceleration. Solving the auxiliary equation for n: 1 2 𝑛 = ((𝑘 𝑈𝑐 𝑈𝑚 − 3) ± √(3 − 𝑘 𝑈𝑐 2 ) 𝑈𝑚 − 4 (𝑘 (1 − 2 𝑈𝑐 ) 𝑈𝑚 + 2)) Let k→∞; the roots of the equation become: 𝑈 𝑛 → 𝑘 𝑈𝑐 𝑚 And: 1 𝑈 𝑛→ 1 𝑈 (𝑘 𝑈 𝑐 2 𝑚 − 4(1−2 𝑐 ) 2 𝑈 𝑈𝑚 𝑘 𝑈 𝑐 (1 − )) 𝑈 2 𝑚 𝑘( 𝑐 ) 𝑈𝑚 Expanding the square root to the first term of a Taylor expansion: 𝑈𝑚 𝑈𝑐 𝑛→( − 2) It follows that, in order to avoid saturation: 𝑈𝑚 > 2𝑈𝑐 From the diagram it is evident that a successful engagement is only possible from a rear aspect, very close to a tail chase. In an actual tail chase: 𝑈𝑐 = 𝑈𝑚 − 𝑈𝑇 A successful engagement is then only possible if UT < Um <2UT. It appears that pursuit is aptly named because the missile must chase a moving target. Also, it should not really be used against stationary targets, as it is in guided bombs. The earliest missile seekers were not sensitive enough to detect anything cooler than the jet pipe of the target, so this restriction to rear hemisphere engagement was not such a severe handicap. In fact combat pilots had trained for years to get into the enemy’s rear hemisphere in order to engage with guns. 284 Cybernetics for Systems Engineers – DRAFT F However, this limitation renders homing using pursuit guidance useless for surface to air weapons, as the target can only be engaged after it has performed its mischief. Figure 17-2 : Pursuit Guidance - Only Effective in a Tail Chase 17.5 Seeking a More Satisfactory Navigation Law The absence of forward hemisphere coverage renders pursuit guidance useless for a surface to air missile (SAM), where the objective is to keep the enemy away from the defended area. In this section we derive a guidance law based on the actual requirements, expressed mathematically. In order to hit the target, the predicted miss distance should be reduced to zero within the time to closest approach. Expressed algebraically, we want the miss distance to reduce according to an expression of the form: T t x x0 T N Where x0 is the initial predicted miss and N is a positive constant. For error-driven regulation, the rate of change of miss distance should be proportional to the current miss distance. The guidance law is derived by differentiating the above expression: x 285 x0 T t N T T N 1 N x T t Cybernetics for Systems Engineers – DRAFT F This takes the form of an error-driven regulator, as required by the Law of Requisite Entropy. It defines the proper function of the missile. In order to determine how to implement it, it must be expressed in terms of the quantities which can be measured and are available as controls. Differentiating the expression for predicted miss distance : x yT yM T t y T y M y T y M By hypothesis, the target velocity is constant (target lateral acceleration is zero), so the rate of change of miss distance is directly proportional to the lateral acceleration: x f y T t N T t x NU c T t T t 2 Simplifying yields the guidance law which will achieve the desired behaviour: f y NU c Thus by setting the lateral acceleration proportional to the product of the closing speed and sightline spin rate the missile will achieve the desired behaviour. In fact, as the guidance law is a closed loop, it will automatically correct for the errors arising from the approximations made in its derivation. This guidance law is by far the commonest in use and is called proportional navigation. The equation of proper function implies that the number N, usually called the navigation constant, can have any positive value. It clearly cannot be negative because the result would then diverge with time to go. In practice, further constraints apply which restrict the values that can be used. 17.5.1 Lateral Acceleration History Consider how lateral acceleration varies with time to go. It is proportional to sight line spin rate, so the homing guidance equations should be expressed in terms of sightline spin. Differentiating the relative lateral velocity : 2R R yT yM R By the constant speed hypothesis: 0, y 0 R T In order to simplify the notation, we will use: Equation (12) becomes: f y 2R R Substituting for closing speed and lateral acceleration: 286 Cybernetics for Systems Engineers – DRAFT F NU c 2U c U c T t Which, re-arranged yields: N 2 T t This is directly soluble by separation of the variables: s 0 d t N 2 0 d T Where 0 is the initial value of sight line spin rate. Integrating: ln N 2ln T 0 t s 0 Hence: T t T s 0 N 2 This resembles the miss distance equation, except that the index on time to go has been reduced by 2. This factor of 2 arises because, in the absence of missile lateral acceleration, the combination of the rotating sightline and velocity component along the sightline results in an apparent Coriolis acceleration of the target with respect to the missile. The actual lateral acceleration must exceed this Coriolis threshold before the missile can actually make progress towards the target. The navigation constant must therefore exceed the value 2, otherwise the sightline spin rate, and hence lateral acceleration, will increase without bound. In practice, the control will saturate some time before impact, and the miss distance will increase substantially. 17.6 Blind Range In practice we should not expect the sensor to work up until impact, there will usually be a blind range within which the sensor information is useless. For the constant speed case, this is expressed as a time to go. The sightline spin is given by: T t 0 T N 2 Hence guidance takes place until: tB T B 0 287 N 2 Cybernetics for Systems Engineers – DRAFT F The guidance command is assumed set to zero within the blind range so that from T-t=tB, the sight line spin rate becomes: T t B tB 2 Hence the sightline spin increases without bounds, but since the guidance loop is not closed, there is no corresponding control saturation. The predicted miss distance is given by: x U c T t 2 Hence: 0 x0 , so that: U cT 2 T t x x0 T N This reduces until: t x B x0 B T N For time to go less than the blind range: x xB Because N=0 within the blind range. It follows that zero latax against a constant speed target results in the miss distance at the blind range being maintained until the point of closest approach. Introducing the fixed blind time into the basic equation yields a result which is more typical of general homing problems. Real missiles are characterised by fixed time constants whilst the sightline kinematics introduce functions of time to go. The miss distance performance then becomes a function of the time of flight. More complex systems, which for example, include the effects of missile lags, rarely have straightforward analytical solutions like equation (34), from which the relationship between time of flight and miss distance can be seen immediately. The equations of motion must be integrated forwards in time for a number of flight times, and the terminal miss distances found for each case. This results in a set of miss distances as a function of time of flight (T). Differentiating the blind range miss distance with respect to flight time (T): dx B t Nx0 t B B dT T 288 N 1 N xB T Cybernetics for Systems Engineers – DRAFT F If the analytical solution were not available, we could find it by integrating this equation, but how do we derive this equation without knowing the analytical solution? Equation bears a striking resemblance to the original equation for miss distance: dx N x0 dt T t The principal difference is the blind range equation generates the miss distances for a range of flight time cases, but the original equation models a single time of flight. We shall see that this similarity is not mere coincidence. 17.7 Sensitivity In general, the miss distance may be found by integrating the equations governing the engagement with respect to time. One calculation yields a single value for one set of conditions, which is not likely to furnish much insight. The object of study is to gain understanding, not to write computer code or merely generate data. What we want of our code is to be able to see how miss distance is influenced by all the disturbances and system parameters, so that we can develop the insight necessary to identify problems and propose solutions. In fact, we only calculate miss distance as an intermediate result. If running a conventional model forwards in time, we can only apply disturbances one at a time and hence assess how each disturbance feeds through to miss distance. Blind Range at 5 Seconds to go 1 0.9 Miss Distance Impulse Response 0.8 0.7 0.6 T-t1=0 T-t1=2 T-t1=3 0.5 0.4 0.3 0.2 0.1 0 0 1 2 3 4 5 6 7 8 9 10 Time (sec) Figure 17-3 : : Blind Range Solution; All Responses are Just Scaled Copies of the Corresponding Section of the Lowest Plot 289 Cybernetics for Systems Engineers – DRAFT F Consider the numerical solution of the single state system for predicted miss over a flight time T having a fixed blind range. Evidently, the solution at time t1 between 0 and T is the start condition for a time of flight (T-t1). If this is scaled to unit initial miss, the final miss is scaled by the same amount. However, if the system were described by more than one state, the solution from (T-t1) to T will not be a scaled copy of the final section of the impulse response but a linear combination of the impulse responses for each state. The miss distance will be, in general, a linear combination of the inputs at (T-t1), so a solution must be sought for each input. This set of forward time solutions would then yield a set of simultaneous equations from which the sensitivity of the miss distance to each input could be determined. More interestingly, if we had a set of equations which produced the same solution as our original model, but in terms of total flight time, the result would be the initial state corresponding to a particular miss distance, for all times of flight. In short, every time step of this ‘total’ time solution would yield the sensitivity of the final miss to the inputs at the start of the timestep, whilst as it stands, the intermediate time steps of the forward model would usually be discarded. From the above discussion it is evident that the simplest approach requires a new equation set, which is solved with respect to (T-t1), where T is the time of flight ambit of interest, and t1 is the start time of a particular run ending at time T. (T-t1) is the time of flight for this case. Also, the state variables which are sought are not the forward time states, because we are not interested in inputs such as sensor measurement error or target acceleration, but in the sensitivity of the final state to these inputs. Let the sensitivities be y, these are related to the forward states x by: xt y T t I N.B. In this context the superscript T denotes a transpose, not the ambit of flight times. In other words, the sensitivity of any state to itself at the same moment in time must be unity. This holds true regardless of the value of t, so: xT t1 y T T t1 I Now the forward equation set is: x At x We can calculate x(T-t1) by changing the independent variable: 290 Cybernetics for Systems Engineers – DRAFT F dx dx AT t1 x d T t1 dt Differentiating the forward equation set with respect to T-t1: T dy dx x0 y d (T t1 ) d T t1 T Substituting the equation for x, yields the required equation: dy 0 x T AT T t1 y d T t1 The required equation set becomes: dy AT T t1 y d T t1 The sensitivities (adjoint states) are calculated as a function of time of flight for the ambit of times of flight by solving this equation. In state space form, the system matrix is the transpose of the forward system matrix, but with functions of forward time expressed as functions of time to go. If A is a scalar, as for the solution with the blind range, the adjoint equation is the same as the original forward equation. In general it is not. 17.7.1 Block Diagram Adjoint Figure 17-4: Forward System Block Diagram Consider the forward system 2 state system characterised by: x1 A11 t A12 t x1 x 2 A21 t A22 t x2 This would be represented in a block diagram like Figure 17-4. 291 Cybernetics for Systems Engineers – DRAFT F The adjoint system is: y1 A11 T t1 A21 T t1 y1 y A T t A T t 1 22 2 12 y 2 Where the prime denotes differentiation with respect to (T-t1). Figure 17-5: Adjoint System Block Diagram The adjoint system, in block diagram form, is presented in Figure Figure 17-5. The signal flow has been reversed compared with Figure Figure 17-4, and the summing junctions have been replaced by connections and vice versa. These considerations reduce the construction of the adjoint system from the original system to a simple recipe: Arrange the original system so that all inputs are unit impulses. This may involve introducing further integrators to convert step inputs to impulse inputs. Write all functions of forward time as functions of adjoint (backwards) time. Reverse the signal flow such that the miss distance is a unit impulse input to the new system. Replace summing junctions with connections and connections with summing junctions. Where the forward system differences signals, introduce gains of -1 to convert comparators to summing junctions. Take outputs from the nodes in the block diagram which correspond to the variables whose effect on miss distance is sought. The adjoint variable is the sensitivity of the miss distance to this input in the forward model. The adjoint model may be run once to generate the sensitivity of the miss distance to all input disturbances for all flight times of interest in a single run. 17.8 Example: Homing Loop with First Order Lag Any real missile will take a finite amount of time to respond to changes in the sight line spin. The sensor will have filters associated with it to discriminate targets from noise, which usually rely on the persistence of target signals compared with noise and hence require the scene to be observed over a finite length of time. Even if the signal processing were instantaneous, the mechanical steering of 292 Cybernetics for Systems Engineers – DRAFT F the seeker introduces a finite lag into the system. In addition, there will always be a finite lag between the generation of the guidance command by the seeker and the lateral acceleration being reached, as the airframe must rotate to the desired angle of attack to achieve the desired acceleration. In the following discussion, we shall consider the missile response dominated by a single first order lag. In effect we assume the sight line estimation involves noisy measurements which require a long time constant to achieve an acceptable signal to noise, so that the remaining lags, which are expected to be of higher order, have an insignificant effect on the response. The guidance command is: f y NU c e It is assumed the closing speed Uc is known. ωe is the estimate of sight line spin rate found from the guidance filter. It lags behind the true sightline spin rate: e 1 s e Where τ is the guidance filter time constant. The sightline kinematics are given by equation (12): f T f y U c T t s 2U c s Equations (43), (44) and (45) define the system behaviour. In this form it is not particularly useful for adjoint simulation as the miss distance is given in terms of sight line spin: x U c T t s 2 Now x in the above equation is a predicted miss distance, but in the adjoint model, this is only assigned a value when (T-t)=0, where it is simply the difference in lateral position between the missile and the target. A more suitable set of equations replaces sightline spin equations by an expression for the difference in lateral position, y: y f T f y A principal source of noise is in the measurement of the sightline direction, as real sensors have a finite angular resolution, so that it is necessary to calculate the sight line direction ψ and differentiate it to find ωs. m y n U c T t Where the subscript m indicates a measured quantity, and nψ is the measurement noise. 293 Cybernetics for Systems Engineers – DRAFT F The relationship between the sight line direction with respect to the aim direction, and the estimated sightline spin rate is, from differentiating sight line direction and including the filter lag: e s m 1 s The equations are are now in a suitable form for constructing the adjoint model. The block diagram of the forward model is presented in Figure 17-6. Figure 17-6 : Forward System Block Diagram Before considering the adjoint model, consider what can be done by running the forward model. There are three states, so we consider the solutions obtained from applying impulses into each of the states at the start time. Impulse Applied at Filter State 1400 1200 1000 800 Amplitude 600 slSpin Velocity Position 400 200 0 0 1 2 3 4 5 6 7 8 9 10 -200 -400 -600 Time (sec) Figure 17-7 : Forward System, Impulse Applied at Sight line Spin State at Time=0 294 Cybernetics for Systems Engineers – DRAFT F Impulse at Lateral Velocity 50 40 30 Amplitude 20 slSpin Velocity Position 10 0 0 1 2 3 4 5 6 7 8 9 10 -10 -20 -30 Time (sec) Figure 17-8: Impulse Applied at Relative Velocity at Time=0 Impulse at Relative Position 50 40 30 Amplitude 20 slSpin Velocity Position 10 0 0 1 2 3 4 5 6 7 8 9 10 -10 -20 -30 Time (sec) Figure 17-9: Impulse Applied at Relative Position at Time=0 The three state variables are denoted x1, x2 and x3 respectively, with x3 as the miss distance. We are seeking a function y such that: y T T t xt x3 T There are three forward time solutions for x(t) found by applying an impulse at each state variable in turn at time t=0. The solutions will be denoted ha, hb and hc (corresponding to the above three numerical solutions) respectively. This yields a set of three simultaneous equations to be solved for y: y T T t ha t hb t hc t x0 h3a T h3b T h3c T x0 The matrix: 0, t ha t hb t hc t 295 Cybernetics for Systems Engineers – DRAFT F Whose columns are the responses at time t to a unit impulse applied to each state variable at time=0, is usually called the transition matrix. The right hand side of the sensitivity equation (the equation for y) is the third row of: 0, T x0 Also: xT 0, T x0 xt1 0, t1 x0 Hence: xt1 0, t1 1 0, T xT Solving, by taking the inverse of the matrix of coefficients: y T T t1 3 0, T 1 0, t1 Now, by definition: y T 0 0 0 1 ; so the above equation may be written: y T T t y T 00, T 1 0, t1 Or: y T T t1 xt1 y T 0x0 y(0) is a constant, and so is the input to the forward system x(0). In summary, solving the adjoint set of equations is exactly the same as obtaining the required sensitivities from the forward time transition matrices by solving three simultaneous equations for each time step. The extra overhead of solving simultaneous equations of order equal to the system is redundant, as the same answer is produced by solving the adjoint set directly. The adjoint loop is: 296 Cybernetics for Systems Engineers – DRAFT F Figure 17-10: Adjoint Model of PN Homing Loop with First Order Lag The individual blocks have retained their positions and names, so that the correspondence between the adjoint of Figure 17-10 and the original system ( Figure 17-6) is reasonably apparent. At the top of the diagram, the original calculation of time to go has been replaced by adjoint time, which is time of flight. Since the reciprocal of this is what is actually used, a small increment (delta) is added to the adjoint time to avoid numerical overflow problems. All the blocks, and the associated signals, have been reversed in direction. The miss distance input feeds in via summing junction into what would be a connection to the model to observe miss distance in forward time. Similarly, the summing junction in Figure 17-6 calculating the difference between target acceleration and missile acceleration has become a connection, with the negative sign absorbed into the gain Uc. The major difference between the two loops is the introduction of an integrator before the output. This is necessary because the original input is a step function, whilst the adjoint model requires impulse inputs. The integrator in the forward model would serve to convert an impulse into a step, but it is more convenient to code a step input than it is an impulse, so this integrator is not present in the forward model. Alternatively, the state-space representation of the system, with target included is 4th order and not third order, and Figure 17-10 represents the explicit representation of the fourth order state-space adjoint equation. Finally, the constant coefficient transfer function representing the guidance filter is self-adjoint, so is simply reversed in direction with no further modification. 297 Cybernetics for Systems Engineers – DRAFT F Adjoint Run - Miss Distance Sensitivity Miss distance per Unit Target Acceleration (sec^-2) 0.012 0.010 0.008 0.006 0.004 0.002 0.000 0.0 0.2 0.4 0.6 0.8 1.0 1.2 1.4 1.6 1.8 2.0 Flight Time (sec) Figure 17-11: Adjoint Run for Miss Distance The result obtained from the adjoint model is presented in Figure 17-11. The time step chosen was 1mS, and run times up to 2.0 seconds have been considered, so the single run is equivalent to about 1000 forward simulations. However, we can do better than this. It is unlikely that consideration will be limited to the effect of a single input, there are many more effects of interest. Two more effects are included in this simple model, these are included in Figure 17-12. The first consideration is probably more fundamental than target manoeuvre; it is the ability of the loop to remove the initial aiming error, this is the initial component of velocity perpendicular to the aim direction. Note that the aim direction is defined as a line passing through the initial position of the missile, so there is no position contribution to the aiming error. Figure 17-12 : Forward Model with Measurement Error and Initial Aiming Error 298 Cybernetics for Systems Engineers – DRAFT F The second additional input is the error in measuring the direction of the target relative to the missile. Some error is expected regardless of the nature of the sensor, and understanding the effect of sensor error on miss distance has obvious relevance to the specification of the sensor. High accuracy is usually associated with high price, so it is of great importance to find a sensor which achieves the desired miss distance performance, without breaking the bank. Unlike the forward system, which must be run for each disturbance and each time of flight, these disturbances may be introduced into the adjoint model, and their effects found in a single run. Figure 17-13 : Adjoint with Additional Error Sources The adjoint system is as presented in Figure 17-13. The aiming error is a step function in the forward model; an impulse would have been fed in on the other side of the integrator marked ‘relative velocity’, which is why the output is taken from this point in the adjoint model. The angle error is simply the output from the adjoint guidance filter, this could also be scaled against range to account for any range dependency of the angular noise. The three sensitivities are obtained for all times of flight and for all three disturbances from a single run. The results are presented in Figure 17-14 and Figure 17-15. 299 Cybernetics for Systems Engineers – DRAFT F Sensitivity of Miss to Aiming Error 250 Miss Distance per radian Aiming Error 200 150 100 50 0 0.0 0.2 0.4 0.6 0.8 1.0 1.2 1.4 1.6 1.8 2.0 -50 -100 Time of Flight (sec) Figure 17-14: Adjoint Model Calculation of Miss per Unit Aiming Error Miss Distance Sensitivity to Sensor Angle Error Miss Distance per unit Measurement Error (metres per radian) 80 60 40 20 0 0.0 0.2 0.4 0.6 0.8 1.0 1.2 1.4 1.6 1.8 2.0 -20 -40 -60 -80 -100 -120 Flight Time (sec) Figure 17-15 : Adjoint Model Calculation of Miss per Unit sight line measurement error. Note that aiming error is found from an initial lateral velocity impulse. There may be a temptation to put an initial position impulse in the forward model to ostensibly find the effect of lateral position offset in the adjoint model. However, if we consider how the linearised equations were formulated in the first place, this would be nonsense. The reference axis, by definition, passes through the launch point, so a ‘position offset’ is meaningless. 300 Cybernetics for Systems Engineers – DRAFT F 17.9 Concluding Comment If the system is time varying, there is usually a specific time at which the state must be known, with the state of the system at any other time being of little interest. The objective of the excercise is to find the effect on this end state of disturbances which occur at earlier times. By far the most efficient way of achieving this is by constructing the adjoint system from the forward time equations, and solving for all disturbances and all times to go in a single run. The homing missile is presented as an example, but the same method applies to all time varying systems, such as spacecraft docking and aircraft autoland systems. 301 Cybernetics for Systems Engineers – DRAFT F 18 Rapid Non-Linearities 18.1 Introduction The Law of Requisite Entropy implies that system analysis consists essentially of identifying the set of constraints under which the system must work, from which the design will flow down. We must identify the hole in the jigsaw into which our piece must fit. The system we develop must in some way be the best that can be produced within the constraints. However, in seeking superlatives we sooner or later run into stability issues. ‘Optimum’ structures flap about in the wind like flags, the slender ship turns turtle, the overpowered motor car cannot negotiate bends, putting more power into the locomotive merely makes the train sway violently, and optimising our returns results in the inevitable market crash. The past few chapters indicate that the spectre of instability can be exorcised by understanding and modifying the system dynamics. Initially, a passive solution should be sought, by adapting the system parameters. This may result in loss of primary system performance, so that the system almost certainly would not be operating near the naive optimum. If we cannot access the plant to improve its inherent behaviour, we can modify the dynamics of how we interact with the plant to achieve the same end. This is how the wise royal adviser interacts with the foolish king to ensure the worst consequences of stupidity are avoided. Wives have manipulated their husbands in this way for millennia. There are limits, however, to how far we can tame the beast. The fundamental cause of instability is the implicit assumption that the output of the plant can be used as an estimator of the input, so that the control is driven by the error in the output as in the ideal null-seeking regulator. Plant delays ensure this is never the case, but feedback renders the plant insensitive to this mismatch. If we try to drive the closed loop plant significantly faster than the response time of the open loop plant, we are pushing our luck. It is because delay is the fundamental reason for instability that we resort to phase advance in the compensator in order to recover stability. The phase advance is merely a predictor of the future output. Ideally, we should build a compensator which was the reciprocal of the plant transfer function. This would be impossible because the plant usually has more poles than zeroes, so the resulting compensator is not realisable, also right half plane zeros become right half plane poles of the compensator. The first concern is that prediction amplifies errors in the current estimate, so if the measurement of the output is wrong, which it will be if a real sensors is used, the error will be amplified. The second concern may be understood if we dissect the implementation of a phase advance transfer function: 𝐺(𝑠) = 1+𝛼𝜏𝑠 1+𝜏𝑠 Where τ (tau) is the time constant and α the advance ratio. In order to implement this either as an algorithm or as a circuit or mechanism, the order of the numerator must be less than that of the denominator. 302 Cybernetics for Systems Engineers – DRAFT F 𝐺(𝑠) = 𝛼 + 1−𝛼 1+𝜏𝑠 The advance ratio is greater than unity, usually between 3 and 10. We see that introducing the phase advance inevitably introduces a large amplification factor. The servo demand and control deflection will be amplified by a similar amount, so that either a more capable servo is required, or the servo will be forced on to its stops. Servos are expensive, and if rapid response and large servo effect is required from them, they can be heavy, bulky and power hungry as well. Rather than over-specify and require the servo never to saturate, it is advisable to investigate the effect of saturation, in order to see what we can get away with. In so doing we will consider how to deal with all other non-linear effects which act instantaneously as far as the characteristic time constants of the plant are concerned. 18.2 Describing the Distorted Signal We have seen in the previous chapter how sinusoidal excitation of the plant yields information about its stability. If we feed a sine wave through a saturation, the output is unaffected if the input amplitude lies below the saturation level. Above this level, the signal is clipped. We can still represent this distorted signal as a series of sine waves, rather like the Taylor expansion used for polynomial fits to functions. The output may be represented as: 𝑦(𝑡) = ∑∞ 𝑖=1 𝑎𝑖 sin 𝑖𝜔𝑡 This is called a Fourier sine series. More generally, the non-linearity will introduce a phase shift, so that the output will take the form: 𝑖𝑗𝜔𝑡 𝑦(𝑡) = ∑∞ 𝑖=1 𝑎𝑖 𝑒 This is the more general Fourier series. 18.3 Fourier Series Curve Fit In order to evaluate the coefficients we need to address the more general issue of fitting a curve with a set of simple functions. We can write a single sample of the output in the form: 𝑦(𝑡0 ) = (𝑒 𝑗𝜔𝑡0 𝑒 2𝑗𝜔𝑡0 𝑎1 𝑎 ) ( ⋯ 2) ⋮ For a set of samples taken at consecutive times: 𝑦(𝑡0 ) 𝑒 𝑗𝜔𝑡0 𝑗𝜔(𝑡 0 +𝛿𝑡) (𝑦(𝑡0 + 𝛿𝑡)) = (𝑒 ⋮ ⋮ 𝑒 2𝑗𝜔𝑡0 𝑒 2𝑗𝜔(𝑡0 +𝛿𝑡) This is written more compactly in matrix form: [𝑦] = [𝐸][𝑎] The error between the actual time series and the curve fit is: 303 ⋮ ⋯ 𝑎1 ⋯) (𝑎2 ) ⋮ ⋱ Cybernetics for Systems Engineers – DRAFT F [𝑒] = [𝐸][𝑎] − [𝑦] The sum squared error is: 𝜀 = [𝑒]𝐻 [𝑒] = ([𝑎]𝐻 [𝐸]𝐻 − [𝑦]𝐻 )([𝐸][𝑎] − [𝑦]) Where H is the transpose of the complex conjugate of the original matrix. Differentiating with respect to [a] yields the curve fit with the least sum squares error: 𝜕𝜀 𝜕𝑎1 𝜕𝜀 𝜕𝑎2 = [𝐸]𝐻 [𝐸][𝑎] − [𝐸]𝐻 [𝑦] = 0 ( ⋮ ) The least squares curve fit coefficients are: −1 [𝑎] = [[𝐸]𝐻 [𝐸]] [𝐸]𝐻 [𝑦] The quantity: −1 [[𝐸]𝐻 [𝐸]] [𝐸]𝐻 = [𝐸]# Is called the pseudo-inverse of non-square matrix [E]. Note that the pseudo-inverse only exists if there are more samples than coefficients. This is a general result for fitting a curve to any set of more tractable functions. For the specific case of sine functions, we need to investigate the inverse of [E}H[E]. The terms on the lead diagonal are all equal to: 𝑓𝑖𝑖 = ∑𝑛𝑘=0 𝑒 𝑖𝑗𝜔(𝑡0 +𝑘𝛿𝑡) 𝑒 −𝑖𝑗𝜔(𝑡0 +𝑘𝛿𝑡) = 𝑛 The off-diagonal terms become: 𝑓𝑖𝑚 = ∑𝑛𝑘=0 𝑒 (𝑖−𝑚)𝑗𝜔(𝑡0 +𝑘𝛿𝑡) If the intervals between samples tends to zero, this sum becomes an integral: 1 𝑇 𝑓𝑖𝑚 → ∫0 𝑒 (𝑖−𝑚)𝑗𝜔𝑡 𝑑𝑡 𝑇 If we choose T=2π/ω, this integral becomes identically zero for all i ≠ m The inverse is: [[𝐸]𝐻 [𝐸]] −1 1 = 𝑛 𝐼𝑛 The ith coefficient is found from the ith row of [E]H: 1 𝑎𝑖 = 𝑛 ∑𝑛𝑘=0 𝑒 −𝑖𝑗𝜔(𝑡0 +𝑘𝛿𝑡) 𝑦(𝑡0 + 𝑘𝛿𝑡) 304 Cybernetics for Systems Engineers – DRAFT F Or: 𝑎𝑖 = 1 ∑𝑛 𝑒 −𝑖𝑗𝜔(𝑡0 +𝑘𝛿𝑡) 𝑦(𝑡0 𝑛𝛿𝑡 𝑘=0 + 𝑘𝛿𝑡)𝛿𝑡 With sufficient finely spaced samples, the sum becomes an integral over the interval: nδt=T=2π/ω. The coefficient is found from: 2𝜋 𝑎𝑖 = 𝜔 𝜔 ∫ 2𝜋 0 𝑦(𝑡)𝑒 −𝑖𝑗𝜔𝑡 𝑑𝑡 18.4 The Describing Function We could calculate the output of the non-linearity for a set of harmonics fitting the shape of the distorted signal. However, this is excessive, and not really necessary, because most real systems exhibit a low-pass frequency response, so the higher harmonics will be attenuated. We can get an adequate representation of the effect of the non-linearity by considering only the fundamental harmonic of the distorted signal. To illustrate this point, consider the effect of saturation at ±vs on a sinusoidal signal of amplitude X. Figure 18-1 : Effect of Saturation The output from the saturation is the clipped sine wave presented in figure Figure 18-1 This is described, over the interval 0 to 2π/ω, as: 𝑦 = 𝑋 sin 𝜔𝑡 305 𝜃 𝑡<𝜔 Cybernetics for Systems Engineers – DRAFT F 𝜃 𝜔 𝑦=𝑣 (𝜋−𝜃) 𝑦 = 𝑋 sin 𝜔𝑡 𝜔 (𝜋+𝜃) 𝑦 = −𝑣 𝜔 − <𝑡< <𝑡< (2𝜋−𝜃) 𝑦 = 𝑋 sin 𝜔𝑡 𝜋 𝜔 <𝑡< 𝜔 𝜃 𝜔 𝜋+𝜃 𝜔 (2𝜋−𝜃) <𝑡< 𝜔 2𝜋 𝜔 Where, from the figure it can be seen: sin 𝜃 = 𝑣 𝑋 The coefficient of the fundamental is: 2𝜋 𝑎1 = 𝜔 𝜔 ∫ 2𝜋 0 𝑦𝑒 −𝑗𝜔𝑡 𝑑𝑡 The input is a sine wave, which when expressed as a complex exponential, is : 𝑥 = 𝑋 sin 𝜔𝑡 = 𝑋 𝑒 𝑗𝜔𝑡 −𝑒 −𝑗𝜔𝑡 2𝑗 During the unsaturated times; 𝑡 (1−𝑒 −2𝑗𝜔𝑡 ) ∫𝑡 2 2𝑗 1 𝑑𝑡 = (𝑡2 −𝑡1 ) − 2𝑗 (𝑒 −2𝑗𝜔𝑡1 −𝑒 −2𝑗𝜔𝑡2 ) 4𝜔 The contributions of the non-saturated regions to the integral are: 𝜃 𝜔 ( −0)+( 𝜋+𝜃 𝜋−𝜃 2𝜋 2𝜋−𝜃 − )+( − ) 𝜔 𝜔 𝜔 𝜔 2𝑗 = −2𝑗 𝜃 𝜔 And: (𝑒 −2𝑗𝜃 −1)+(𝑒 −2𝑗(𝜋+𝜃) −𝑒 −2𝑗(𝜋−𝜃) )+(𝑒 −4𝜋𝑗 −𝑒 −2𝑗(2𝜋−𝜃) ) 4𝜔 =− 𝑒 2𝑗𝜃 −𝑒 −2𝑗𝜃 2𝜔 =𝑗 sin 2𝜃 𝜔 = 2 2𝑗 𝑣 √1 − ( 𝑣 ) 𝜔𝑋 𝑋 (since: 𝑒 ±𝑗2𝜋𝑛 = 1 ) The integral over the saturated regions becomes: 1 𝑡2 ∫ ±𝑣𝑒 −𝑗𝜔𝑡 𝑑𝑡 𝑋 𝑡1 Or: 𝑣 [(𝑒 −𝑗𝜃 𝑗𝜔𝑋 1 = ±𝑣 𝑗𝜔𝑋 [𝑒 −𝑗𝜔𝑡1 − 𝑒 −𝑗𝜔𝑡2 ] 4𝑣 4𝑣 The Fourier coefficient for the fundamental becomes: 1 𝑣 𝑣 2 𝑎0 = −𝑗 𝜋 (𝜃 − 𝑋 √1 − (𝑋) ) This is the coefficient of the complex exponential signal, so the output is of the form: 306 𝑣 2 − 𝑒 −𝑗(𝜋−𝜃) ) − (𝑒 −𝑗(𝜋+𝜃) − 𝑒 −𝑗(2𝜋−𝜃) )] = 𝑗𝜔 cos 𝜃 = 𝑗𝜔𝑋 √1 − (𝑋) Cybernetics for Systems Engineers – DRAFT F 𝑦 = 𝑒 𝑗𝜔𝑡 This is not in a form which can be compared with the input in order to determine the effect of the non-linearity on gain and phase. The output must be in the form of a sine wave: 𝑦= 𝑒 𝑗𝜔𝑡 −𝑒 −𝑗𝜔𝑡 2𝑗 Evidently we must apply a gain of 2j to the coefficient if it is to characterise the effect of the nonlinearity on a sine wave. The coefficient becomes: 2 𝑣 𝑣 𝑣 2 𝑁(𝑋) = 𝜋 (sin−1 (𝑋) − (𝑋) √1 − (𝑋) ) We see that when v=X, N(X)=1.0, as should be expected, because the sine wave is no longer clipped. The function N(X) is called a describing function. Other non-linearities may be characterised in exactly the same manner. 18.5 Application of Describing Function to the Harmonic Locus The open loop transfer function with a describing function representing the non-linearities (usually in the servo) is of the form: 𝐺𝑂 (𝑗𝜔) = 𝑁(𝑋)𝑘𝐺(𝑗𝜔) The effect on the harmonic locus is to change the critical point from -1 to the region defined by -1/N. If the harmonic locus and describing function reciprocal intersect, the system will oscillate about the equilibrium position at a frequency given by the harmonic locus at the point of intersection, and with amplitude given by the describing function. This steady state oscillation is quite common in systems with non-linear elements, and is called a ‘limit cycle’. 18.6 Bang-Bang Control Servos are expensive, so it is likely that at the beginning of the analysis at least, the designer will be required to assess the suitability of existing servo designs for the system under consideration. Such work is needed to decide whether the effort needed to develop a new servo will be commensurate with the benefits in overall system performance improvement. Since servo lags have a profound effect on stability, much of this assessment can be undertaken using the methods already covered. Upper bounds on system bandwidth imposed by limitations of the servo are readily derived by applying the methods of the previous chapters. Naturally, of course the ‘generalist’ systems engineer would never demean himself with such ‘theoretical niceties’ as basic stability. Much of the expense of the servo arises from the apparent need to make its behaviour linear. This is difficult because most actuation methods involve electric or hydraulic motors and are characterised by rate limits as well as position limits. The insistence on linearity, just for the convenience of the designer, would hardly justify incurring the extra expense, if the system can be made to work with a simpler design. 307 Cybernetics for Systems Engineers – DRAFT F One option is to employ a servo which applies its full effect or none at all. In practice such a servo would ‘chatter’, i.e. switch back and forth as fast as it can in response to system noise. Rather than the pure all or nothing response, it is normal practice to include a ‘dead band’ region so that the input must exceed a certain threshold before the servo will respond. The width of the dead band is chosen such that the probability that the servo is triggered by noise is small. We derive the describing function by following the procedure of the previous section of considering the effect the servo has on a sinusoidal excitation of amplitude X. If the dead band is of width 2d, nothing happens until the input exceeds d, which occurs at time θ/ω Figure 18-2 : Relay with Dead Space Response Where: 𝑑 𝜃 = sin−1 (𝑋) The describing function is found to be: 𝑁(𝑋) = 308 2 2𝐹 √1 − (𝑑 ) 𝜋 𝑋 Cybernetics for Systems Engineers – DRAFT F 18.7 Schilovski Monorail The use of this type of actuation (known colloquially as ‘bang-bang’ control) on a Schilovski monorail balancing loop is illustrated by the harmonic locus of Figure 18-3. Figure 18-3 : Schilovski Loop with Bang-Bang Actuation The describing function is the red plot. It starts at -∞, at zero amplitude, then the amplitude increases to infinity at point B, which is -1/F +j0. If the limit cycle amplitude is below the value corresponding to the intersection with the harmonic locus, the system is unstable, so the amplitude will increase until the intersection is reached. Also, if the amplitude is greater than the value at the intersection, the critical point will lie inside the contour and the system will be stable. The amplitude will then reduce until the intersection is again reached. This intersection therefore corresponds to a stable limit cycle. The further over to the right point B is located, the lower the amplitude of the limit cycle, unless the describing function crosses the harmonic locus again at A. If the servo force F is too low, the describing function will lie entirely to the left of the harmonic locus and stability will be impossible. If it is too high, point B will lie between point A and the origin. The system will then be unstable, causing the response to move to point B, and infinite amplitude. Schilovski actually implemented control in this manner using a clutch mechanism to extract the actuation torque from the gyro itself, obviating the need for a separate servo. 309 Cybernetics for Systems Engineers – DRAFT F 18.8 More General Servo Considerations Most means of applying actuator displacements are rate, as well as position, limited. This introduces a phase lag on the input, which is likely to have a de-stabilising effect on the closed loop. Since such constraints are inevitable, it may be more sensible to operate the servo either at its maximum rate or not at all, rather than reducing system bandwidth just to keep the servo within its linear operating range. The describing function for a servo of this type is found from the output in response to a sine wave input signal. Figure 18-4 : Servo with Rate and Position Limits We see immediately that there is a 90° phase shift between input and output, which should be expected from a component containing an integrator. If the input amplitude is greater than the dead band (denoted ±v/2), the velocity output is zero, which implies the servo stays where it is. There is no reason to believe that this will be the zero position. The description of a servo with both rate limits and position saturation as two consecutive non-linearities is incomplete, as it is evident that there is no ‘open loop’ control over the servo position. In assuming the equilibrium position is about the zero deflection, we fail to take account of the fact that we are usually dealing with linearisations about an operating point. If the servo represents, for example, the elevator actuation of an aircraft, a finite deflection will normally be required to trim 310 Cybernetics for Systems Engineers – DRAFT F the aircraft at its steady state angle of attack. Similarly, the presence of lateral wind forces, or accelerometer bias, in the Schilovski loop become manifest as non-zero gimbal deflections at equilibrium (inertial side forces do not introduce any bias). In order to apply describing functions, we implicitly assume the zero level is an operating point which will change slowly compared with the response time of the control loop. In the aircraft example, the steady load is taken out with a trim tab, in the monorail the centre of gravity may be shifted laterally. By implication, the controller containing the non-linearities is part of a hierarchy of control systems. The presence of the integrator implies the servo must itself contain a feedback loop comparing the actual deflection with the desired deflection. Figure 18-5: Non-Linear Servo Schematic These points are illustrated by the schematic loop of Fig Figure 18-5. The presence of the integrator would render any attempt at open loop control precarious, to say the least, so the actuator is controlled by the difference between the desired deflection and the current deflection. The velocity limit usually arises from a flow rate or current limit in the actuator motor, which is represented here as a limit on the input. The continuous dynamics are second order, so the harmonic locus will not intersect the negative real axis. The gain margin is therefore infinite. Since we are assumed to operate well away from the deflection limits, stability is determined by the relay with dead space non-linearity. The phase margin is found from the intersection with the harmonic locus at a distance given by point B on the describing function plot. If this is less than about 30°, there is a risk of a limit cycle. A system which oscillates about the equilibrium state is described as ‘hunting’. The phase margin is dictated by the maximum servo rate and the time constant of the motor. 311 Cybernetics for Systems Engineers – DRAFT F In order to use the servo within a loop it is necessary to represent it as a linear system, for inclusion in the parent subsystem. We note from the describing function that for signals significantly greater than the dead band, the relay acts as a gain of 2F/π. Using this approximation, the closed loop transfer function is: 𝐺(𝑠) = 2𝐹1 𝜋𝜏 1 2𝐹 𝑠2 + 𝑠+ 𝜏 𝜋𝜏 The undamped natural frequency is: 2𝐹 𝜔02 = 𝜋𝜏 Note that the damping depends entirely on the motor response, which therefore dictates the maximum servo rate and hence bandwidth. These are hardware limitations, which introduce penalties in terms of cost, size, weight and power consumption. These latter considerations impose restrictions on the servo bandwidth, but at the same time, the servo bandwidth imposes significant limitations on overall system bandwidth and proper function. The servo design is therefore a compromise of maintaining adequate system response time, whilst avoiding excessive weight, etc.. Note these latter constraints are secondary to achieving adequate response time, but the modern tendency is to employ a tail-wag-dog approach of dictating subsystem performance on the basis of secondary system requirements, thus rendering primary function difficult, if not impossible, to achieve. 18.9 The Phase Plane 18.9.1 Dominant Modes There is no shortage of authorities which recommend that the designer should aim to ensure the system modes are closely spaced by, for example, employing a Butterworth pole pattern. This author has consistently insisted that the mode frequencies should be determined from their proper function. This understanding comes from experience with the specific plant to be controlled. In general, we would no more set the modes to the same frequency than we would tune the strings of a violin to the same note. If you lack this understanding of the original open loop plant, you are in no position to consider modifying it. We use the differences in frequency to decouple different behaviours. This enables us, for example, to control the attitude of an aircraft, as well as its trajectory, from a single set of controls. Imagine trying to ride a bicycle in which the balancing modes and steering modes were strongly coupled together. If we cannot find an approximate factorisation of the characteristic equation, the implication is that modes are closely coupled, which indicates there is something wrong with the plant itself. With this decoupling, we can study modes in isolation. When we try to speed up our system response, or improve its behaviour in some other way, the result is that coupling between modes and cross-coupling between functionally independent 312 Cybernetics for Systems Engineers – DRAFT F channels increases to such an extent that this assumption breaks down. Our fundamental concern is with how much feedback we can have before the effects of all our approximations and uncertainty in the plant dynamics undermines the assumptions we have made. In short, the most important single concern when introducing active feedback control is with the robustness of our design. At present we use the idea of gain and phase margins to hopefully account for such uncertainty, together with separations in characteristic frequencies of modes of at least a factor of 3, (preferably 10). With the proviso that we shall investigate robustness explicitly, it is often the case that we can deal with modes in isolation. A mode is represented as either a simple or a quadratic factor, and familiarity with the specific plant behaviour will usually indicate which modes are of immediate interest. We have already done this in the case of the rocket weathercock mode, in assuming it to be separable from the trajectory mode (the phugoid). As a further example, consider the gyro monorail; our concern is with the stability of the precession mode, which is dynamically unstable. We can write the equations of motion, ignoring the terms which are significant at high frequency, as: Gimbal: −𝐻𝜙̇ = 𝑀𝜃 𝜃 + 𝑀𝑐 Roll motion: 𝐻𝜃̇ = 𝑊ℎ𝜙 Experience with a toy gyro will soon show that the most obvious effect is the tendency of the spin axis to align itself with the precession axis. The effects of the expected inertia resistance (proportional to angular acceleration) are transient, and not noticeable. This counter-intuitive behaviour unfortunately makes the gyroscope an ideal candidate for the active element in the crackpot inventor’s proposals. This is now a second order system. 18.9.2 Representation of a Second Order Plant The importance of a second order system is its motion can be represented completely on a twodimensional diagram. We have seen from the works of M C Esher, that pictorial representation of higher dimensions, although possibly aesthetically pleasing, is ambiguous , and the reader is warned against such self-deception. Beyond two dimensions; algebra reigns supreme. Consider the general linear, constant coefficient second order system: 𝑣̇ = −2𝜁𝜔0 𝑣 − 𝜔02 𝑦 + 𝑓 𝑦̇ = 𝑣 We eliminate time by expressing it in terms of y: 313 Cybernetics for Systems Engineers – DRAFT F 𝑣̇ = 𝑑𝑣 𝑑𝑡 = 𝑑𝑦 𝑑𝑣 𝑑𝑡 𝑑𝑦 =𝑣 𝑑𝑣 𝑑𝑦 = 1 𝑑𝑣 2 2 𝑑𝑦 The second order plant is described by the single equation: 1 𝑑𝑣 2 2 𝑑𝑦 = −2𝜁𝜔0 𝑣 − 𝜔02 𝑦 + 𝑓(𝑣, 𝑦) 18.9.3 Unforced Plant A diagram is only of any value if we know how to interpret it. The most important features become apparent if we consider the simplest case of a system with no input and zero damping. Figure 18-6 : Phase Plane - Linear Oscillator - Locus Direction is Clockwise The equation is usually plotted with displacement as ordinate (horizontal axis) and its rate of change as abscissa. With no damping, the system conducts a simple harmonic motion so that the displacement and speed are in quadrature (displacement lagging speed by 90° phase lag). This results in the elliptical contours shown, produced by starting the system at y=1, v=0 and y=2,v=0. The most important feature of the phase plane is that all possible trajectories form clockwise paths around the origin as time increases. In quadrant A, the velocity is positive, so it is impossible for the displacement to reduce in this region, the slope can never become negative here. In quadrant B the speed is negative, so the slopes must be negative here. By similar reasoning, the slope must be positive in quadrant C and negative in quadrant D. 314 Cybernetics for Systems Engineers – DRAFT F We know that adding damping will cause the amplitude to decrease with time. It could be argued that in many circumstances, drag is more nearly proportional to the square of the speed, so the plant may be better represented with an equation of the form: 𝑑𝑣 2 𝑑𝑦 = −𝑐|𝑣 2 | − 𝜔02 𝑦 + 𝑓(𝑣, 𝑦) Figure 18-7 : Non-Linear Damping As expected, adding the damping causes the plot to spiral in towards the origin. Using the square law, the damping at low amplitudes is poor, so it takes a large number of cycles for the disturbance to decay. This is worth bearing in mind when we approximate a square law damping term with a linear (viscous) damping representation. Introducing negative damping, as expected, causes the plot to spiral away from the origin. Phase plane representations are used extensively in the analysis of non-linear systems, which have several unstable and stable equilibrium positions, rather than just the origin. These equilibrium points are called ‘attractors’ because of the tendency to spiral in towards them, or spiral out from them. With a second order system, such as a linear mode studied in isolation, it is only possible to have the three types of attractor presented here. With a highly non-linear model, the perturbation equations become problematical, as it is extremely difficult to separate ‘states’ from ‘coefficients’. In effect the dominant mode changes from one centred on one attractor to one centred on a different attractor, but is not stable about either. The result is a trajectory which orbits both with an apparently random switching between the two attractors. Although the trajectories are practically unpredictable, they stay in the vicinity of the 315 Cybernetics for Systems Engineers – DRAFT F attractors, exhibiting non-periodic orbits about both. This ‘multiple attractor’ which can affect continuous time systems of order three or higher, has been given the rather pretentious name of ‘strange’ attractor, and is fundamental to characterising chaotic motion. Figure 18-8: Negative Damping - Locus Spirals Outwards As engineers, we are not concerned with chaos as such, but on determining the circumstances under which it occurs, and making sure our systems remain in an ordered operating regime. Chaos theory comes into its own when studying natural phenomena beyond human control. Our job is to bring nature to heel, as far as is possible. It seems a reasonable speculation that left to themselves, uncontrolled natural selection processes will lead to chaos, because that is the limit of adaptation before actual instability sets in. The reason evolution cannot tell us which species will dominate in a thousand years’ time is the population equations are chaotic in nature. However, to operate an entire economy, on which the survival of mankind depends, in the chaotic region, must be considered the most extreme act of folly of the human race. 18.10 Introducing Control It is evident from the basic stability analysis of the equations of motion that the precession mode of the gyro stabilised monorail is unstable. The phase plane plot for the basic plant resembles the outward spiral of the unstable attractor. Using the parameters derived earlier, with the level of friction we would associate with welllubricated gimbal bearings, results in the above open loop phase plane. The system is clearly unstable, but what is not presented is the time taken for the oscillation to build up from about half a degree to about 1.5 degrees. This is about 50 seconds. The result is important because we need to know how the system will behave in the event of servo failure. Provided the servo does not actually cause the gimbal to lock, there should be ample time to bring the vehicle to a stop. 316 Cybernetics for Systems Engineers – DRAFT F Figure 18-9 : Gyro Monorail with no Feedback Control If we consider the Fourier series fit to the output from a bang-bang actuator, it is evident that there will be significant energy in the higher harmonics, so there is a risk of de-stabilising the nutation, if gimbal rate feedback is employed. In order to avoid this risk, we restrict our consideration to roll angle feedback. The phase plane is just another tool which provides insight into the system behaviour, it is not a panacea solution, nor is it a numerical ‘sausage machine’ to be used by unskilled labour. The designer must be aware of the potential dangers of taking a single mode in isolation. We shall introduce the bang-zero bang control used to illustrate the describing function approach. This is represented on the phase planes as switching curves which indicate the conditions when the servo force is applied. In this case, if the roll angle lies within the deadspace, the gimbal control moment is zero, otherwise it provides a constant control torque in a sense which stabilises the system. 317 Cybernetics for Systems Engineers – DRAFT F Figure 18-10: Effect of Relay with Dead Space Control –Roll Angle Figure 18-11: Effect on Gimbal Angle The roll angle plot indicates that the trajectory will spiral in until it is approximately bounded by the switching curves. Superficially, this appears reasonable. However, a glance at the phase plane plot 318 Cybernetics for Systems Engineers – DRAFT F for the gimbal angle shows that the situation is far from satisfactory. The maximum gimbal deflection approaches 1 radian, which is well beyond the small angle assumptions implicit in the equation of motion. In practice the gimbal would hit the stops and the vehicle would topple. Figure 18-12: Excessive Gimbal Deflection Indicates a Requirement for Larger Gyros The gimbal limits may be respected by increasing the gimbal deflection feedback. However, this requires an increase in gyro angular momentum in order to maintain separation between nutation and precession modes. A system which appears stable according to our basic analysis is easily rendered unstable by system non-linearities. Typically, the high gains and phase advance which appear to ensure rapid system response are precisely the measures which aggravate the effects of non-linearities. Any method of analysis of the robustness of the system must take into account the need to avoid saturation of the system states. 18.11 Hierarchical System Considerations Automatic control is frequently introduced with words to the effect; ‘given a system described by the linear time invariant equation: 𝑥̇ = 𝐴𝑥 + 𝐵𝑢...’ The view of the author is that much ingenuity is involved in deciding what aspects of system dynamics are relevant to the problem in hand, so the equations are rarely ‘given’. Indeed if our sole knowledge consists of the numerical values of the elements of the matrices, it is difficult to see how we could offer advice as to how the plant could be modified to render it more readily controllable. 319 Cybernetics for Systems Engineers – DRAFT F Furthermore, the implicit assumption is that the states are similar in nature at all levels in the system. This is never the case. The higher level functions are different in nature from the lower level functions, but are equally describable as linear differential equations. The skills of the soldier on the field of battle are different in nature to those of the general. What are constants to one are variables to another. However, at the end of the day the evolution of any dynamic process can be modelled as a set of differential or difference equations. We do not throw the baby out with the bathwater, just because snobbery and self-importance does not admit that ‘high level’ system function is just as amenable to dynamic analysis as any other level. To illustrate this point, consider a practical example of a simple hierarchical system. 18.12 Scanning System A scanning system for a certain type of sensor consists of a mirror driven by a pair of servomotors. In order to avoid backlash the mirror is supported on flexible mounts, and the direction of pointing in azimuth and elevation are measured using shaft encoders. The mirror is required to follow a scan pattern consisting of sinusoidal motions in azimuth and elevation. The frequencies of the azimuth and elevation scans are to be controllable, as is the phase of each with respect to a specified time datum, and the amplitude of each. At first sight, this appears a simple matter of feeding in a demand which varies at the correct frequency, amplitude and phase into a pair of tracking loops. However, we know that the output will be corrupted by higher harmonics arising from non-linearities, and even if the tracker were linear, there would inevitably be a difference in amplitude and phase, between the input and output. If the scan pattern were fixed, a simple solution would drive a resonant system open loop. The disadvantage of this approach is the problem inherent in all open loop approaches; it is extremely sensitive to the system parameters. So a relatively inflexible system could be built at great expense, as the scan frequency requirement places severe constraints on the mass of the mirror employed. In order to employ lower quality hardware, and to obtain control over the scan pattern, the loop must be closed, and the inherent robustness of closing the loop would obviate the need for high quality components. The system therefore consists of the plant, consisting of the mirror, its mounts, the motors and the encoders, and two controllers, one to point the mirror, and the other to control the amplitude and phase of the scan pattern. Now amplitude and phase are not describable as linear functions of the bore sight angle, so the controller is non-linear. It introduces the type of non-linearity we would expect in any hierarchical system. The states of the parent system are not linearly related to the states of the subordinate system. Evidently, this is always the case in a hierarchical system. Evidently if the loop which adjusts amplitude and phase has similar response time to the inner pointing loop, the system will be non-linear, possibly even chaotic. We need to know how close in frequency the two may be brought before the system starts to misbehave. 320 Cybernetics for Systems Engineers – DRAFT F An open loop system would be expected to operate near resonance in order to minimise the motor power required. This would also be desirable for a closed loop system. However, we have seen that at resonance the slope of the phase curve against frequency becomes infinite, so we should anticipate that controlling phase is likely to become problematical. Figure 18-13 : Amplitude and Phase Response of Basic Plant A frequency error is the derivative of a phase error, so if we are to control phase we need to operate in a reasonably linear part of the response curve. This corresponds to 90 degrees phase lag, but unfortunately, the amplitude plot is at its peak at this point. It is concluded that control of both amplitude and phase near resonance is unlikely to be possible. 321 Cybernetics for Systems Engineers – DRAFT F The inner control loop is therefore designed to ensure a near linear response in both amplitude and phase. One option is to increase both frequency and damping by means of a phase advance compensator, to ensure the response is at least monotonic, if not reasonably linear. Demand + k 2 1 s 1 0.1s Achieved k1 s 2 2 0 s 02 Compensator Plant Figure 18-14: Inner Control Loop With suitably chosen parameters, the worst effects of operating near resonance can be avoided. The final value theorem states that in order to track an input expressed as a polynomial in time, the loop must contain as many integrators as the order of the input polynomial plus one. This is unfortunate if the task is a sine wave, which is represented as a Taylor series with an infinite number of terms. 322 Cybernetics for Systems Engineers – DRAFT F Figure 18-15 : Gain and Phase Responses with the Inner Loop Controller One approach is to control amplitude and phase explicitly using an outer loop, which generates a sinusoidal input corrected for amplitude and phase lag. The relationship between the states of the outer loop (amplitude and phase) are not linearly related to the inner loop states, so the system cannot be represented as: 𝑥̇ = 𝐴𝑥 + 𝐵𝑢 A moment’s reflection will reveal that this is the usual situation for a hierarchical system. The different levels have different functions. Higher level system functions determine operating points and parameters for subordinate systems. The overall system is far from linear, but that should not worry us if we know the limitations of our linear theory. Indeed, we are far better off than the buffoons who imagine the dynamics of complex systems are adequately represented as finite state machines. 18.12.1 Outer Loop A possible linear model of the outer loop is presented in figure Figure 18-16. The phase and gain responses are their steady state values, so it follows that the characteristic time constants of the outer loop must be chosen to allow time for the transients of the inner loop to die out, this is fundamentally where we get the 3 to 10 times rule of thumb from. Subject to this constraint, we may deduce values for the various feedback gains. However, this conceptual representation of the outer loop bears very little resemblance to what is actually implemented, its purpose is to enable the control loop function to be understood and analysed. The actual implementation requires the amplitude and phase of the output signal to be estimated. 323 Cybernetics for Systems Engineers – DRAFT F Figure 18-16: Outer Loop Controller - Controls Gain and Phase, not the Position and Velocity Given that the output is of the form: 𝑦𝐼 = 𝑎 sin(𝜔𝑡 + 𝜙) We can find the amplitude and phase by constructing a quadrature signal: 1 𝑑𝑦𝐼 𝑑𝑡 𝑦𝑄 = 𝜔 This is produced by passing the output through a differentiator of the form: 𝐺(𝑠) = 𝑠 1+𝜏𝑠 The amplitude of the output is: 𝑎 = √𝑦𝐼2 + 𝑦𝑄2 The phase lag due to the inner loop lags is added as a phase lead to the input signal. The phase lag is given by: sin 𝜙 = sin(𝜔𝑡 + 𝜙) cos 𝜔𝑡 − cos(𝜔𝑡 + 𝜙) sin 𝜔𝑡 cos 𝜙 = cos(𝜔𝑡 + 𝜙) cos 𝜔𝑡 + sin(𝜔𝑡 + 𝜙) cos 𝜔𝑡 Or: 324 Cybernetics for Systems Engineers – DRAFT F 1 sin 𝜙 = (𝑦𝐼 cos 𝜔𝑡 − 𝑦𝑄 sin 𝜔𝑡) 𝑎 1 cos 𝜙 = 𝑎 (𝑦𝑄 cos 𝜔𝑡 + 𝑦𝐼 sin 𝜔𝑡) Where sinωt is the input to the inner loop. It should now be obvious why the closed loop avoids the resonant region, with its potentially infinite gain. 18.13 Concluding Comments If control theory were restricted to linear systems, its value would be severely limited, because of all dynamic systems, very few are actually linear. The issue is not one of controlling non-linear systems but of knowing under what circumstances the linear approaches break down. They are considerably more robust and universally applicable than they are given credit for. By maintaining a high level of abstraction, which only a mathematical approach permits, we can study system behaviour independently of any specific implementation. In stark contrast, most current, non-mathematical approaches, are (in practice) strongly influenced by specific hardware characteristics, and are anything but abstract. 325 Cybernetics for Systems Engineers – DRAFT F 19 The Representation of Noise 19.1 Introduction Noise has been mentioned in earlier chapters as one of the reasons for avoiding very short time constants in compensators. It is hoped the reader has some intuitive idea of what is meant – consider for example the sound of static from a radio receiver which is not tuned to any station. A noise signal is one in which contains no useful information. What constitutes noise depends on the context, but it is evident that we are concerned with values of system states, which are corrupted by random disturbances and instrument errors, these limit the accuracy to which the states may be known. This chapter considers how we represent such noise sources. Whilst saturation yields the upper bound on the magnitude of the signals passing through the system, the separation in signal levels which can be distinguished is determined by the level of noise. Hence the maximum and minimum intervals, which ultimately determine the amount of information which can pass through the system are dictated by saturation and noise. The ratio of the saturation value to the measurement interval is called the system dynamic range. In principle, the noise level would define the quantisation interval to convert analogue signals to digital. It is therefore reasonable to characterise even the original analogue signal as so many bits, corresponding to the logarithm to base 2 of the dynamic range. A moment’s reflection will reveal that this is the same as the information capacity of the process conveying the information. The reason our earlier ruminations, based on response time, are incomplete, is we have only addressed the time response aspect of the system ‘intelligence’, in order to define it completely, we need to know the quantities of information flowing. In the real world this information is represented as positions of controls, angles of attack, gimbal deflection angles, etc., but cybernetics views these as no more than writing of the same information on a different medium with a different type of ink. 19.2 Correlation What is fundamental to any broadening of the application of a mathematical theory of systems is the ability to compare two signals and determine whether they are the same. Ideally, we should also seek some metric of how dissimilar the two signals are. The latter requirement is really a luxury, because as Ross Ashby’s homeostat, and indeed, evolution by natural selection illustrate, the adaptation can be random, but the ability to identify the null state is essential for error driven regulation. To recap fundamental results, error driven regulation is the only way of dealing with an open set of disturbances. How, in general, do we compare two signals? 19.2.1 The Same, Or Different? As an example, an estate agent might list all the attributes of the properties on his books. In order to avoid ambiguity of natural language, each field describing the property would be selected from an 326 Cybernetics for Systems Engineers – DRAFT F appropriate keyword list, which would appear as a pull down menu when the data is entered, thus ensuring exact matches between enquiries and data fields. Data specific to particular properties, such as a photograph, or any additional comments on the property would probably be entered explicitly. We might expect price, location, type of dwelling, number of bedrooms, number of reception rooms, etc. to be specified. In an ideal world, the estate agent would enter the customer’s requirements using the standard keyword list and out would come a list of properties matching the customer’s requirements. In practice, only rarely would anything at all come out. Depending on how many properties are on the books, the estate agent would play with the inputs until about five or six options emerged. The simplest option would be to score a one for each attribute which matched , and zero for each which did not, and summing the result, the five or six with the highest score would be selected. This has the difficulty of weighting less negotiable attributes, such as price, number of bedrooms, and perhaps location equally with cosmetic or irrelevant attributes. Most databases require a certain ingenuity both in entering data and on retrieving it. Does a two bedroom house with planning permission for a loft conversion constitute a 2 and a bit bedroom house? What if a separate dining room could be used as a bedroom? The funny little room currently filled with junk, could that be treated as a potential bedroom? Architect’s aberrations, and weird extensions and alterations quickly defy the naïve options of the input lists. We could extend the database by making the options themselves database fields into which we can enter a number reflecting how accurately the term describes the actual property. In essence, we deal with the vagueness of the real world by replacing the crisp sets of the original database with fuzzy sets. This would yield a, somewhat arbitrary, metric indicating how well each attribute meets the customer’s requirements. This could be extended further by deciding the relative importance of each attribute, by applying a large weighting to the important attributes, and low weighting to unimportant. This results in an overall figure of merit, from which the ‘best’ options may be selected. The problem arises, as in all ‘soft’ analysis methods, of how we attach meaningful weightings to the attributes. Staying with single bit data, consider the representation of letters of the Latin alphabet represented as arrays of dots, as was once actually done on low cost printers. For simplicity, each character will be defined as a 59 dot array. The characters A and B are presented. The object of the exercise is to be able to distinguish between these symbols using an algorithm. We will use the image of one letter, say A, as a template to compare with a candidate, to see if they match. 327 Cybernetics for Systems Engineers – DRAFT F Figure 19-1: Pattern Matching Requires the Noise to be Characterised Following the database matching idea, we increment our metric by 1 when the candidate and template pixels are both high and 0 when they do not. Comparing the A and B, For the top row, 3pixels match, two don’t = 3 For the 2nd, two pixels are high on both=2 For the 3rd, 2 pixels match=2 For the 4th, 4 match, 1 doesn’t=4 For the 5th, 2 match=2 For the 6th, 2 match=2 For the 7th, one matches=1 For the 8th, none are set=0 For the 9th, none are set=0 The total value is 16, whilst for a correct match it would be 18 so the ratio=0.89, could be interpreted as a measure of how much the letter B resembles the letter A in this script. 19.2.2 Pattern Matching As it stands, this is not particularly useful. It could be argued that the legibility of the script could be assessed pseudo-objectively by comparing every character with every other and summing the ambiguity values of all pairs. Meaning can only be ascribed to these numbers, if due to imperfections in the print process pixels are omitted and additional ones included at random. These spurious pixels constitute noise. Knowledge of the process would enable us to assign a probability of spurious pixel mismatch from which the probability of confusion of characters may be determined. Without this knowledge, the metric is meaningless. In order to make the results more general, we treat the pattern we are seeking as a (not necessarily continuous) function of two variables: 328 Cybernetics for Systems Engineers – DRAFT F f f x, y The region being checked for the presence of this pattern is also a function of the position coordinates, but these are referred to the region itself, and not the template, so the coordinates will be denoted h,k rather than x,y. g g h, k We shall assume that the region is to the same scale and the same orientation as the template, so that an exact match of the template with the region is possible. We test an area equal to the template at the coordinate (h,k) by comparing each point Dealing with functions of two variables adds unnecessary complexity to the problem. Imagine the page is scanned as a raster pattern, in much the same way it is read. This is always possible in practice because the scan line width will be equal to the resolution, which will always be finite. In order to convert from a 2D matrix to a 1D signal we need to know the raster width, W. The area of interest is located between the k th scan and the (k+b) scan, where k is the vertical coordinate in the raster, and b is the height of the area of interest. Expressed as a time series, the start of the area of interest is offset from the start of the scan by kW samples, and has duration bW samples. Within this interval, the area of interest is offset by h samples (where h is the x coordinate), and is ‘a’ samples in duration, where a is the width of the symbol, followed by a delay of W samples and another block of samples of length ‘a’, this repeats across the entire range of scans. The mapping from two dimensions to one involves a fairly simple algorithm, and indeed corresponds to how data is usually stored in digital computers. The algorithm may be extended to higher dimensions enabling searches in scale, orientation and other common distortions, to be mapped into a single dimension. The total number of samples, and the magnitudes of individual samples is not affected by this process, so the entropy is unchanged. The same data could be scanned vertically, and there would be no entropy change, so the mapping to a single dimension does not alter the information content, nor should it matter how the data is scanned, provided each sample is covered only once per scan. Using the algorithm, the area of interest is transformed to a 1D sequence covering just the area of interest. So it is now a function of the sample number, which may correspond to time, or some more abstract independent variable. The problem is now in a form amenable to mathematical analysis. The problem of pattern matching, whether printed characters, images of objects or spoken words, reduces to ‘how well does function f(t) fit function g(t), over the range (0,T), where t is an arbitrary independent variable?’. The difference between the two signals is the error signal: t f t g t 329 Cybernetics for Systems Engineers – DRAFT F We have already solved this problem when we considered a least squares fit of a set of elementary functions to a given time series, when seeking to characterize the sine wave signal distorted by rapid non-linearities. In that case the ‘template’ function was the first harmonic of the output signal. If we fit f(t) to a set of elementary functions so that it takes the form: 𝑓(𝑡) = ∑𝑛1=1 𝑐𝑖 𝑓𝑖 (𝑡) For m samples, these quantities become matrices: Fc g Where is now an m-vector, F is an mn matrix, c is an n-vector, and g is an m-vector of samples. We found in Chapter 18 that the coefficients are given by: c FT F 1 FT g In the special case of a single function (template) fit to the data g(t), there is only one function, and hence one coefficient, which we may interpret as the measure of the fit in a least squares sense. In this case: F T F f t j m 2 1 F T g f t j g t j m 1 Hence the metric becomes: f t g t m c j 1 j f t m 2 j 1 In other words, we compare every sample with the template and sum the result, we then scale the result by the same process applied to the template itself. This comparison process is called correlation. We can map n-dimensional data to a single dimension using a raster-scanning approach, because we are always dealing with finite dynamic ranges, and in fact the quantities f and g are not limited to scalar numbers, the approach applies to single bit (text string) data as well as continuous data, multidimensional data and scalars. This metric therefore appears fairly universal, it is used explicitly in experimental curve fitting and implicitly in practically every form of pattern analysis. It is usually given the name correlation coefficient. 330 Cybernetics for Systems Engineers – DRAFT F In analysing correlation problems, it is usually more convenient to deal with continuous expressions, assuming the sampling interval is infinitesimal. The sums reduce to integrals. Furthermore it is usual to define the template such that its self-correlation is unity. The correlation becomes: 1T cT f t g t dt T 0 Returning to the original problem, the objective is not just to see whether a pattern is present in the data set, but whereabouts it is located. However, the setting of thresholds to decide whether the correlation function is sufficiently high, requires knowledge of processes which degrade the pattern, i.e. the noise which may be present. We need to know how likely it is that the threshold crossing is spurious. 19.3 Auto Correlation If we select symbols (e.g. letters) at random and put them in sequence, it is unlikely that they will spell out English words. We expect a constraint to be present which restricts the options which may be selected following a particular letter. In fact we need to take account conditional probabilities of up to about four consecutive characters before the resulting sequence starts to generate credible English words. Conveying meaning appears to require a restriction on the entropy of the message to below the potential capacity of the channel. Conversely, if we compare successive samples of a message, we expect to see significant correlation, if the message conveys meaning. We can use this idea to detect whether a time series contains a useful signal is to investigate it for trends. We have seen from our deliberations with the z transform that functions of time may be represented as recursive algorithms, so that the current output depends in some way on previous outputs. We can detect whether such a relationship exists by comparing the output with itself, shifted in time: 𝑐𝑘 = ∑∞ 𝑖=0 𝑦𝑖 𝑦𝑖+𝑘 The resulting sequence is called an autocorrelation. For a continuous signal, the sequence becomes a function of a time offset: 𝑇 𝑐(𝜏) = ∫0 𝑦(𝑡)𝑦(𝑡 + 𝜏)𝑑𝑡 Where it is understood that T is long enough to cover the maximum system time constant of interest. If we consider the impulse response from a first order lag: 𝑦(𝑡) = 𝑒 −𝑎𝑡 The autocorrelation function becomes: 331 Cybernetics for Systems Engineers – DRAFT F 𝑇 𝑐(𝜏) = ∫0 𝑒 −𝑎𝑡 𝑒 −𝑎(𝑡+𝜏) 𝑑𝑡 = − 𝑒 −𝑎𝜏 [𝑒 −2𝑎𝑡 ]𝑇0 2𝑎 → 𝑒 −𝑎𝜏 2𝑎 There is an observable pattern in the signal. Fig Figure 19-2 is a plot of the outcome of throwing two dice 100 times. The average (7) has been subtracted from each throw in order to produce a nominally zero-mean sequence. 6 4 2 0 0 20 40 60 80 100 120 -2 -4 -6 Figure 19-2: Sequence of Dice Throws 700 600 500 400 300 200 100 0 0 20 40 60 80 -100 -200 Figure 19-3 : Autocorrelation of Random Sequence 332 100 120 Cybernetics for Systems Engineers – DRAFT F The correlation plot for this sequence indicates that the correlation is high only if there is no offset in time. For all finite time offsets the correlation is much lower than the peak. One might expect that the longer the sequence, the lower the correlation will be, away from the zero time offset. With an infinite series with samples separated by infinitesimal intervals, it looks like the autocorrelation for a random sequence will become a delta (impulse) function. We can then define a random process as one whose autocorrelation function reduces to an impulse. 19.4 Characterising Random Processes A random process is characterized as an infinite series of joint probability density functions (pdf), each one of which characterizes the probability of finding the amplitude x of the process at some time. For example, for three discrete time samples we may define a pdf: pdf x1 , x 2 , x3 : t1 , t 2 , t 3 x1 x 2 x3 probabilit yx1 xt1 x1 x1 , x 2 xt 2 x2 x 2 , x3 xt 3 x3 x3 As the population of individual samples is infinite, we must revert to a statistical description of the process, and leave the lies and damn’ lies to the software developers. The process is characterized typically by the first and second moments: Mean: x t x pdf x, t dx Mean Square: x t 2 x 2 pdf x, t dx Variance: 2 t x x t 2 pdf x, t dx Correlation function: t , x x 1 2 pdf x1 , x 2 : t , dx1 dx 2 These may be defined for vector functions, in which case the correlation function becomes a correlation matrix. Note: t , t x 2 t Because x can have only one value at any instant in time, so if t , then x1 x2 . (we have seen this with the dice example) In the following, and indeed implicitly in Monte Carlo simulations, it is assumed that the statistical processes are stationary (the statistics are invariant with respect to translation in time), and ergodic (the statistics taken from a single run over time are the same as the statistics of an ensemble of processes). 333 Cybernetics for Systems Engineers – DRAFT F It is practically impossible to prove that these conditions ever apply, but any error will be present in a Monte Carlo just as in an analytical approach. These assumptions simplify the expressions for the process statistics: x t lim 1 T T xt dt T 2 T 2 lim 2 t t1 T 2 1 T T lim 1 T T xt x t dt 2 T 2 T 2 xt xt t dt 1 T 2 19.4.1 White Noise Taking a simple example, if there is no correlation between one instant and the next: xt xt t1 0, t1 0 n , t1 0 2 Where n is taken from the input random sequence (it is assumed x(t) has zero mean). The individual samples must have finite duration Δt , so that the time sequence of is in fact a set of samples of length: N T t The variance of the sequence is given by: 2 1 N N n 2 1 But the correlation is given by: t1 2 t W A random sequence for which this equation applies is called white noise. In practice white noise is a process which has correlation time much shorter than any of the time constants relevant to the current problem. 334 Cybernetics for Systems Engineers – DRAFT F It is evident that the correlation function of white noise does not have units of the input quantity squared; it is multiplied by a time constant, in this case, Δt . A white noise source is one for which Δt→0, implying the amplitude of individual samples increases without bound. Any practical attempt to measure white noise intensity will be limited by the sample interval that can be used, so a given noise source might be expected to yield different answers for different intervals. However, we ascribe to white noise the property that the variance of the noise reduces proportionately as the sample interval increases, so that the estimate of white noise intensity calculated by correlation ought to be independent of the actual sample length used. The justification for this is the noise ‘energy’ (proportional to the square of the amplitude, and thus the variance), is proportional to the sample length. Provided the sample size is small enough to detect any potential high frequency cut-off within the range of frequencies of interest, the sample time should not matter. For this reason, the arbitrary sample length is absorbed into the term W, which is what the correlation process actually measures, and is independent of the sample length for a white noise process. Like geometric points, particles, weightless pulleys, inextensible strings and rigid bodies, white noise is another useful fiction which is never absolutely true, but yields deep insights into the problem. There is no shame in approximation; it is ignorance of the error bounds which renders results useless. 19.4.2 System Response to White Noise White noise intensity is therefore always expressed as a signal squared over a bandwidth, e.g. volts2 per Herz. It may be thought of as the density the noise spread across the frequency domain, hence the more usual term; power spectral density. The random signal at the output of a first order lag is the limit of the following sum: t it xt nit exp Where τ is the time constant of the lag. n(iΔt) is the value of the random input at time iΔt, and i is an integer. This is assumed uncorrelated from one instant to the next, like successive throws of a dice. Also if (t-iΔt) is negative, it will be taken as zero in this equation. It follows that: t t1 it xt t1 nit exp If the random sequence has zero mean, the only significant contributions over a population of samples will be from terms resulting from the same inputs: t 2t it xt xt t1 n 2 it exp exp 1 335 Cybernetics for Systems Engineers – DRAFT F In this form, the correlation depends on the time of application of each random impulse, as well as the current time, so that it does not take the simple form of the original white noise sequence. The time of application of the impulse relative to current time is important, whereas in the white noise case, the time samples could be taken in any order. We are dealing with the forward solution of a linear homogenous differential equation of constant coefficients, whilst the right hand side of our response equation is expressed in adjoint time. The input is taken from a random sequence, so is not an explicit function of time. The statistics are not changed by taking adjoint time. If we let ta=t-iΔt the impulse response of the first order lag when expressed as an adjoint, is common to all inputs, so we are justified in summing them over adjoint time, because a transfer function with constant coefficients is self-adjoint. The impulse response of the first order lag does not take account the finite width of any real excitation. The input is assumed to be a rectangular pulse of finite width. So in order to evaluate the correlation, it is necessary to calculate the output from a first order lag to a pulse of width Δt: The response during the pulse is: t xt n1 exp So the amplitude at the end of the pulse is: t t xt n1 exp n Since Δt will always be chosen such that it is much less than τ. The finite width pulse does not affect the correlation beyond modifying the input. The correlation therefore becomes: T t 1 exp 2 a T0 t t T 1 exp 2 exp 1 exp 1 dt a 2T The left hand side of equation of this is common to all samples, of which there are N., so the correlation coefficient is: T T t t t W t t1 n 2 exp 1 exp 1 t 1 2T 2 2 As t1→0 this tends to the variance: E x 2 0 Where E() is the expected value. 336 W 2 Cybernetics for Systems Engineers – DRAFT F What we have calculated is the output variance of a system fed by white noise. This time domain calculation is seriously tedious. An alternative approach is considered later in this chapter. 19.4.3 Implications for Time-Varying Systems A practical time-domain model of white noise uses a sample period much shorter than any time constant of the system (which, by implication, assumes the response times of the components have been found long before we think about computer modelling). If this sample period is Δt, a reasonable approximation to the white noise process is a sequence of pulses of rms amplitude: y W t Where W is the white noise intensity. This is not very precise, but since the only use for explicit time series white noise generators is for inherently poor Monte Carlo simulations, the small error introduced by this approximation is unlikely to be noticed. The time domain solution in effect replaces the time domain signal fed into a single system with an ensemble of systems, each fed with an impulse taken from the random distribution. The random process must be ergodic for this to be valid. The correlation integral is evaluated using the adjoint solution. Now, the above reasoning could be applied to a system with time varying coefficients. In order to reduce the problem to a single time variable, the equations must be evaluated in adjoint time. Hence the result is directly applicable to adjoint models, but not to their equivalent forward time models. The output yi2 integrated over adjoint time is the rms value of the corresponding (forward model) input taken over the ensemble of run times. The rms miss caused by this input is: T x T Wi yi dt a 2 2 0 Where Wi is the white noise description of the input noise signal. Thus, by squaring and integrating the adjoint outputs with respect to adjoint time, we obtain the miss distance statistics without resorting to Monte Carlo techniques. Also, unlike a typical Monte Carlo, the partitioning of miss distance between the noise sources is found from a single run, so the relative significance of each of the noise sources is readily apparent. Not only have we saved a thousand runs to cover the time to go ambit, this figure is multiplied by the number of disturbances and the number of Monte Carlo replications. In a single afternoon we can complete an analysis equivalent to several million individual runs. Target manoeuvre may be described as a white noise process, but more typically a target would be expected to conduct a deterministic manoeuvre, but at random time. The adjoint for this is similar to equation, but divides the integral by the time of flight. f x T T T 2 337 2 T y 0 2 fT dt a Cybernetics for Systems Engineers – DRAFT F Also, for aiming error, which is a discrete distribution, rather than a white noise process, the interpretation of the input miss distance as an rms value taken over the ensemble implies the deterministic aiming error should also be interpreted as an rms value, there is no need to modify the existing deterministic description. The signal can usually be interpreted as a deterministic time varying signal with a zero mean random signal superimposed on it, so restricting our consideration to zero mean random processes does not lose generality. 19.5 Frequency Domain Calculating the time domain system response to white noise is very tedious, as is calculating a linear system response to any arbitrary input in the time domain. It was only presented here to illustrate the point that it applies directly to the adjoint model of a time-varying system. We found that the Laplace transform furnished a means of overcoming the difficulty of obtaining time domain solutions by mapping the time series functions into the s operator domain, where convolutions were converted to straightforward multiplications. We need a similar process to deal with noise signals. 19.5.1 The Fourier Transform When considering the distortion of a signal passed through a non-linear element we fitted the distorted signal with a series of complex exponentials, representing sine and cosine series. We found that the coefficient of the ith harmonic is given by: 𝜔 2𝜋 𝑎𝑖 = 2𝜋 ∫0𝜔 𝑦(𝑡)𝑒 −𝑖𝑗𝜔𝑡 𝑑𝑡 Now, consider the case where the fundamental frequency is so low, that the resulting time series signal is indistinguishable from d.c over the time scale of interest: 2𝜋 𝑎(𝑖𝛿𝜔) = 𝛿𝜔 𝛿𝜔 ∫ 𝑦(𝑡)𝑒 −𝑗𝑖𝛿𝜔𝑡 𝑑𝑡 2𝜋 0 This can be thought of as a finely sampled continuous function of frequency: 𝑎(𝑖𝛿𝜔) = 𝛿(𝑖𝛿𝜔)𝑔(𝜔)𝛿𝜔 i.e. the a(iδω) is the area of the infinitesmal strip under the function of frequency. When δω→0, the continous function in the frequency domain becomes: 1 ∞ 𝑔(𝜔) → 2𝜋 ∫0 𝑦(𝑡) 𝑒 −𝑗𝜔𝑡 𝑑𝑡 This is called a Fourier Transform, it resembles a Laplace Transform, with s=jω. Hence the frequency domain description of transfer functions is precisely the same as the Fourier transforms of their impulse responses. This has immediate practical application in finding the frequency response of the system. Subjecting the system to an impulse and then applying a Fourier transform to the resulting time series response, yields the frequency response. The experimenter testing a ship’s roll stabilizer need not become unpopular with the crew, after all. 338 Cybernetics for Systems Engineers – DRAFT F 19.5.2 Application to Noise In the time domain, the use of white noise implies infinitesmal time intervals and infinite amplitudes, so that in practice rather clumsy approximations need to be employed. When considering the Fourier series decomposition of the random signal, these difficulties disappear, and white noise becomes a very handy signal description. This also reflects the way in which noise was originally measured. Rather than deriving the power spectral density by explicitly calculating the correlation coefficient over a set of samples, a narrow notch filter was scanned over the frequency range of interest. As with the time domain calculation, the uniform distribution of the noise energy implied that the intensity (power spectral density) obtained by this method should be insensitive to the notch width. The frequency domain description of white noise is obtained from the Fourier transform of its autocorrelation function: Q T 2 1 exp jt1 xt xt t1 dtdt1 T T 2 The time domain solution transforms this to adjoint time, producing a time dependent term common to all members of the ensemble. We represent the sequence as finite width pulses, rather than impulses: xt xt t1 n, 0 t1 t 0, otherwise 1 Q exp jt1 ntdt1 T Now t1 only adopts positive values. So taken over the ensemble: Q 2 exp jt1 dt1 0 Q 2 1 exp jt 2 t j This is the same as the time domain definition obtained by correlation, so white noise may be described by the same constant whether in the time or frequency domains. The frequency response of a constant coefficient system is obtained from the transfer function expressed as a Laplace transform by the substitution: s j If the system is described by transfer function G(s), the amplitude of the response to a harmonic input at frequency ω is given by: 339 Cybernetics for Systems Engineers – DRAFT F G j1 G j G j 2 The rms value taken over all frequencies is: S t 2 2 t G j WG j d 2 t As Δt→0. The maximum frequency corresponds to the minimum sample period, so that the frequency domain and time domain descriptions are consistent, and the same signal energy is present in both. In the time domain we consider components which change in time in the range: t t s T With white noise we assume the energy is equally partitioned between time ranges, hence the variation of variance with sample interval. The equivalent frequency domain description partitions the same energy equally over the corresponding range of frequencies: 2 2 T t In both calculations we ignore the lower limit altogether, but there is an infinitesimal region about the origin which is usually ignored. What this means in practice is the sample period and observation duration must be such that this region contains insignificant energy in both integrals. Integrating over the frequency range yields the same answer as calculating ρ(0). Too many texts appear to have forgotten why the 2π factor is included in the integral over the frequency domain. So we expect from energy considerations: S 0 E x 2 Equation (73) uses the power spectral density of the white noise directly, so we do not need to mess about with finite width pulses which can only be interpreted as white noise inputs at the final stage of the calculation. Also, this is an inherently easier calculation than a correlation, but is restricted to time-invariant systems. As an example, consider the first order lag: Gs Hence: 340 1 1 , G j 1 s 1 j Cybernetics for Systems Engineers – DRAFT F S W 2 d 1 2 2 W W tan 1 tan 1 0 2 2 The same answer as was derived in the time domain is obtained much more readily. 19.6 Application to a Tracking System A sensor detects a distant target, and contains all the processing needed to follow the target within its field of view. If the target is distant, the amount of energy arriving at the sensor is expected to be small, so some magnification is necessary. Typically, the higher the magnification, the narrow is the possible field of view, so if the target is moving it may be lost from the field of view. The detector, and its associated signal processing, will usually generate noise as well as the desired signal. If the target is a distant astronomical object, the tracking system may use knowledge of the rotation of the Earth to maintain the approximate sight line direction, but usually feedback control will be needed as well. In the military context, where no such cooperation is expected of the target, feedback is all that can be employed. There are two basic design options for the feedback. The boresight error may be measured and fed back, using phase advance compensation to stabilise the loop, or sight line rate as well as boresight error may be estimated and fed back. Figure 19-4: Typical Tracking Loop We shall ignore any a priori knowledge we may have, so that the tracker must work completely by feedback. Also we shall stick to the single input/single output loops that we have studied so far. A possible control loop is presented in fig Figure 19-4. 341 Cybernetics for Systems Engineers – DRAFT F The detection, recognition and target loyalty require processing in the sensor which is expected to take a finite amount of time. The presence of the double integrator arising from the equation of motion implies that some phase advance compensation must be necessary. In addition, we assume the servo force takes a finite amount of time to build up. We require the boresight error to be kept within the field of view when following a rotating sight line. The open loop transfer function is: 𝐾 𝐺(𝑠) = 1+𝑇 𝑝𝑠 1+𝛼𝜏𝑠 1 𝐼 1+𝜏𝑠 1+𝑇𝑎 𝑠 𝑠2 The boresight error is: 𝜀(𝑠) = 𝜃𝐷 − 𝜃 = 𝜃𝐷 − (𝜀(𝑠) + 𝑛)𝐺(𝑠) Where θ is the sight line direction and the subscript D denotes a demanded quantity, n is the noise sequence. It follows that; 𝜀 𝑛 𝐺(𝑠) = 1+𝐺(𝑠) From which the rms error may be calculated: 1 ∞ 𝐺(𝑗𝜔) 𝐺(−𝑗𝜔) ̅̅̅ 𝜀 2 = 2𝜋 ∫0 (1+𝐺(𝑗𝜔)) 𝑊 (1+𝐺(−𝑗𝜔)) 𝑑𝜔 Where W is the power spectral density of the measurement noise. Note, that in order to attenuate the noise G(s)→0. The deterministic tracking error is: 𝜀 𝜃𝐷 1 = 1+𝐺(𝑠) So accurate tracking of the input requires G(s)→∞. There is clear a conflict between attenuating noise and at the same time accurately following a moving sight line. The transfer function: 1 𝑆 = 1+𝐺(𝑠) is known as the sensitivity, and: 𝑇= 𝐺(𝑠) 1+𝐺(𝑠) is the transmission, or complementary sensitivity. These become important when we consider design methods based on closed loop behaviour. 342 Cybernetics for Systems Engineers – DRAFT F 20 Line of Sight Missile Guidance 20.1 Introduction The earliest designs of anti-air target missiles used some form of line of sight guidance. This was a natural development of anti-aircraft artillery, which unlike the homing guidance, covered earlier, does not require an expensive sensor to be expended with the missile. Unlike the homing system, the coefficients of the transfer functions are usually constant, so we may use this example to illustrate the ideas of the previous chapter. We should expect a deterministic miss distance to arise because the line of sight is unlikely to be fixed, or even rotate at constant angular velocity if a target is flying a straight line trajectory at constant speed, offset from the launch point. However, for the moment, we shall restrict ourselves to the statistical miss distance which arises from the noise sources within the loop. 20.2 Line of Sight Variants From the point of view of fundamental kinematics, line of sight guided missiles may be implemented in one of two basic forms, the oldest is called a beam rider, but later systems tended to use a form known as command to line of sight (CLOS). The reduced cost of modern seekers has rendered this technology in either form, more or less obsolete. The pacifist may therefore read this section with a clear conscience knowing that this example is chosen for its efficacy in explaining an academic point, rather than any practical value it might have in the construction of weapons of war. 20-1 Line of Sight Guidance Basic Options Some authorities treat beam riders and CLOS as equivalent but there is an important difference, which ensures that CLOS usually has superior performance to beam riders. In both systems a tracker is located on the launch platform, which usually is either stationary or moving slowly compared with the speed of the target. In a beam rider system, the tracker maintains its bore sight pointing at the target, and a guidance beam is transmitted, having a common bore sight with the tracker. The missile has rearwards facing 343 Cybernetics for Systems Engineers – DRAFT F sensors which serve to detect its position from the bore sight. How this is achieved is the subject of subsequent iterations of the system analysis. In a command to line of sight system the target is again tracked with a sensor on the platform, but the same sensor also tracks the missile. The error between the target and missile positions is used to derive a guidance signal which is transmitted to the missile via a command link. Superficially, these appear identical guidance philosophies, but a moment’s reflection will reveal that the CLOS system does not require the tracker bore sight to be exactly aligned with the line of sight to the target as the sensor can measure the difference in direction from the true sight line to the target and the missile sight line. In short, the tracking loop in the beam rider is connected in series with the guidance loop, whilst with CLOS the two are in parallel. Another advantage of the CLOS loop is the ease with which the acceleration arising from the rotation of the beam may be added into the guidance command as a feed forward (open loop) signal, to improve the tracking behaviour of the guidance loop itself. 20.3 Analysis The two guidance loops are presented in Figure 20-2. In both cases the system has been drawn in a form which represents its proper function, the input is the target behaviour and the output is the miss distance. 20-2 Comparison of Beam Rider and CLOS Loops The principal difference is evident from these diagrams; the target position is fed directly into the guidance stage in the case of CLOS, whilst it is fed through the tracker response in the case of the beam rider. There are several noise sources in the loop, but for simplicity we have restricted our consideration to that originating from the measurement errors. For reasons of clarity, generality and security, the exact forms of the transfer functions G(s) and H(s) are not given. 344 Cybernetics for Systems Engineers – DRAFT F In the form presented in Figure 20-2 the response of the miss distance to target behaviour can be found, but for the current analysis, we shall restrict ourselves to the effect of sensor noise on miss distance. Since the system is linear over the range of state variable magnitudes considered we may repeat the analysis for all the noise sources and find the total effect on miss distance by summing the variances due to each source and taking the square root. For the moment, we shall just consider the noise in measuring the target position. This is associated with two-way path loss which is expected to be much greater than the one-way path loss of the beam rider, and the CLOS missile would usually carry some form of beacon, so that its signal strength is expected to be high. In order to determine the effect on miss distance, we require the transfer functions from the noise source to the miss distance. In the CLOS case, this is fairly simple because the response of the tracker is irrelevant to miss distance, provided the target remains at all times within the field of view of the sensor. The relationship between the target tracking noise and miss distance is determined entirely by the guidance. Denoting the miss distance ‘d’ and the tracker noise ‘n’, it is evident from Fig ??? that: 𝑑 = −𝐻(𝑠)(𝑛 + 𝑑) Or: 𝑑 𝑛 −𝐻(𝑠) = 1+𝐻(𝑠) The variance of the miss due to the tracker noise is found from the frequency domain noise integral: 1 ∞ 𝐻(𝑗𝜔) 𝐻(−𝑗𝜔) 𝜎𝑑 2 = 2𝜋 ∫−∞ (1+𝐻(𝑗𝜔)) (1+𝐻(−𝑗𝜔)) 𝑁 𝑑𝜔 Where N is the power spectral density of the tracker noise, and the symbol σ is used to denote the standard deviation in this context.. For the beam rider the transfer function includes both tracker and guidance response: 𝑑 𝑛 345 𝐺(𝑠) 𝐻(𝑠) = (1+𝐺(𝑠)) (1+𝐻(𝑠)) Cybernetics for Systems Engineers – DRAFT F For the same bandwidth guidance loop the beam rider response will be slower. 20.4 Comments We note that in order to respond to deterministic target behaviour, such as evasive manoeuvres, the system bandwidth must be high. However, increasing the bandwidth admits more sensor noise into the loop causing an increase in RMS miss distance. A compromise must be found which ultimately depends on the expected target behaviour and its signature at the wavelength at which the sensor operates. Needless to say, military targets are highly manoeuvrable and have the smallest signature possible at the wavelength the enemy sensor operates. If the target is un-manoeuvrable, slow and easy to see, the chances are it is a civilian vehicle, and not a legitimate military target at all. Note that we do not describe missile systems according to the sensor type or any other constructional details (e.g. ‘wire-guided’, ‘heat seeking’), instead we start from the definition of the target characteristics, the platform characteristics, environment, available budget and state of the art constraints. Once these are defined it is possible to flow down to an appropriate set of potential design solutions. There is a worrying modern tendency to reverse this process, and try to build a system from pre-defined parts, rather like Lego, in the forlorn hope that the resulting collection of parts will form a system. The astute reader will doubtless notice that the target noise is expected to be a function of the range because the energy received from the target is expected to diminish with increasing range. For this reason the assumption of constant power spectral density throughout an engagement is a bit fishy. The reality is, our noise integral is not an estimate of the variation of RMS miss over a particular fly out. Instead, it is the miss distance of an ensemble of fly outs with intercept occurring at a particular range. In fact, the result is equivalent to an adjoint run, but with a system which is self-adjoint. The analysis would be repeated for a range of nominal intercept ranges. The limitation to small perturbations is not really a limitation, if it is understood that a system in which the miss distance were a large quantity would not merit extensive further analysis, but would be immediately dismissed as useless. 346 Cybernetics for Systems Engineers – DRAFT F 21 Multiple Outputs 21.1 Introduction The compensator design methods presented so far have been for the classical case where the system proper function is to control a particular state, so we design the system to drive the error in the state of interest to zero. This is a single-input/single output (SISO) system. We should expect the plant to be designed such that functionally independent channels will be decoupled from each other, so more complex systems are expected to consist of a collection of SISO systems in close formation. For the above reasons, SISO systems are by far the commonest, and since sensors are expensive compared with processing, it is not immediately obvious why we should spend money on additional sensors, unless there were clear benefits in doing so. Adding sensors is expected to reduce the amount of phase advance which is typically required to stabilise a SISO system, and usually provides a means of increasing the closed loop bandwidth without running into noise and saturation problems. Actually, most of the examples considered in earlier chapters do contain implicit multiple feedback loops, but they are inherent in the dynamics of the plant. The unstable gimbal mount of the gyro monorail is a gimbal feedback implemented mechanically. The weathercock stiffness and damping of a rocket arises from its configuration and its interaction with the airflow. The way we treat a system with a single input, but multiple outputs is much the same as we accommodate the inherent feedback of the plant. As there is a single control input, there is a corresponding reference input consisting of the required value for the output. This means that once we have designed a compensator for the multi-output case, we can assess its robustness by generating Bode plots, harmonic loci or root loci for all loops. Note that this is only possible because we have a phase reference in the form of a single reference input. In the multi-input multi-output (MIMO) case (i.e. a collection of cross-coupled SISO systems) no such phase reference exists, and the idea of a ‘phase margin’ becomes meaningless. The problem arises that the dynamics of these additional artificial loops are not ‘given’, but a matter of design. The following sections suggest possible approaches. 21.2 Nested Loops The simplest (and quite common) case to consider is when the additional measurements control a loop which is nested completely within the main loop. This is illustrated by the sensor tracking loop which has some means of measuring the slew rate of the sight line (e.g. a rate gyro mounted on the antenna). 347 Cybernetics for Systems Engineers – DRAFT F Figure 21-1: System with Nested Loops Evidently, the innermost loop is designed first, and then the outer loop. This works because there is a direct causal relationship between velocity and position. Rate feedback Acceleration demand G1 k as b s 2 n n s n 2 r 2 Achieved acceleration s 2 s 2 s 2 2 n n s n Figure 21-2 : More Typically, the Loops are not Nested – the Outputs are Caused by a Common Input The idea can be extended to more general systems. Taking the rocket autopilot example, we saw that the natural airframe response is usually lightly damped, so the first vice we will want to eliminate is the sustained weathercock oscillation. We can do this by applying body rate feedback. Now it is evident that the body rate loop is not nested within the lateral acceleration loop, both are caused by the control deflection, and consequently the two are in parallel. 348 Cybernetics for Systems Engineers – DRAFT F By re-arranging the transfer functions we can get an apparent relationship between the lateral acceleration and the body rate. The transfer functions relating lateral acceleration to control defection, and yaw rate to control deflection are (from chapter ???) 𝑓𝑦 𝜁 𝑟 𝜁 𝑌𝜁 = = 𝑌 𝑁𝜁 𝑌𝜁 𝑁𝑣 𝑠−( 𝑣 − )𝑈 𝑚 𝐶 𝑚 𝐶 𝑚 𝐶 𝑌𝑣 𝑁𝑟 𝑁𝑣 𝑈 𝑌𝑣 𝑁𝑟 2 𝑠 −( + )𝑠+( + ) 𝑚 𝐶 𝐶 𝑚 𝐶 𝑚 𝑠2 − 𝑌𝜁 𝑁𝑟 𝑁𝜁 𝑌 𝑁𝜁 𝑌𝜁 𝑁𝑣 𝑠−( 𝑣 − ) 𝐶 𝑚 𝐶 𝑚 𝐶 𝑌 𝑁 𝑁 𝑈 𝑌 𝑁 𝑠2 −( 𝑣 + 𝑟 )𝑠+( 𝑣 + 𝑣 𝑟 ) 𝑚 𝐶 𝐶 𝑚 𝐶 These characterise the physics of the problem. We can contrive a relationship between the body rate and the lateral acceleration by dividing one transfer function by the other. 𝑓𝑦 𝑟 𝑌𝜁 = 𝑚 𝑌𝜁 𝑁 𝑌 𝑁𝜁 𝑌𝜁 𝑁𝑣 𝑟 𝑠−( 𝑣 − )𝑈 𝑚 𝐶 𝑚 𝐶 𝑚 𝐶 𝑁𝜁 𝑌𝑣 𝑁𝜁 𝑌𝜁 𝑁𝑣 𝑠2 − 𝐶 𝑠−( 𝑚 𝐶 − 𝑚 𝐶 ) Using this relationship, we can draw the loop as if the body rate loop were nested within the lateral acceleration control loop: k a as b s 2 2 11 s 1 Closed rate loop s 2 s 2 as b ‘constraint’ transfer function Figure 21-3 : They may be Treated as Nested by Introducing a the Ratio of the Numerators as an Additional Transfer Function Although the transfer function looks odd (the order of the numerator is greater than that of the denominator), it does yield useful insights into the plant dynamics. The constraint transfer function consists entirely of plant zeros, which are inherent in the plant dynamics. Unlike the plant poles, these cannot be moved about by feedback. Left half plane zeros may be cancelled by compensation, but why bother, when it is the right half plane zeros that are the source of so many woes? If we want high manoeuvrability from the missile, it will be designed to be statically neutrally stable (i.e. Nv≈0), also the direct lift due to the fin deflection (Yζ) is expected to be small, as is the yaw damping (Nr). The constraint equation becomes: 349 Cybernetics for Systems Engineers – DRAFT F 𝑓𝑦 𝑟 𝑌 𝑁𝜁 −( 𝑣 )𝑈 = 𝑁𝜁 𝑚 𝐶 𝑌 𝑁𝜁 𝑠−( 𝑣 ) 𝐶 𝑚 𝐶 = 𝑈 𝑚 −( )𝑠+1 𝑌𝑣 The steady state centrifugal acceleration is Ur. What this equation means is that the achieved lateral acceleration lags behind the steady state value implicit in the body rate. The time constant of the lag is –(m/Yv), which is consequently known as the ‘incidence lag’. In fact the numerator for the body rate transfer function cancels with the denominator of the constraint equation, so we design the lateral acceleration loop with the plant formed by closing the loop around the body rate. Increasing the number of feedback states beyond two leads to issues of how the loops are to be nested. 21.3 Pole Placement The state space representation of a system with input is: 𝑥̇ = 𝐴𝑥 + 𝐵𝑢 Where x is the state variable and u is the control input. For a single input system B will consist of a single column. The control will have least effect on the closed loop system if it affects only a single state directly, i.e. if it has only a single non-zero element. With a control matrix (B) in this form, any feedback can only affect a single row of the system matrix. If we modify a single row of the system matrix, then, in order to change all n coefficients of the characteristic equation, we must modify all n row elements. As each column corresponds to a state variable, it follows that if we want to place the closed loop poles anywhere in the Argand diagram, we must feedback all n state variables. However, whilst it is obvious that we need the n feedback gains to modify n coefficients, it isn’t clear under what circumstances full state feedback will actually present a solution. To illustrate this point, consider a system governed by the equations: 𝑥̇ 1 = 𝑎11 𝑥1 𝑥̇ 2 = 𝑎21 𝑥1 + 𝑎22 𝑥2 + 𝑢 Evidently, no matter what we do with the control u, we will not affect x1. If the system described by the first of the above equation is stable, this may not matter; we simply do not have complete control over the closed loop poles. The example is easy to identify as a pair of disparate systems, with one exciting the other. More often than not the coupling of the control into the system is not so obvious, and the presence of states which are not modifiable by the available control cannot be determined by inspection. This ability for the control to influence every state is, somewhat mischievously, called ‘controllability’, even though plant which cannot meet this criterion may indeed be controllable in the more usual sense. 350 Cybernetics for Systems Engineers – DRAFT F We shall call this type of controllability, ‘state controllability’, to indicate that it is no more than an essential pre-requisite for pole-placement methods. The following section presents the constraints on the system and control matrices required for this type of controllability. 21.4 Algebraic Controllability If we consider the special case of a system represented as an nth order linear differential equation, it is represented in state space with a system matrix of the form: −𝑐𝑛−1 𝐴=( 1 ⋮ 0 −𝑐𝑛−2 ⋯ −𝑐1 −𝑐0 0 ⋯ 0 0 ) ⋮ ⋱ ⋮ ⋮ 0 ⋯ 1 0 Where the top row consists of the coefficients of the characteristic equation. The remaining rows consist of the (n-1)×(n-1) identity matrix, followed by an (n-1) element column vector of zeros. This is known as a companion form, for doubtless a good reason, which isn’t too clear from this context. If the control matrix takes the form: 𝑏 𝐵 = (0 ) ⋮ 0 It is evident that full state feedback of the form: 𝑢 = ∑𝑛𝑖=1 𝑘𝑖 𝑥𝑖 = 𝐾 𝑇 𝑥 Where the ki are gains, will modify each of the elements of the first row: −𝑐𝑛−𝑖 ′ = −𝑐𝑛−𝑖 + 𝑏𝑘𝑖 Where the prime denotes the coefficient of the closed loop characteristic equation. In this special case, it is evident that each coefficient may be assigned arbitrary values, so that the poles may be placed anywhere we choose. More generally, we need to find out whether a control exists which can move the system states from any value to any other value. The governing equation is: 𝑥̇ = 𝐴𝑥 + 𝐵𝑢 Differentiating: 𝑥̈ = 𝐴𝑥̇ + 𝐵𝑢̇ = 𝐴(𝐴𝑥 + 𝐵𝑢) + 𝐵𝑢̇ = 𝐴2 𝑥 + [𝐴𝐵 Repeating this process n times: 351 𝑢̇ 𝐵] ( ) 𝑢 Cybernetics for Systems Engineers – DRAFT F 𝑑𝑛𝑥 𝑑𝑡 𝑛 = 𝐴𝑛 𝑥 + [𝐴𝑛−1 𝐵 ⋯ 𝑑 𝑛−1 𝑢 𝑑𝑡 𝑛−1 𝐵] ( ⋮ ) 𝑢 In order to proceed we need to use a result known as the Cayley-Hamilton theorem. We know that the eigenvalue matrix must satisfy the characteristic equation: Λ𝑛 + 𝑐𝑛−1 Λ𝑛−1 ⋯ 𝐼𝑐0 = 0 The eigenvalue matrix is obtained from the system matrix by means of a similarity transform: Λ = 𝑇 −1 𝐴𝑇 Multiplying both sides by the eigenvalue matrix: Λ2 = (𝑇 −1 𝐴𝑇)𝑇 −1 𝐴𝑇 (Note: this assumes there are no repeated eigenvalues) It follows that: Λ2 = 𝑇 −1 𝐴2 𝑇, ⇒ Λ𝑛 = 𝑇 −1 𝐴𝑛 𝑇 The characteristic equation is: 𝑇 −1 (𝐴𝑛 + 𝑐𝑛−1 𝐴𝑛−1 ⋯ 𝑐0 𝐼)𝑇 = 0 Since T is non-singular, it follows that the matrix satisfies its own characteristic equation. It also follows from this result, that: 𝐴𝑛 = −𝑐𝑛−1 𝐴𝑛−1 − 𝑐𝑛−2 𝐴𝑛−2 ⋯ 𝑐0 𝐼 Differentiating the plant equation for the n+1 th time: 𝑑 𝑛+1 𝑥 𝑑𝑡 𝑛+1 = (−𝑐𝑛−1 𝐴𝑛 − 𝑐𝑛−2 𝐴𝑛−1 ⋯ 𝑐0 𝐴)𝑥 + [𝐴𝑛−1 𝐵 ⋯ 𝐵] ( 𝑑 𝑛−1 𝑢 𝑑𝑡 𝑛−1 − 𝑐𝑛−1 𝑢 ) ⋮ 𝑢 − 𝑐0 𝑢 We can repeat this process an infinite number of times, substituting for An as appropriate. We can represent x as an infinite time series by repeatedly integrating the highest order derivative. However, it is impossible to find the control input corresponding to an arbitrary state history unless the matrix in square braces has an inverse. The state controllability criterion reduces to the requirement that [An-1B An-2 B…B] has an inverse. In our simple example: 𝐴=( 352 𝑎11 𝑎12 0 0 ), 𝐵 = ( ) 𝑎22 1 Cybernetics for Systems Engineers – DRAFT F 0 𝐵] = (𝑎 [𝐴𝐵 22 0 ) 1 This has a zero row, so the inverse does not exist. 21.4.1 Monorail The state controllability requirement does not necessarily determine whether or not a viable controller can be designed, but can provide insights into the system dynamics. For example, consider the gyro monorail. We wish to know whether it is state controllable, and also why. The plant equation with no feedback, either passive or active, is: 𝐽𝜃̈ + 𝐻𝜙̇ = 𝑀 𝐴𝜙̈ − 𝐻𝜃̇ = 𝑊ℎ𝜙 The system state vector is (𝜃̇ 𝐴= 𝐻 0 0 −𝐽 1 0 0 0 0 0 1 𝐻 𝐴 (0 𝑇 𝜙̇ 𝜃 𝜙) , for which the system matrix is: 0 1 𝐽 0 , 𝐵= 𝑊ℎ 𝐴 0 ) 0 0 ( 0) The controllability test matrix is: 𝐻2 0 − Γ= 𝐻2 𝐴𝐽2 𝐻3 − 𝐴2 𝐽2 + ( 0 − 𝐴𝐽2 0 1 𝐽 0 1 𝐽 0 𝐻 𝐴𝐽 0 0 0) 𝐻𝑊ℎ 𝐴2 𝐽 0 𝐻 𝐴𝐽 The determinant evaluates to: |Γ| = 𝐻 2 𝑊ℎ 𝐴3 𝐽2 It is actually the toppling moment which renders the system controllable. Physically, this means that the equilibrium position is determined by the balancing of the moment due to weight against the disturbance, with the inertial (gyroscopic) torques providing a zero mean correction to perturbations around the equilibrium. 21.4.2 Road Vehicle A fairly crude representation of the directional stability of a road vehicle was derived in Chapter ???: 353 Cybernetics for Systems Engineers – DRAFT F a2 b2 CU k k a b v CU 2 k a b v k U v mU mU Where ψ is the body axis direction and v is the slip velocity. U is the vehicle speed, a the distance of the front axle ahead of the centre of gravity and b the distance of the rear axle behind the cg. C is the moment of inertia in the yaw plane, m the mass of the vehicle and k is the slip force coefficient. The vehicle is steered by deflecting the front wheels with respect to the body. This generates a yawing moment; kaζ, where ζ is the steering deflection, in addition there will be a side force; kζ. The equation of motion, with the steering control input becomes: a2 b2 a b ka k k v C CU CU k 2 k a b v k U v m mU mU If the steady state slip at the centre of gravity where zero, the body would point in the direction of motion. Re-arranging the plant equations with 𝑣 = 𝜓̈ = 𝑣̇ = 0: 𝑎 2 +𝑏2 ) 𝜓̇ 𝐶𝑈 +𝐶𝜁 = 0 𝑘(𝑎−𝑏) + 𝑚𝑈 𝑈) 𝜓̇ + 𝑚 𝜁 = 0 ( ( 𝑎 𝑘 Since neither the steering deflection, nor the yaw rate, will in general be zero, so this is only possible if the matrix of coefficients is singular: 𝑎 2 +𝑏2 𝐶𝑈 | 𝑘(𝑎−𝑏) ( 𝑚𝑈 + 𝑈) 𝑎 𝐶 | 𝑘 𝑚 =0 This corresponds to a speed: 𝑏 𝑘 𝑈 2 = (𝑎 + 𝑏) (𝑎) (𝑚) At this speed the vehicle neither oversteers nor understeers, but from the point of view of handling the steering is perfect. It is usually preferable for the car to oversteer than understeer, so this speed ought to be low. This implies the centre of gravity should be nearer the rear axle than the front, which conflicts with the basic directional stability requirement. The controllability test matrix determinant is: 354 Cybernetics for Systems Engineers – DRAFT F 𝑘 𝑚 𝑏 𝑎 |Γ| = 𝑈 2 − ( ) (𝑎 + 𝑏) ( − 𝐶 ) 𝑚𝑎 2 At this speed using feedback gives us no control over the frequency of the associated mode, but we can adjust the damping. This is hardly a cause of concern. Cars have been built for years oblivious of the fact that there is a speed at which they are apparently ‘uncontrollable’. We obtain a further interesting result if we replace the slip velocity with the actual velocity direction: 𝑣 𝛾 =𝜓+𝑈 (𝑎−𝑏) 2 𝑘 𝛾̇ = 𝜓̇ − 𝑘 (𝑚𝑈2 ) (𝛾 − 𝜓)𝑈 − (𝑘 𝑚𝑈2 + 1) 𝜓̇ + 𝑚𝑈 𝜁 Or: 2𝑘 𝑘(𝑎−𝑏) ) 𝜓̇ 𝑚𝑈 2 𝛾̇ = − (𝑚𝑈) 𝛾 − ( 2 2𝑘 𝑘 + (𝑚𝑈) 𝜓 + 𝑚𝑈 𝜁 2 𝑎−𝑏 𝑎 +𝑏 𝑎−𝑏 𝑘𝑎 𝜓̈ = −𝑘 ( 𝐶 ) 𝛾 − 𝑘 ( 𝐶𝑈 ) 𝜓̇ + 𝑘 ( 𝐶 ) 𝜓 + 𝐶 𝜁 The system matrix then becomes: 2𝑘 − 𝑚𝑈 𝐴=( 𝑎−𝑏 ) 𝐶 −𝑘 ( 0 𝑘(𝑎−𝑏) 𝑚𝑈 2 𝑎 2 +𝑏2 −𝑘 ( 𝐶𝑈 ) 1 2𝑘 𝑚𝑈 𝑎−𝑏 ) 𝑘( 𝐶 ) 0 The determinant of this is zero, which implies that the equilibrium yaw orientation and heading are not calculable from the control input. This is also the constant term in the characteristic equation, so the system has a pole at the origin. This is consistent with experience; we turn the wheel to change direction, and centre it when the new heading is reached; this neutral mode is not a cause for concern. When we find neutrally stable modes, we need to interpret them in terms of physical reality before we can decide whether we should worry about them. 21.5 Comment on Pole-placement For low order systems, we can evaluate the closed loop characteristic equation directly. For higher order systems there exist standard numerical algorithms (e.g. the Bass-Gura algorithm) to calculate the gains needed to place the closed loop poles in their specified positions. This raises the question of where the closed loop poles should be placed. We cannot answer this question without knowledge of the proper function of the plant. Usually we seek to separate functionally independent modes. Imagine trying to use a mouse whose interval between the key presses of a double-click was comparable to the time taken to move it on to an icon. If, for some reason, such understanding of the plant dynamics is unavailable, a Butterworth pole pattern might be considered. Since we expect our states to be limited by saturation in practice, we might determine our gains using a method which tries to keep the states within their amplitude bounds. This is the idea behind linear/quadratic, so-called ‘optimal’ control. 355 Cybernetics for Systems Engineers – DRAFT F Full state feedback appears to allow us to place the closed loop poles anywhere we like on the Argand diagram. This seems too good to be true, and of course, it is. We have the same constraints on pole locations as we did with the root locus method. There is nothing in the method which reveals any requirement for integral feedback needed to track time varying inputs. Integrators must be anticipated as necessary and included in the plant equations before pole placement can be applied. We can put the poles where we like, but what about the zeros? In general, we are interested in a outputs from the plant which are linear combinations of the states. For this reason we distinguish the states from the outputs: 𝑦 = 𝐶𝑥 + 𝐷𝑢 Where y is referred to as the plant output, it consists of the vector of quantities which are available for measurement. In the previous section we have assumed C to be the identity matrix. The matrix D is included to cover cases such as the missile autopilot, which has a direct link from fin deflection to lateral acceleration. The structure of C and D determine the system zeros. Consider a system in companion form: −𝑐𝑛−2 ⋯ −𝑐1 −𝑐0 0 ⋯ 0 0 ) ⋮ ⋱ ⋮ ⋮ 0 ⋯ 1 0 −𝑐𝑛−1 𝐴=( 1 ⋮ 0 (any system matrix can be transformed into this form). The output tracking the reference input is expected to include the nth state variable. If the full state feedback, as is needed for pole placement is applied, C becomes the identity matrix and D=0. The transfer function relating input to output will feed back all derivatives of the nth variable, so that the order of the numerator will be one less than that of the denominator in the closed loop transfer function: 𝑏 𝑠𝑛−1 +⋯ 𝑛−1 +⋯ 𝑛−1 𝑠 𝑛−1 𝐺𝐶 (𝑠) = 𝑠𝑛 +𝑐 At high frequency, this approximates an integrator: 𝑏 𝐺𝐶 (𝑠) → 𝑛−1 𝑠 𝑠→∞ The high frequency roll-off is 20dB per decade, whilst for an nth order system we should expect it to be n times steeper. From a noise perspective, this is not a good feature, because we wish to exclude all the noise above our design bandwidth. This can only be achieved with a sharp cut off. Quite apart from the poor high frequency roll-off, it is usually impossible to measure all the states of a system. The quantities which are available for measurement are the outputs y. These are related to the actual state variables via matrix C, which is determined by the plant dynamics. The elements 356 Cybernetics for Systems Engineers – DRAFT F of C and D determine the system zeros, which are inherent in the plant. Unlike the poles, we cannot place the zeros anywhere we like by applying feedback. Usually, not all the states are available for feedback. However, we found that systems with many states but a single measured output could be made to work, so presumably having more than one output ought to improve matters. We can still apply full state feedback if we can estimate all the states from the outputs. 21.6 The Luenberger Observer Since, in order to design the feedback, we must have a set of differential equations describing the system, it is reasonable to use the same equations to estimate the state of the plant. The astute reader will see the flaw in this reasoning, if we knew the dynamics of the plant with sufficient precision that we can estimate its states from a model of the plant, we would not need closed loop control in the first place. We correct the model estimates by feeding the difference between the measurements and the estimated values of the measurements. Let the observer take the form: 𝑥̂̇ = 𝐸𝑥̂ + 𝐹𝑢 + 𝐾(𝑦 − 𝐶𝑥̂) The circumflex (hat) is used to indicate that it is an estimate of the corresponding state variable (read it as ‘x-hat’). E is an n×n matrix F is n×1 and K is n×m, where m is the number of measurements. The plant equation is: 𝑥̇ = 𝐴𝑥 + 𝐵𝑢 Evidently, E=A and F=B. A system model, fed with the error between measurements and measurement estimates, is called an observer of the system. The closed loop system equation for the observer is: 𝑥̂̇ = (𝐴 − 𝐾𝐶)𝑥̂ + 𝐵𝑢 + 𝐾𝑦 Comparing this with the full state feedback controller: 𝑥̇ = (𝐴 + 𝐵𝐺)𝑥 Where G is the gain matrix. Consider the transpose of the homogenous part of the observer: 𝑥̂̇ 𝑇 = 𝑥 𝑇 (𝐴𝑇 − 𝐶 𝑇 𝐾) This is similar in form to that of the plant with full state feedback. In the latter case we derived a controllability condition on the A and B matrices, this problem is analogous. Following a similar reasoning to the derivation of the controllability criterion, we can obtain an ‘observability’ condition: 357 Cybernetics for Systems Engineers – DRAFT F Γ𝑂 = [𝐶 𝑇 𝐴𝑇 𝐶 𝑇 ⋯ (𝐴𝑛−1 )𝑇 𝐶 𝑇 ] In the case of a single input system this is n×n and essentially determines whether the full state can be estimated from the one input. More generally, it will be n×(nm), so the criterion reduces to the requirement for ΓOT to have at least n independent rows, or in matrix jargon; the observability matrix must have full rank. As with the controllability criterion, failure to meet this criterion doesn’t mean the system is necessarily unsatisfactory – it merely means we can’t construct a full state observer. If the associated mode is well behaved, who cares? Consider the simple case: 𝑢̇ = 0 𝑥̇ = 𝑢 𝑦=𝑢 Or: 0 𝐴=( 1 Hence: 1 Γ𝑂 = ( 0 0 ) , 𝐶 = ( 1 0) 0 0 ) 0 This is singular. It is impossible to estimate position from a measurement of velocity, as should be intuitively obvious. The observer is a feedback system in its own right, using measurements to correct the state estimates rather than to control the states themselves. For this reason the observer and the controller are often called ‘dual’ problems. The gain matrix K, for the observer may be found by pole placement methods, as in the case of the controller. However, we noticed that compensator design required poles well over to the left of the Argand diagram, implying that performance would be limited by excessive noise bandwidth. It is possible to design the observer such that the variance in our state estimate is a minimum, and this approach leads to the Kalman filter. If the observer is designed using pole placement, it is usually called a Luneberger Observer. 21.7 The Separation Theorem A system which uses full state feedback together with an observer to estimate the states is governed by the following equations: 𝑥̇ = 𝐴𝑥 + 𝐵𝐺𝑥̂ 𝑥̂̇ = (𝐴 − 𝐾𝐶 + 𝐵𝐺)𝑥̂ + 𝐾𝐶𝑥 The error in the estimates is: 𝜀 = 𝑥 − 𝑥̂ 358 Cybernetics for Systems Engineers – DRAFT F The equations become: 𝑥̇ = (𝐴 + 𝐵𝐺)𝑥 − 𝐵𝐺𝜀 𝜀̇ = (𝐴 − 𝐾𝐶)𝜀 The poles of the observer equation are not affected by the controller poles, neither are the controller poles influenced by the observer poles. The observer gains may be calculated independently of the controller gains. This result is known as the separation theorem. It permits the observer to be designed independently of the controller. The separation theorem relies on the plant equations used in the observer matching the actual plant equations. Usually, however, the plant equations are not known, or they at least constitute a gross approximation to physical reality. Assume the system, control and output matrices are in error, the observer becomes: 𝑥̂̇ = ([𝐴 + Δ𝐴] + [𝐵 + Δ𝐵]𝐺 − 𝐾[𝐶 + Δ𝐶])𝑥̂ + 𝐾𝐶𝑥 The error equation becomes: 𝜀̇ = ([𝐴 + Δ𝐴] + Δ𝐵𝐺 − 𝐾[𝐶 + Δ𝐶])𝜀 − (Δ𝐴 + Δ𝐵𝐺 − Δ𝐶)𝑥 Where the Δ denotes an uncertainty. We can apply the separation principle in the initial design, but it is evident that some adjustment may be necessary when plant uncertainty is taken into account. 21.8 Reduced Order Observers It seems a bit odd to design an estimator for those states which are actually measured. In cases where the measurement noise is high, this is actually a reasonable approach. However, more typically we are better off using the estimator only for states which are not available for measurement. We re-order the state vector, and by implication, the system and control matrices, so that it is partitioned between the states that are measured and the states that are estimated. There will be a corresponding partitioning of the measurement matrix. 𝐶 = ⌈𝐶1 |0⌉ Where C1 has an inverse. The system and control matrices will be similarly partitioned: 𝐴 |𝐴 𝐴 = [𝐴11 | 𝐴12 ] 21 22 𝐵 𝐵 = [𝐵1 ] 2 The full state observer is: 𝑥̂̇ = 𝐴𝑥̂ + 𝐵𝑢 + 𝐾(𝑦 − 𝐶𝑥̂) 359 Cybernetics for Systems Engineers – DRAFT F 𝑥̂̇2 = 𝐴22 𝑥̂2 + 𝐴21 𝐶1−1 𝑦 + 𝐵2 𝑢 The poles of this observer are the same as the eigenvalues of matrix A22. In general, these will not be favourably located. However, we should have realised by now, that in order to move poles to more favourable positions, we must include zeros. These are implemented by feeding the input both through a dynamic process and directly to the output, and combining the two. In other words we estimate the unmeasured states using a compensator of the form: 𝑥̂2 = 𝐿𝑦 + 𝑧 = 𝐿𝐶1 𝑥1 + 𝑧 𝑧̇ = 𝐹𝑧 + 𝑀𝑦 + 𝑁𝑢 Differentiating the first of this pair: 𝑥̂̇2 = 𝐿𝐶1 (𝐴11 𝐶1−1 𝑦 + 𝐴12 𝑥2 + 𝐵1 𝑢) + 𝐹(𝑥̂2 − 𝐿𝑦) + 𝑀𝑦 + 𝑁𝑢 Or: 𝑥̂̇2 = 𝐹𝑥̂2 + 𝐿𝐶1 𝐴12 𝑥2 + (𝐿𝐶1 𝐴11 𝐶1−1 − 𝐹𝐿 + 𝑀)𝑦 + (𝐿𝐶1 𝐵1 + 𝑁)𝑢 The error in the observer estimates is: 𝜀̇ = 𝑥̇ 2 − 𝑥̂̇2 = 𝐹𝜀 + (𝐴22 − 𝐹 − 𝐿𝐶1 𝐴12 )𝑥2 + ((𝐴21 − 𝐿𝐶1 𝐴11 )𝐶1−1 − 𝐹𝐿 + 𝑀)𝑦 + (𝐵2 − 𝐿𝐶1 𝐵1 − 𝑁)𝑢 The observer poles are the eigenvalues of F. In order to ensure these are distinct from the controller poles, the remaining matrices must be chosen such that: 𝐴22 − 𝐹 − 𝐿𝐶1 𝐴12 = 0 𝑀 = 𝐹𝐿 − (𝐴21 − 𝐿𝐶1 𝐴11 )𝐶1−1 𝑁 = 𝐵2 − 𝐿𝐶1 𝐵1 With these values, the separation theorem applies, with the usual proviso that the separation theorem never really applies, but this approach is not a bad method of deriving a reasonable compensator quickly. The limitations of the separation theorem only represent a problem if we delude ourselves that control system design is a once through ‘sausage machine’ process requiring no broader understanding of the plant behaviour, than is needed to handle-crank algorithms devised by wiser minds. Such an approach betrays spectacular ignorance of the engineering enterprise. 21.9 Concluding Comments On re-introducing the state space representation, with its matrix algebra rather than complex numbers and operator notation, we have also subtly changed the approach from predicting how modifications to the open loop affect the closed loop, to designing the closed loop directly. When we come to consider multi-input/multi-output systems, it becomes evident that extrapolating from SISO methods is not particularly helpful, and the most successful methods tend to work directly with the design requirements of the closed loop plant. 360 Cybernetics for Systems Engineers – DRAFT F If good reason is required to justify the expense of additional sensors for multi-state feedback, a positively watertight case must be presented for introducing the expense, weight, volume and power requirements of additional servos. For this reason MIMO systems tend to be rare; found more frequently within the hallowed cloisters of academia, rather than in the world of practical technology. 361 Cybernetics for Systems Engineers – DRAFT F 22 Respecting System Limits – Optimal Control 22.1 Introduction Introducing artificial feedback, if adjusting system parameters does not yield an adequate solution, has introduced the problem that constraints on compensators beyond ensuring stability, do not emerge directly from the mathematics. In practice, the designer is limited to off-the shelf sensors and servos and is restricted to seeing the best he/she can do with what is available. Once these hardware constraints are introduced, this problem largely disappears. In order to decide which bought-in items are most critical to achieving system performance the loop is designed with incremental hypothetical improvements in component noise, saturation and bandwidth to understand the sensitivity of overall system performance to component characteristics. However, we are not always in the position of seeking the best that can be achieved with ‘given’ components. Sometimes the bandwidth requirement flows down from overall system requirement, in which case, such things as servo requirements are the outputs from the study. As an example, the miss distance of a guided weapon against a specified target depends critically on the guidance loop bandwidth, which flows down from the target signature and behaviour. There is little point considering noise or saturation constraints if the basic kinematics are inadequate. In such circumstances, where bandwidth is driven from fundamental system function, the output of the design process consists of servo and sensor requirements specifications. Now, contrary to a few user requirements I have actually come across, we cannot specify both system performance and subsystem constraints. We can either flow down component requirements from system requirements, or we can see what we can do with what is available. Sometimes one component shortfall can be compensated by another’s exceeding expectations. However, one must apply the methods presented here to decide if such trade-offs are possible. When flowing down requirements, it is quite easy to characterise system components as bandwidths, but as we have seen, saturation and noise prevent high bandwidth from being achieved. In order to flow down to noise levels and saturations, our frequency domain and impulse response methods need to be supplemented with methods based on dynamic range constraints. If both methods yield similar results, we can have reasonable confidence in our design. The methods presented in this chapter complement classical methods, and were once briefly considered the ultimate approach to control system design. If we dispense with the notion that any single method will meet all our needs and make use of all the tools at our disposal, at least we will stand a chance of understanding why our controller doesn’t work, which will enable us to do something about it. 22.2 Optimal Control In every walk in life, there is always some naive zealot who insists on ‘the best’, blissfully unaware of the fact that what constitutes ‘the best’ of anything is very much a matter of debate, it all boils down 362 Cybernetics for Systems Engineers – DRAFT F to what our criteria for quality are. Most of the time this juvenile pre-occupation with superlatives tends to be the domain of the snob, ascribing fantastic value to artefacts of minimal, if not zero utility, beyond appealing to even bigger fools. The ultimate performance of time varying systems is expected to be stability limited, hence the emphasis on stability throughout this book. Optimisation algorithms tend to be notable for their extremely poor numerical stability, and usually need to be used with extreme caution by an individual familiar with both the problem domain, and the details of the optimisation algorithm. All too often they are presented as turnkey applications with an attractive user interface, which allows the naive user to venture freely where angels fear to tread. Some codes are merely stupidity amplifiers. The definition of the ‘best’ regulator in the context of what is known as ‘optimal’ control is one which minimises a quadratic performance index: 𝑇 𝑡 𝐽 = 𝑥(𝑡𝑓 ) 𝑆𝑥(𝑥𝑓 ) + ∫0 𝑓 𝑥 𝑇 𝑄𝑥 + 𝑢𝑇 𝑅𝑢 𝑑𝑡 Where Q is a weighting matrix which can be used to scale the states, or combinations of states, by their saturation values, and R is similarly a weighting matrix for the controls. This does not mean that the method ensures that saturation limits will not be reached, but renders saturation equally likely on all states and controls. Q, R and S are n×n symmetric matrices, where n is the number of states. The objective is to choose a control (u) which minimises the least squares error in the state. This is implicitly a regulator, rather than a tracking problem. The performance index is open to a range of criticisms, which we shall not go into here, except to mention a result which has crept into the folklore of proportional navigation missile guidance. Applying the quadratic performance index to the terminal homing problem yields a navigation constant of 3 as the ‘optimum’. However, the only reason for choosing a quadratic performance index is to make the problem mathematically tractable. If we based the performance index on n rather than 2, the optimal control solution would yield a navigation constant of: 𝑁= 2𝑛−1 𝑛−1 When n=2, this reduces to the much-quoted value of 3, but as n→∞, this reduces to the value 2, which as we have seen, is the limiting value of navigation constant. Using n→∞, is more appropriate for the frequency domain, where we wish to suppress resonance peaks, rather than to average out errors over the entire frequency range. This is the basic idea behind H-infinity optimisation (a name strongly indicative of academic in-breeding), which we shall consider presently. 363 Cybernetics for Systems Engineers – DRAFT F 22.3 Optimisation Background 22.3.1 Principle of Optimality Practically all optimisation problems consist of finding a path in the solution space which minimises or maximises some performance index. Usually we assume only one such path exists, but particularly if there is uncertainty in the parameters, this may not be the case. The principle of optimality states that if an optimal path exists between a certain intermediate point and the end point, the optimal path from any other point must pass through this same intermediate point. This is illustrated with the simple network consisting of nodes which might represent events on a PERT chart, or way points on a routing problem, and connecting arcs, which could represent distances, activity durations, survival probabilities, or the myriad of other data relevant to problems which can be represented as networks. The object of the exercise is to find a path through the network, which maximises or minimises some cost function. Figure 22-1: Illustration of Principle of Optimality A typical minimising problem might be to find the critical path through a complex plan. In our example we need to find the path from A to B which minimises the total cost. The cost of traversing each arc is marked on the network. We could perform an exhaustive search of all possible routes, but the number of nodes does not need become very large before the number of potential routes becomes prohibitive. 364 Cybernetics for Systems Engineers – DRAFT F In order to reduce our search space we apply the principle of optimality by considering first the potential paths which lead to B. If we zoom in on the right hand edge, we can examine the paths from the nodes closest to the target node B, e.g. node x. There are two potential paths; the red one with a cost of 16, and the blue with a cost of 18. To advance from x we follow the red path. Evidently, when trying out paths up to node x, we need advance no further, we just use this result. Similarly we can derive costs of (18, 29) and (20,32) for the direct and indirect (via node x) paths from the nodes above and below x. Similarly, we no longer need to consider the paths in the network to the right of these nodes as we have the minima 18,16 and 20 as costs to completion for the three start nodes of all paths to the right. Figure 22-2 : Final Stage of the Optimal Path This process may be repeated, only considering the path as far as the set of nodes for which the minimum cost functions are known. In this example, applying the optimality principle does not appear to save much effort, but the benefit increases with the complexity of the network. For obvious reasons, an illustrative example must be simple in order to explain anything. 365 Cybernetics for Systems Engineers – DRAFT F 22.3.2 Application to the Control Problem The principle of optimality implies that we must work backwards from the end state in order to derive candidate control schemes which minimise the weighted sum square of control and state from an arbitrary start condition. The idea may be illustrated using a scalar example, as the usual matrix presentation is probably a distraction to most readers, it certainly causes writer’s cramp. The system to be controlled has a single state and a single control input: 𝑥̇ = 𝑎𝑥 + 𝑏𝑢 Where a and b are constants (actually, they can be time-varying). The performance index is: 2 𝑡 𝐽 = 𝑠𝑥(𝑡𝑓 ) + ∫0 𝑓 𝑞𝑥 2 + 𝑟𝑢2 𝑑𝑡 We shall assume that control is applied at discrete intervals of duration (Δt) and is maintained at a constant level throughout the interval. The performance index J is calculated over a fixed time interval (tf=NΔt). Using the principle of optimality, we work backwards from the end state. We need to define cost to completion associated with the control strategy: 2 𝑡 𝐽∗ (𝑡𝑓 − 𝑡) = 𝑠𝑥(𝑡𝑓 ) + ∫𝑡 𝑓 𝑞𝑥 2 + 𝑟𝑢2 𝑑𝑡 We have: 2 𝐽∗ (0) = 𝑠𝑥(𝑡𝑓 ) Stepping back a single control interval: 2 𝑡 𝐽∗ (∆𝑡) = 𝑠𝑥(𝑡𝑓 ) + ∫𝑡 𝑓−∆𝑡 𝑞𝑥 2 + 𝑟𝑢𝑁−1 2 𝑑𝑡 𝑓 From the plant equation: 𝑥̇ = 𝑎𝑥 + 𝑏𝑢 Solving: 𝑏 𝑥𝑁 = 𝑥𝑁−1 𝑒 𝑎∆𝑡 + (𝑒 𝑎∆𝑡 − 1)𝑢𝑁−1 𝑎 If the time interval is small compared with 1/a, this becomes: 𝑥𝑁 = (1 + 𝑎∆𝑡)𝑥𝑁−1 + 𝑏∆𝑡𝑢𝑁−1 The cost to completion at time tf – Δt is: ∆𝑡 2 2 𝐽∗ (∆𝑡) = 𝑠(𝑥𝑁 )2 + ∫ 𝑞(1 + 𝑎𝑡)2 𝑥𝑁−1 + 2𝑞𝑏𝑡(1 − 𝑎𝑡)𝑥𝑁−1 𝑢𝑁−1 + (𝑏 2 𝑡 2 + 𝑟)𝑢𝑁−1 𝑑𝑡 0 366 Cybernetics for Systems Engineers – DRAFT F The integral to first order in Δt becomes: ∆𝑡 2 2 + 2𝑞𝑏𝑡(1 − 𝑎𝑡)𝑥𝑁−1 𝑢𝑁−1 + (𝑏 2 𝑡 2 + 𝑟)𝑢𝑁−1 𝑑𝑡 ≈ ∫0 𝑞(1 + 𝑎𝑡)2 𝑥𝑁−1 2 2 (𝑞𝑥𝑁−1 )∆𝑡 + 𝑟𝑢𝑁−1 The cost to completion is: 2 2 2 𝐽∗ (∆𝑡) = 𝑠 (((1 + 𝑎∆𝑡)𝑥𝑁−1 + 𝑏∆𝑡𝑢𝑁−1 )) + 𝑞𝑥𝑁−1 ∆𝑡 + 𝑟𝑢𝑁−1 ∆𝑡 This is a minimum if the control is chosen such that: 𝜕𝐽∗ (∆𝑡) 𝜕𝑢𝑁−1 =0 Or: 2𝑠𝑏∆𝑡((1 + 𝑎∆𝑡)𝑥𝑁−1 + 𝑏∆𝑡𝑢𝑁−1 ) + 2𝑟𝑢𝑁−1 ∆𝑡 = 0 𝑏 𝑟 𝑢𝑁−1 = 𝑠𝑥𝑁−1 This yields the result: The ‘optimal’ control is linear state feedback. Also, the second differential is positive, so this yields a minimum cost function. Both these results are a consequence of the quadratic form of the performance index, which was chosen for convenience, not because of its relevance to the problem. The problem reduces to finding a gain k such that the performance index: 2 𝑡 𝐽(𝑡𝑓 ) = 𝑠𝑥(𝑡𝑓 ) + ∫0 𝑓 𝑞𝑥 2 + 𝑟 (𝑥̇ 𝑏2 − 𝑎𝑥)2 𝑑𝑡 Is a minimum. The time series solution for x is: 𝑥 = 𝑥(0)𝑒 ((𝑎+𝑏𝑘)𝑡) If the system is stable, the final value of the state is expected to be small if the final time is taken as a multiple of the settling time. The performance index is then expected to be dominated by the integral. The integral becomes: (𝑞+𝑟𝑘 2 ) 𝑡 𝑥(0)2 ∫0 𝑓(𝑞 + 𝑟𝑘 2 ) 𝑒 (2(𝑎+𝑏𝑘)𝑡) 𝑑𝑡 = 𝑥(0)2 2(𝑎+𝑏𝑘) [𝑒 2(𝑎+𝑏𝑘)𝑡 − 1] It is difficult to see how the performance index could be minimised if a+bk were positive, so it is evident that the exponent must be negative. It follows that as the final time tends to infinity, the performance index tends to: (𝑞+𝑟𝑘 2 ) 𝐽(∞) → −𝑥(0)2 2(𝑎+𝑏𝑘) Differentiating with respect to k to find the gain to minimise J(∞): 367 Cybernetics for Systems Engineers – DRAFT F 𝑑𝐽 𝑑𝑘 = 0 = 𝑥(0)2 4𝑟𝑘(𝑎+𝑏𝑘)−2𝑏(𝑞+𝑟𝑘 2 ) 4(𝑎+𝑏𝑘)2 = 𝑥(0)2 2𝑟𝑘𝑎−𝑏𝑞+𝑏𝑟𝑘 2 2(𝑎+𝑏𝑘)2 The optimal gain is given by: 2𝑟𝑘𝑎 − 𝑏𝑞 + 𝑏𝑟𝑘 2 = 0 For consistency with the rest of the optimal control literature, and the form of the feedback control derived in the previous section, we shall introduce a variable p defined as: 𝑘=− 𝑝𝑏 𝑟 The optimal gain is given, in terms of this parameter, by: 2𝑝𝑎 + 𝑞 − 𝑝2 𝑟 =0 The derivation for an nth order system, is more cumbersome on account of the constraints of matrix algebra. It may teach us a lot about matrix algebra and constrained optimisation, but very little about control. The matrix equation is: 𝑃𝐴 + 𝐴𝑇 𝑃 + 𝑄 − 𝑃𝐵𝑅 −1 𝐵𝑇 𝑃 = 0 This is sometimes called an algebraic Riccati equation. The first three terms are recognisable as Liapunov’s Equation, which we have already met. The fourth term is quadratic, so is expected to be at least positive semi-definite. We require a positive definite solution for P for the system to be stable. The gain matrix is given by: 𝐾 = −𝑅 −1 𝐵𝑇 𝑃 The closed loop system equation becomes, with this feedback: 𝑥̇ = (𝐴 + 𝐵𝐾)𝑥 This form of ‘optimal’ control is called ‘LQ’ for Linear/quadratic, indicating that the plant equations are linear, whilst the performance index is quadratic. 22.4 Gyro Monorail Usually the Riccati equation must be solved numerically for specific cases, so several solutions are needed to provide the insight into how the system parameters relate to the plant behaviour. With modern computers, hundreds of solutions may be generated quickly, and this to some extent offsets the disadvantage of a method which has all the hallmarks of a sausage machine to be cranked by unskilled labour. Used sensibly, however, this method has much to recommend it. One of the few cases for which the Riccati equation has an analytical solution is the gyroscopic monorail. This, and the fact that it exhibits both right half plane poles and zeros, is the principal reason why this engineering curiosity features so prominently in this book. 368 Cybernetics for Systems Engineers – DRAFT F We shall make the assumption that the gyro momentum is large, so that the nutation and precession modes are widely separated. We shall use the LQ method to obtain values for the gimbal feedback and roll angle feedback gains. The plant equations governing the precession are: 𝐻𝜙̇ = 𝑀 −𝐻𝜃̇ = 𝑊ℎ𝜙 To recap; H is the gyro angular momentum, W the monorail weight, h the height of the centre of gravity above the rail and M is the controlling moment applied to the gimbal. The roll angle is φ, and the gimbal deflection is θ. The system matrix is, therefore: 0 𝐴 = (− 𝑊ℎ 𝐻 0 0) The control matrix is: 1 𝐵=( ) 0 We characterise the servo moment by a maximum value Mmax, so that the control weighting matrix becomes: 𝑅= 1 2 𝑀𝑚𝑎𝑥 We scale the roll angle and gimbal deflection by φm and θm respectively: 𝑄= 1 2 1 𝜙𝑚 ( 2 0 0 1 ) 2 𝜃𝑚 The factor 2 is included so that: |𝑥 𝑇 𝑄𝑥|~1 When the states are near their scaled values. The control weighting is similarly scaled if there is more than one control input. The algebraic Riccati equation is: 𝑝11 (𝑝 12 0 𝑝12 𝑊ℎ ) ( 𝑝22 − 𝐻 0 ) + (0 0 0 𝑊ℎ 𝑝 11 − 𝐻 ) (𝑝12 0 1 2 1 𝜙𝑚 𝑝12 ) + 𝑝22 2 0 ( This reduces to three simultaneous quadratic equations. 369 0 1 2 𝜃𝑚 ) 1 2 )(1 0)𝑃 = 0 + 𝑃 ( ) (𝑀𝑚𝑎𝑥 0 Cybernetics for Systems Engineers – DRAFT F 2 𝑊ℎ 𝑝 𝐻 12 𝑊ℎ 𝑝 𝐻 22 − 1 2 2𝜙𝑚 2 2 + 𝑀𝑚𝑎𝑥 𝑝11 =0 2 + 𝑀𝑚𝑎𝑥 𝑝11 𝑝12 = 0 1 2 2 𝑀𝑚𝑎𝑥 𝑝12 − 2𝜃2 = 0 𝑚 Which may be solved for the elements of P: 𝑝12 = − 1 √2𝜃𝑚 𝑀𝑚𝑎𝑥 𝑝11 = 1 2 2𝜙2 𝑊ℎ √1 + √ 𝑚 𝜃𝑚 𝑀𝑚𝑎𝑥 𝐻 √2𝜙𝑚 𝑀𝑚𝑎𝑥 𝑝22 = 𝐻 1 2 2𝜙2 𝑊ℎ √1 + √ 𝑚 𝑊ℎ 2𝜙𝑚 𝜃𝑚 𝜃𝑚 𝑀𝑚𝑎𝑥 𝐻 Actually, p22 is redundant. The resulting feedback gains are: 𝑀𝜙 = − 𝑀𝑚𝑎𝑥 √1 √2𝜙𝑚 + 2 𝑊ℎ 2√2𝜙𝑚 𝜃𝑚 𝑀𝑚𝑎𝑥 𝐻 For the roll angle feedback. 𝑀𝜃 = 𝑀𝑚𝑎𝑥 √2𝜃𝑚 For the gimbal angle feedback. This isn’t a bad approach to the monorail problem as it is evident that gimbal deflection must be kept small. In fact, there will always be an amplitude limit on the states. If not set by physical saturations, they will be determined by the effects of the small perturbation assumptions implicit in the plant equations. It should be noted that this example is very much the exception, and most real systems require a numerical solution. Indeed, the emergence of commercial analysis packages has been a mixed blessing, as experience in using bespoke software has become valued over the ability to do anything useful with it. Given that neither managers nor the HR departments know the first thing about the subject, this placebo approach of valuing software packages over skill has become almost universal. Why technically naive managers are given responsibility over cutting edge technology remains a mystery. 22.5 The Optimal Observer The linear-quadratic approach is merely pole placement with a particularly fancy algorithm, and suffers from the same 20dB per decade roll-off at high frequency. Like the pole placement method, it relies on full state feedback, so the separation theorem must be invoked to permit an observer to be constructed to estimate those states which are not measured. 370 Cybernetics for Systems Engineers – DRAFT F As there is an algorithm based on signal amplitude which appears to generate reasonable first-cut designs, the question arises as to whether a similar approach is possible with respect to the observer design. The controller appeared to be limited by saturation, so the LQ approach sought to keep states and controls within bounds. The corresponding observer limitation, which prevents us from placing observer poles too far over to the left, appears to be noise. If we just introduced noise originating from sensors, the problem cannot be formulated as an observer; we would need to trace the noise through to, for example, the servo, and determine levels which give rise to excessive power consumption, vibration or some other system limitation. What follows is a derivation of a type of Kalman filter, limited in scope to the problems that interest us, i.e. linear time-invariant systems. If our interest were in signal processing, we would adopt a more general approach. We assume the plant equations contain uncertainty, not in their parameters or missing dynamics, but in the form of random disturbances. Examples might be wind gusts or vibration. The plant equations are modified to take this random excitation into account: 𝑥̇ = 𝐴𝑥 + 𝐵𝑢 + 𝐹𝑣 Where v is a vector of zero mean white noise processes. Also, and less controversially, the measurements are assumed corrupted by noise: 𝑦 = 𝐶𝑥 + 𝑤 Where w is also a vector of zero mean white noise processes. Our objective is to use our knowledge of the disturbance and noise statistics to design an observer which produces an estimate with the least variance. 22.5.1 Covariance Equation In order to proceed, we must extend our scalar expression for the response of a system to white noise to the vector case. To simplify matters we shall ignore the control input and consider the response of the simpler system: 𝑥̇ = 𝐴𝑥 + 𝐹𝑣 The white noise input is tricky to define as an explicit function of time, so we shall treat it as a discrete random sequence which is updated at a sampling interval of Δt, which is short. compared with the time constants of the system we are dealing with. A single sample has a vector of values v(i) at time iΔt. The time series solution is: 𝑥 = 𝑒 𝐴𝑡 𝑥(0) + ∆𝑡[𝑒 𝐴𝑡 𝐹𝑣(0) + 𝑒 𝐴(𝑡−∆𝑡) 𝐹𝑣(1) + ⋯ + 𝑒 𝐴(𝑡−𝑘∆𝑡) 𝐹𝑣(𝑘)] In the vector case, the scalar variance is replaced with a covariance matrix, the diagonal terms of which are the component variances, whilst the off-diagonal terms account for correlation between the components. 371 Cybernetics for Systems Engineers – DRAFT F Figure 22-3 : White Noise Sequence It is given by: 𝑋 = 𝐸(𝑥𝑥 𝑇 ) Where E is the expected value, i.e. the value found by averaging over a large ensemble of cases. The covariance is: 𝑋 = 𝑒 𝐴𝑡 𝑥(0)𝑥(0)𝑇 [𝑒 𝐴𝑡 ]𝑇 + (∆𝑡)2 [𝑒 𝐴𝑡 𝐹𝑣(0)𝑣(0)𝑇 𝐹 𝑇 [𝑒 𝐴𝑡 ]𝑇 + 𝑇 𝑒 𝐴(𝑡−∆𝑡) 𝐹𝑣(1)𝑣(1)𝑇 𝐹 𝑇 [𝑒 𝐴(𝑡−∆𝑡) ] ⋯ ] All the terms in x(0)v(i)T are zero, when taken over the sample, because v has zero mean. Since the noise is uncorrelated between samples v(i)v(j)T=0, unless i=j. Also, the disturbance is a stationary process, so that: 𝐸(𝑣(𝑖)𝑣(𝑖)𝑇 ) = 𝑉 ∀ 𝑖 𝑇 𝑋 = 𝑒 𝐴𝑡 𝑋(0)[𝑒 𝐴𝑡 ]𝑇 + (∆𝑡)2 ∑𝑘0 𝑒 𝐴(𝑡−𝑖∆𝑡) 𝐹𝑉𝐹 𝑇 [∑𝑘0 𝑒 𝐴(𝑡−𝑖∆𝑡) ] As Δt→0 𝑡 𝑡 𝑋 → 𝑒 𝐴𝑡 𝑋(0)[𝑒 𝐴𝑡 ]𝑇 + ∫0 𝑒 𝐴(𝑡−𝜏) 𝑑𝜏 𝐹𝑉𝐹 𝑇 [∫0 𝑒 𝐴(𝑡−𝜏) 𝑑𝜏] 𝑇 𝑋 = 𝑒 𝐴𝑡 𝑋(0)[𝑒 𝐴𝑡 ]𝑇 + 𝐴−1 𝑒 𝐴𝑡 𝐹𝑉𝐹 𝑇 [𝐴−1 𝑒 𝐴𝑡 ]𝑇 Differentiating yields an expression for the time evolution of the state covariance: 372 Cybernetics for Systems Engineers – DRAFT F 𝑋̇ = 𝐴𝑋 + 𝑋𝐴𝑇 + 𝐹𝑉𝐹 𝑇 Assuming A is stable, this converges on to a steady state solution: 𝐴𝑋 + 𝑋𝐴𝑇 + 𝐹𝑉𝐹 𝑇 = 0 This correlation equation bears a strong resemblance to the Liapunov equation used to test stability. 22.5.2 The Observer The observer structure is the same as for a Luneberger observer: 𝑥̂̇ = 𝐴𝑥̂ + 𝐵𝑢 + 𝐾(𝑦 − 𝐶𝑥̂) The error in the observer estimate is: 𝑒̇ = 𝑥̇ − 𝑥̂̇ = 𝐴𝑒 + 𝐹𝑣 + 𝐾𝐶𝑒 − 𝐾𝑤 𝑒̇ = (𝐴 + 𝐾𝐶)𝑒 + (𝐹𝑣 − 𝐾𝑤) Or: This takes the form of a homogenous equation fed with white noise. The white noise source is a combination of disturbance and measurement noise. 𝑞 = 𝐹𝑣 − 𝐾𝑤 From which: 𝑞𝑞 𝑇 = 𝐹𝑣𝑣 𝑇 𝐹 𝑇 − 𝐹𝑣𝑤 𝑇 𝐾 𝑇 − 𝐾𝑤𝑣 𝑇 𝐹 𝑇 + 𝐾𝑤𝑤 𝑇 𝐾 𝑇 We would not normally expect the disturbance noise to be correlated with the measurement noise, so the expected value of the observer noise input is: 𝐸(𝑞𝑞 𝑇 ) = 𝐹𝑉𝐹 𝑇 + 𝐾𝑊𝐾 𝑇 Where W is the measurement noise covariance matrix. The steady state covariance equation for the observer becomes: (𝐴 + 𝐾𝐶)𝑋 + 𝑋(𝐴 + 𝐾𝐶)𝑇 + 𝐹𝑉𝐹 𝑇 + 𝐾𝑊𝐾 𝑇 = 0 In order to get some idea of the form of the solution, let us consider the single state case, in which the matrices become scalars. 2(𝑎 + 𝑘𝑐)𝑥 + 𝑓 2 𝑣 + 𝑘 2 𝑤 = 0 Or 𝑥=− (𝑓2 𝑣+𝑘 2 𝑤) 2(𝑎+𝑘𝑐) , 𝑑𝑥 𝑑𝑘 = 4𝑘𝑤(𝑎+𝑘𝑐)−2𝑐(𝑓2 𝑣+𝑘 2 𝑤) 4(𝑎+𝑘𝑐)2 This has a turning point where: 2𝑘𝑤𝑎 + 𝑘 2 𝑐𝑤 − 𝑐𝑓 2 𝑣 = 0 For consistency with the Kalman filter literature, let; 𝑐 𝑘 = −𝑤𝑝 The minimum variance gain is found from: 373 =0 Cybernetics for Systems Engineers – DRAFT F 2𝑝𝑎 + 𝑓 2 𝑣 − 𝑝2 𝑐 2 𝑤 =0 The matrix equivalent of this expression is an algebraic Riccati equation, similar to that derived for LQ control: 𝐴𝑃 + 𝑃𝐴𝑇 + 𝐹𝑉𝐹 𝑇 − 𝑃𝐶 𝑇 𝑊 −1 𝐶𝑃 = 0 Since the gains are calculated from a similar equation, the observer problem is often called a ‘dual’ to the control problem. The gain matrix becomes: 𝐾 = −𝑃𝐶 𝑇 𝑊 −1 We call this a minimum variance filter, rather than a Kalman filter, to indicate that Kalman filtering is a major area of study in its own right, which we have only touched on. All we have done is presented the fundamental ideas behind the Kalman filter, without, I hope, obscuring the principles with gobbledegook. 22.5.3 Tracker Example Analytical solutions of the algebraic Riccati equation for the observer are as rare as they are for the controller. We are restricted to specific numerical solutions, waving the Kalman magic wand, and abracadabra - we have ‘the best’ observer. We shall consider the problem of tracking a faint object by means of a steerable antenna. The antenna is subjected to wind gusts, which usually have a defined spectrum, but we shall assume a Gaussian white noise process. The tracker specification is expected to be based on a requirement to track a target of known brightness to a specified accuracy at a particular range, whilst minimising the power and antenna size and weight. In such circumstances, we might expect performance to be noise limited. We shall assume only the boresight error is available as a measurement. The system equation is: 0 𝐴=( 1 0 ) 0 𝑉 = 𝜎𝐺2 1 𝐹 = (𝐽 ) 0 𝐶 = (0 1) 2 𝑊 = 𝜎𝑚 Where σ denotes standard deviation, subscript G refers to the gust torque, J is the moment of inertia of the antenna, and subscript m refers to the measurement noise. Evaluating the terms in the algebraic Riccati equation. 374 Cybernetics for Systems Engineers – DRAFT F 2 𝜎𝐺 0 𝑝11 0 ) , 𝑃𝐴𝑇 = (𝐴𝑃)𝑇 = ( ) , 𝐹𝑉𝐹 𝑇 = ( 𝐽2 𝑝12 0 𝑝12 0 0 𝐴𝑃 = ( 𝑝11 0) 0 And: 𝑃𝐶 𝑇 𝑊 −1 𝐶𝑃 = 2 𝑝12 1 ( 2 𝜎𝑚 𝑝12 𝑝22 𝑝12 𝑝22 ) 2 𝑝22 The Riccati equation becomes: 𝜎𝐺2 𝐽2 1 2 2 𝑝12 𝜎𝑚 1 2 𝑝12 𝑝22 𝜎𝑚 − ( 𝑝11 − 1 2 𝑝12 𝑝22 𝜎𝑚 ) 1 2 2𝑝12 − 𝜎2 𝑝22 𝑚 𝑝11 − =0 Solving: 𝑝12 = 2 √𝜎𝐺2 𝜎𝑚 𝐽 2𝑝 ,𝑝 , 𝑝22 = √2𝜎𝑚 12 11 = 𝑝12 𝑝22 2 𝜎𝑚 The gain matrix is: 1 𝜎𝐺2 √ 2 𝐽 𝜎𝑚 𝐾= 1 ( 2 𝜎2 4 √𝐽 (𝜎2𝐺 ) 𝑚 ) And p11 is redundant. The gain matrix contains the ratio of the variances. This is a general feature of the solution, so white noise intensities could be used to characterise the noise sources, instead of variances, without affecting the result. 22.5.4 Alpha beta Filter ‘Alpha beta’ and ‘alpha beta gamma’ filters are observers based on the assumption that target motion is adequately represented as constant velocity, or constant acceleration. It is assumed that position is measured, differenced with the filter estimate of position and fed back via gains, usually denoted α, to the position estimator , β/Δt to the velocity estimator, and 2γ/Δt2 to the acceleration estimator. Here Δt is the sampling period. Values of α, β and γ are usually found by trial and error using a system model. 22.6 Comment The gain matrix indicates that as the measurement noise increases, the observer gains reduce, and the observer pays more attention to its internal model of the plant, and less attention to the measurements. This effectively moves the observer poles to the right, into the region on the Argand diagram where we should expect to find the controller poles. We know that the plant equations in the observer are approximations, so the separation theorem doesn’t strictly apply, but we encounter the most serious interactions between observer and controller when their poles are adjacent. 375 Cybernetics for Systems Engineers – DRAFT F It is this ‘open-loop’ behaviour, which contravenes the fundamental principles of the error-driven regulator. This renders the Kalman filter approach questionable when applied to control systems. Kalman filters provide the best observers when plant equations are known accurately, and all uncertainty is indeed attributable to disturbances and measurement errors. A classical example is in inertial navigation, where the sensors are characterised and calibrated to the limits of available precision, and the statistics of the residual error bounds obtained. Elsewhere, the claim to optimality is a bit mischievous. It is reasonable to claim that if the system designed using a Kalman filter based observer becomes unsatisfactory due to the interaction with the controller poles, we need to consider whether the candidate sensors are really up to the job, and higher performance sensors might be needed. The Gaussian disturbance noise may be interpreted as a characterisation of the variation in system dynamics over an ensemble of systems, rather than an actual physical disturbance of an individual system. All observers based on the separation theorem remain blissfully ignorant of the conflict between accurate kinematic tracking of a moving target and minimisation of statistical tracking error. As with all the other ideas presented in this book, the reader is encouraged to grasp the underlying principles, and to understand where each has its application. We are not looking for the Philosopher’s Stone, merely a set of tools, which if used skilfully, will yield insights and effective designs. For a while it appeared that the control problem was ‘cracked’; with controllers designed using LQ ‘optimal’ methods and observers with Kalman filter methods (the resulting systems are called LQG controllers), each designed in isolation by invoking the separation theorem. Indeed, before 1980, very little was accepted for publication unless it was based on the virtues of these methods. When practitioners started to use them in anger, however, their limitations rapidly became manifest. Designing our observer on the one-dimensional criterion of minimum variance is only ever going to be a partial solution. The absence of the noise spectrum from the Kalman filter may not be a cause of concern if it is used in a more or less open loop manner, as is frequently the case. If the output is to be fed directly into other elements in a closed loop system, we really need to know something about its frequency content so that we know which modes it might excite. Typically, we would desire a much sharper high frequency cut off than 20dB per decade. Such a gradual roll off is in danger of exciting modes which have been neglected from the analysis, in particular, lightly damped structural modes could be excited, with catastrophic consequences. 376 Cybernetics for Systems Engineers – DRAFT F 23 Multiple Inputs 23.1 Introduction We have seen that additional outputs give us considerable control over the closed loop pole locations, which may justify the inclusion of additional sensors within the loop. Multiple inputs imply multiple servos, which remain potentially expensive system elements, and will only be considered if there is good reason. The ability to manipulate matrices does not qualify as a good reason. It is difficult to see how the proper function of any plant could be specified as requiring crosscoupling between functionally independent channels. MIMO systems emerge from processes which are inherently strongly cross coupled, or become cross coupled for over-riding cost, weight or power consumption reasons. An example of the former is the environmental control of a greenhouse where the temperature, humidity and carbon dioxide concentration interact, but each must somehow be controlled independently of the other two. The most ubiquitous example of the latter is the classical aeroplane. Unlike a missile, which is designed for manoeuvrability and rapid response, it is not feasible in an aeroplane to have separate lifting surfaces for yaw and pitch. Only a single lifting surface is available, and this introduces inevitable-cross-coupling between roll and yaw, which is absent from the more complex configuration of the missile. As should be expected from the principle of requisite entropy, simplification of the configuration leads to more complex (NOT simpler) behaviour. It cannot be emphasised strongly enough that we do not begin control system design until the dynamics of the open loop are fully understood, and we know the reasons why cross-coupling exists and cannot be avoided. If we find that the plant is heavily cross-coupled with no frequency separation between its modes, the wisest course of action is probably to scrap it, and not waste time trying to make the proverbial silk purse from the sow’s ear. We have seen that instability is generally encountered as we try to speed up the response time in a single-input/single output system. In the MIMO case this is aggravated by parasitic cross coupling between functionally independent channels, and this must be taken into account when we try to achieve high system bandwidth, so there is usually practical merit in studying MIMO systems. For a while it appeared that the LQG approached furnished a means of designing systems on the basis of state limits and noise levels. Indeed, used sensibly, these methods can be very effective. However, the absence of explicit stability margins and the questionable assumptions behind the separation theorem has led to some precarious designs. This is an on-going area of research, and the literature is very much in a state of flux. With few practical applications and even fewer opportunities for practitioners to publish, the available text books generally remain opaque to the average engineer. 377 Cybernetics for Systems Engineers – DRAFT F 23.2 Eigenstructure Assignment We found in the monorail example that decoupling between modes becomes possible by separating the modes in frequency. This may be thought of as analogous to having different effects arising from double-clicking and single clicking a computer mouse. The idea behind modal separation is not dissimilar to the idea behind the derivation of the linear small perturbation equations from inherently non-linear equations. We are really only concerned with the stability of set points, which are defined outside the current loop. The value of a particular state variable, controlled by a low frequency mode, is effectively an equilibrium value as far as a high frequency mode is concerned. Similarly, the rapid fluctuations in the state associated with the faster mode represent a zero mean noise source, above the cut off frequency of the slower mode. This is the reason why, if we nest loops, we cannot expect the outer loop to respond more quickly than the inner loop. It is why the ‘low level’ dynamics dictate the behaviour of the overall system. So much for the nonsensical dogma of ‘top down’ analysis. Thus, we can have independent control over the pitch transients of an aircraft as well as the long term control of the flight path, through a single elevator control. We found that full state feedback gives us the option to place the closed loop poles where we wish, although we included the health warning that they should reflect the modal behaviour of the plant. A similar result exists for the multi-input case. Not only can the mode frequencies (eigenvalues) be influenced by feedback, but the extra control can be used to suppress or encourage the participation of different states in each mode (i.e. the eigenvectors may be altered). We would not begin to think of using this method without first acquiring a fundamental understanding of the open loop plant behaviour. As with all methods in this and later chapters, only the bare bones will be presented here. 23.3 Using Full State Feedback We shall assume that estimates of all the state variables are available via some form of observer. With full state feedback, and a state-controllable system, we can place the closed loop poles anywhere we choose. We also have (potentially) control over as many elements of the eigenvectors as we have control inputs. Thus with two inputs we can set to zero elements of the eigenvector corresponding to states which we do not wish to take part in the corresponding mode. This manipulation of the eigenvector must be undertaken with care. It is evident that any attempt, for example, to remove the pitch rate from the short period pitch oscillation mode, is likely to lead to some bizarre results. 23.3.1 Basic Method We start by specifying a desired closed loop eigenvalue and the elements associated with the eigenvector (mode shape). The number of eigenvectors which may then be specified is equal to the maximum of the rows of the measurement matrix and the columns of the control matrix (assuming 378 Cybernetics for Systems Engineers – DRAFT F both have full rank). The number of elements of each eigenvector which may be assigned is equal to the minimum of these two. Taking lateral autopilot of a monoplane, the modes of interest are Dutch roll and roll subsidence. The spiral mode is taken care of by the trajectory guidance. The state vector is: v r p Where v is sideslip, r is yaw rate and p is roll rate, is the perturbation roll angle with respect to the plane of incidence in a steady turn. The plant equations contain cross-coupling derivatives e.g. Lv (roll due to sideslip, or dihedral effect) and the obvious manoeuvre coupling via the roll perturbation. Any particular mode consists of an exponential (simple or complex) function of time multiplied by a vector defining the amount by which each state is affected by the particular mode. The plant equations reflect the fact that the open loop plant is usually designed such that the each control input is associated with a mode which excites specific states, whilst avoiding excitation of others. It is this inherent decoupling which permits separate autopilots to be designed for the yaw, pitch and roll channels of skid to turn missiles. Even here there is weak cross-coupling, but adequate stability margins generally ensure these have minimal effect as far as the closed loop control is concerned. The monoplane configuration arises largely because of the economy in terms of weight and profile drag of having a single principal pair of lifting surfaces. The penalty is an inability to apply lateral forces rapidly, as the limited control authority is required for simultaneous roll and pitch manoeuvres, whilst regulating the sideslip near zero. This is particularly serious during a reversal of (fixed axes) yaw command direction, when saturation is almost inevitable. This is inherent in monoplane configured vehicles, and no amount of clever autopilot design will compensate for this intrinsic limitation of the plant. What we can do with eigenstructure assignment is ensure that the Dutch roll and roll subsidence modes are adequately decoupled. However, we can do more than that. Whether bank to turn or skid to turn, an explicit method for decoupling roll and yaw is necessary, and eigenstructure assignment is a potential option to consider for this purpose. 23.3.2 Fitting the Eigenvectors In the monoplane case we have two modes of interest, and what we wish to do in the first instance, is use the control to decouple Dutch roll from roll subsidence, although the two are strongly coupled in the plant equations. We have two control inputs, so we can specify two elements of the eigenvector, e,g. consider Dutch roll, with the rows of the eigenvector corresponding to the rows in the state vector: 379 Cybernetics for Systems Engineers – DRAFT F Sideslip: 1 * * 0 In other words, the Dutch roll pole is associated with sideslip but not roll perturbation. The asterisk denotes that these other elements do not concern us. Similarly for yaw rate we expect. Yaw rate: * 1 * 0 The eigenvalues will form a complex conjugate pair for the Dutch roll mode. Similarly, we have for the roll subsidence mode: Roll rate: 0 * 1 * Roll perturbation: 0 * * 1 The roll perturbation, influences sideslip and yaw rate via the lateral force equation, but the only feedback to the roll perturbation is via the roll rate. We specify two pairs of complex conjugate eigenvalues corresponding to the time responses of these two modes. For each eigenvalue, we calculate the values of the transfer matrix from output state to control input: Ti Isi A B 1 Where the plant equation can be written in the form: x Ax Bu Where x is the state, u the control input, A the system matrix and B the control matrix. The control eigenvector νi corresponding to the specified state eigenvector is given by: i Ti i Where ηi is the specified state eigenvector. The transfer matrix must be re-ordered so that the rows and columns associated with the * entries are distinct from those containing numerical values and can consequently be ignored. 380 Cybernetics for Systems Engineers – DRAFT F The resulting re-ordered transfer matrix has nu rows and nx columns and will be denoted Li (note: this is not to be confused with a roll derivative). The control vector corresponding to the ith specified eigenvector is obtained from the pseudoinverse of Li : 1 i Li T Li Li i T The eigenvectors which are actually achievable are: i Ti i It is usual to display this intermediate result, to allow the designer to adjust the eigenvalue to more nearly fit the desired eigenvector. Absurd combinations of eigenvector elements, e.g. requiring zero values for the derivative of a state which is desired, are not expected to fit the actual plant very well. Some understanding of the dynamics of the plant is required before the technique can be used effectively. 23.3.3 Calculation of Gains Once a set of eigenvectors have been decided on and appear feasible, it is necessary to calculate the feedback gain matrix F, where: u Fx We shall assume full state feedback for the sake of clarity. Methods are described in the literature for output feedback, but for the moment it will be noted that this is really a pole placement method for which the separation theorem applies, so a full state observer may be constructed to estimate the state from the available output. The robustness of this process must be tested afterwards. From our achievable eigenvector we have: i si Ai BFi For all nx we have: A BF Or: Z BFN Where N is the matrix whose columns are the achievable eigenvectors, Λ is the diagonal matrix of corresponding eigenvalues. If the eigenvalues are distinct, the inverse of Ν exists, so the gain matrix is given by: F BT B 381 1 B T ZN 1 Cybernetics for Systems Engineers – DRAFT F This is a least squares fit, which may result in the closed loop poles being displaced from their nominal positions, further analysis is required to ensure the resulting system meets stability and robustness criteria. Quite often the control matrix is of the form: I B u 0 In which case F is calculated by ignoring the lower nx-nu rows of ZN-1. More generally, the control matrix is transformed into this form by the following method: The control matrix is of the form: X B Y Where X is nuхnu and Y is (nx -nu )хnu. Pre-multiply this by matrix E which is partitioned to match that of B: L E N M P Such that: I EB 0 The sub-matrices are found from: LX MY I NX PY 0 N PYX 1 L X 1 MYX 1 Sub-matrices M and P may be chosen at random. The matrix ZN-1 is pre-multiplied by E, and the final nx-nu rows ignored. 23.3.4 Comment on Stability The process of assigning eigenvectors amounts to a least squares squares fit, which can go badly wrong, so this is not a once-through handle-turning exercise. The resulting loops will need to be analysed, and additional compensation introduced as required. However, the method has the advantage of relating design closely to the actual system proper function requirements, rather than requiring a working solution to spontaneously appear through esoteric magic. 382 Cybernetics for Systems Engineers – DRAFT F 23.4 Loop Methods The single-input/single-output design methods are all based on understanding how the open loop dynamics influences the closed loop plant. This information is used to design compensators to distort the appropriate plot into one which has more favourable characteristics, as measured by gain and phase margins. We have deviated from this approach when we introduced the state-space methods, in the form of pole placement and its variations, and have worked directly with the closed loop. The problem with this approach is the impossibility of defining the stability margins, so the robustness of the design solutions remains little better than a matter of faith. At about the same time as most people were adopting LQG exclusively as the way ahead, extensive work was undertaken to extend the ideas of SISO design, to MIMO system. Whilst these methods have practical merit, the whole loop-based approach is inherently unsuitable for MIMO systems, which is why LQG was heralded as the only general method which could tackle this type of system. 23.4.1 General Considerations Transfer functions are derived either by combining the transfer functions of the individual components making up the plant or from the differential, or difference, equations defining the plant behaviour. In a multi-input multi-output system we are evidently dealing with matrices of transfer functions. As servos tend to be considerably more expensive than sensors, we would typically expect the number of outputs to exceed the number of inputs, so our transfer matrices are rarely expected to be square. We note that the plant poles remain the eigenvalues of the system matrix, regardless of the numbers of inputs . When deriving the transfer matrices from component transfer functions we need to be careful with pole-zero cancellation, especially if the loop contains improper transfer functions, such as was present in the missile two-loop autopilot example. Most controllers are designed to maintain performance about an ensemble of operating points, so we should expect the commonest reference input to be a fixed set point. To begin with, therefore, we would seek combinations of inputs and outputs are most strongly related at equilibrium. In order to perform this analysis of the inherent controllability of the plant it is necessary to scale the inputs and outputs by dividing them by their limiting values. A set of n feedbacks (where n is the order of the system) are chosen using this information together with the state controllability criterion, such that the poles are placed appropriately. In practically all MIMO analysis and design it is necessary to scale all quantities, and modify the plant equations to reflect the scaled values of the variables. To date, we have used ‘input’ to mean the inputs to the open loop plant, and ‘outputs’ for the sensor measurements. We shall introduce the term ‘reference’ to mean the input to the closed, and ‘outputs’ to refer to the measurements of the quantities the system is required to control. With this proviso, that we are dealing with a system that has already passed through at least one design iteration, we are justified in treating the modified plant transfer matrix as ‘square’. 383 Cybernetics for Systems Engineers – DRAFT F 23.4.2 Nyquist Criterion When we consider extending frequency domain analysis to multi-input systems, we really need to consider whether what we are doing makes sense. Any method which considers phase information must be clear regarding what the phase reference is. The single input case has the phase of the input as a reference. If we have more than one input, is it reasonable to imagine that both inputs will necessarily be synchronised in phase? Surely the phase of one input is independent of the other and any combination of phase is possible. We can derive a stability criterion analogous to the SISO Nyquist criterion, using the determinants and eigenvalues of the transfer function matrix, and an identical encirclement criterion arises. However, as an extrapolation of the SISO case, it implicitly assumes phase synchronicity between the inputs, which is never the case. Figure 23-1: Gershgorin Circles In fact, in the MIMO case, we have great difficulty in defining what we mean by ‘phase’. This is unfortunate because we have identified phase lag as the principal cause of loss of stability in the single input case. There is therefore a culture shock awaiting the traditional control engineer, in discovering that the familiar approaches are of questionable value, and we are better off dealing with the closed loop directly, rather than using the more familiar loop-shaping approach. However, if the plant is not strongly cross-coupled we can still deal with it as if it were decoupled. Consider a plant having transfer function: 𝐺 𝐺(𝜔) = ( 11 𝐺21 𝐺12 ) 𝐺22 The matrix elements are responses to sinusoidal inputs of unit magnitude. 384 Cybernetics for Systems Engineers – DRAFT F The first output is given by: 𝑦1 = 𝐺11 𝑢1 + 𝐺12 𝑢2 Now u1 and u2 have unit magnitude but the phase of u2 with respect to u1 is random, so any point on the actual harmonic locus lies within the circle centred on G11, of radius |G12|. More generally, the harmonic locus lies within a band defined by circles of radius equal to the sums of the magnitudes of the off diagonal elements, centred on the diagonal elements. This is known as Gershgorin’s theorem, and furnishes a means of applying Nyquist’s criterion to a system which isn’t too far from diagonal. In fact, such systems are quite common. The circles superimposed on the harmonic locus are called Gershgorin circles, and likewise the regions encompassing all the Gershgorin circles are called Gershgorin bands. If the Gershgorin bands are narrow near the -1 point, the system is said to be diagonally dominant. If the Gershgorin band includes the -1 point, the system is not necessarily unstable, so large diameter Gershgorin circles near the -1 point will usually lead to excessively conservative designs. 23.5 Sequential Loop Closing It is often the case that additional controls serve specifically to control modes which are widely separated from each other. The throttle control of an aircraft usually has a very slow response, as it is dictated by the dynamic behaviour of the engines, but it usually influences the trajectory modes, having negligible effect on the short period handling modes. The same can be said about the trim tab. The pitch plane control appears to be a three input system, but it is evident that the associated modes are decoupled by virtue of their frequency separation. Similarly, we may refine the gyro monorail by including an additional input causing the centre of gravity to shift laterally, so that the vehicle will not be affected by steady state contact forces such as offset loads or side winds (inertial forces are automatically dealt with by the basic system, which will cause the vehicle to bank on bends). Since the objective is to balance almost steady state disturbances, the associated mode is expected to be very slow compared with those of the balancing system. A submarine has ballast tanks to control buoyancy, but they are useless for controlling depth. That requires the dynamic lift, and rapid response, of hydroplanes. Many systems are of this form, where it is impossible to force the actuation of slow modes to influence higher frequency modes without excessive power or saturation, by virtue of the fundamental dynamics of the system. In such cases, we should expect to be able to close the loops one at a time, starting from the highest frequency mode. Designing the loops to accepted values of gain and phase margin should be sufficient to ensure the interaction between the modes should be minimal. 385 Cybernetics for Systems Engineers – DRAFT F 23.6 Diagonalisation Methods 23.6.1 ‘Obvious’ Approach A naive approach might imagine designing the compensator in two stages; the first to render the plant/compensator system diagonal, and the second to adjust the dynamic behaviour. This will not work for a number of reasons, the first of which is that the inverse of a general plant is expected to introduce improper transfer functions (numerator order exceeding denominator), which are impossible to realise. Secondly, in the subsequent decoupled single loop analyses, the effect on robustness of the diagonalisation compensator is effectively ignored. Since in many cases this may be where most of the phase loss occurs, the resulting loop design becomes meaningless. We may design the system such that for quasi-static signals (such as we should expect from the reference input), the system is decoupled. This amounts to making the actual error signals linear combinations of the errors between reference inputs and plant outputs, to offset the static crosscoupling in the plant. The simple diagonalisation of the plant by calculating its inverse also raises issues of integrity, in that each output is fed to every input, so that failure of a single sensor could cause simultaneous failure of all channels rather than just one. 23.6.2 Commutative Compensators The idea behind commutative compensation is to produce a transformation which diagonalises the plant transfer matrix using a similarity transform, rather like the eigenvalue/eigenvector decomposition of a system matrix. We seek a matrix W such that: 𝐺(𝑠) = 𝑊(𝑠)Λ(𝑠)𝑊(𝑠)−1 Where Λ(s) is diagonal. The similarity transform itself does not affect the actual plant dynamics, so that a compensator may be designed using single loop methods. The diagonal compensator may then be transformed back into the same space as the original plant, where it would actually be implemented. Obvious candidates for W are the eigenvectors of G(s). Unfortunately, these invariably contain improper transfer functions, which cannot be realised in practice. However, there is extensive literature, and even design software, available which enable close approximations to the ideal commutative compensators to be designed. The details of such methods lie outside the scope of a presentation which aims at breadth, rather than depth, of cover. 23.7 Comment Loop based approaches, in which the effect of changes to the open loop plant on the closed loop is used to design compensators, are extremely effective for single-input/single output systems. Their extension to MIMO systems requires a certain ingenuity and creativity together with comprehensive understanding of the plant dynamics. The former is usually effectively suppressed by our academic institutions, while the need for the latter is seldom recognised by a technically naive management. 386 Cybernetics for Systems Engineers – DRAFT F The absence of a single phase reference, which is so fundamental to single input systems, would indicate that loop methods are not really transferrable to the MIMO case. This is perhaps the reason why methods which worked directly with the closed loop, such as pole placement and LQG, quickly overshadowed loop based methods, until their potential lack of robustness became apparent. 387 Cybernetics for Systems Engineers – DRAFT F 24 Closed Loop Methods 24.1 Introduction The control systems which have been designed in the past tended to be mechanisations of tasks originally performed by human beings. Contrary to what appears to be the current wisdom, human beings have severe limitations, which require tasks to be largely decoupled, either in separate physical channels or in time, with different modes requiring attention over different time periods. Just channelling all the information to a human being without filtering it down to the relatively few bits per second he/she can handle, is a recipe for disaster. Admittedly, there is always the cop-out of ‘pilot error’ as a fig leaf for all manner of poor system design. The control engineer views the human being as merely another dynamic component, whose behaviour when performing a specific task is not as difficult to characterise as human vanity believes. These limitations are reflected in the plant. So when we come to automate existing tasks, they are usually nicely decoupled into functionally independent channels or separated in frequency for which intervention is required. SISO methods are adequate, so that multivariable control techniques seem to lack relevance in the real world. For this reason, and the fact that most presentations are well beyond the event horizon as far as clarity is concerned, there has been a reluctance on the part of practitioners to study the methods which have emerged over the past 30 years. If control system specialists see no merit in investing time in the newer methods, how much less likely is the world of systems engineering to be impressed. The more we ask of the machinery, the greater the effect all the cross-coupled neglected dynamics will have. The chemical industry furnishes us with many processes which are strongly cross-coupled, and the consequence of merely increasing flow rates to speed up production is unlikely to be a pretty sight. Systems of the future will be muti-variable, and as we have seen, SISO methods really are not up to the job. At best we can pretend that ‘square’ systems can be treated more or less like multiple SISO systems, but generally systems only become square AFTER we have designed the controller. Square plant is a rarity. In general, we cannot use loop methods for MIMO systems, so we are forced to adopt methods which apply directly to the closed loop. Also, our characterisation of robustness, in terms of gain and phase margin, has served well as ‘rules of thumb’, but outside the hallowed cloisters of control engineering, they are meaningless. We are better off describing our stability margins in terms of the variability in the values of the plant parameters which can reasonably be expected. These are meaningful to other specialists. They provide a means by which the control engineer can specify tolerances on components. In principle, the information could be used to define the probability of a design becoming unstable in service. This is what the safety people actually want to know. The following begins with SISO systems, to see how we characterise the closed loop in order to address the actual design issues which confront us. We use the idea hinted at when we considered 388 Cybernetics for Systems Engineers – DRAFT F the effect of noise on a tracking sensor, that there is a conflict between suppressing noise originating within the loop and at the same time ensuring accurate kinematic tracking of the reference signal. Unlike the traditional SISO approaches, these methods may be extended readily to the MIMO case. 24.2 The H-infinity Norm Most practitioners would have come across the term H∞, and wondered, like the author, about the publishing standards which permit mathematical symbols to enter the abstract, let alone the title, of a paper. Perhaps it indicates a limited command of the language, which cannot express the idea in plain English. Perhaps the concept is not fully understood by the originator, to the extent that it can be expressed in natural language. Perhaps that once explained, it loses the mystic air of expertise which seems to be associated with gobbledegook. Well, the term is now established and must be added to the list of control engineer’s jargon (together with M and N circles, phase margin, etc.) Although the idea is not needed immediately, we shall dispel the mysticism before proceeding. We mentioned, when optimising a signal in the time domain, that minimising least squares error seemed a reasonable basis for finding the ‘best’ control. It was pointed out that this was just a matter of convenience; we could have raised the error to the power p. As p tends to infinity, large deviations from the general trend of the data become penalised more heavily. In the time domain the rms value of a quantity is often associated with the energy conveyed, so does have some physical significance. In the frequency domain the energy of the signal distributed over the spectrum does not have the same significance. In the frequency domain our desire is to avoid resonance peaks, so that our characterisation of the mean deviation from the desired frequency response must avoid smoothing out such features. We can define a norm of a set of data values as: 1 𝑥̿𝑝 = 1 𝑝 𝑝 (𝑁 ∑𝑁 𝑖=1|𝑥𝑖 | ) The norm of the random data set for several values of p is presented in the figure. We see that as p→∞, the norm tends to the maximum peak, or the supremum, of the set. This is the vector infinity norm. This illustrates the idea behind H∞ methods, but is not in itself a definition of the H∞ norm. Perhaps calling it the ‘resonance peak alleviation’ method would have made up in accessibility what it lacks in accuracy. H∞ seems a spectacularly pretentious way of saying ‘the biggest’. 389 Cybernetics for Systems Engineers – DRAFT F Figure 24-1 : Illustration of the Effect of Index on Sample Norm 24.3 MIMO Signals It has been suggested that in the absence of a common input to act as a phase reference, phase information is meaningless in MIMO systems. By implication, the sinusoidal excitation of the system is of little use without exploring all possible phase relationships between the inputs. Nyquist, Bode and Nichols charts become misleading. We must deal with signal magnitudes. Consider the plant: 𝑦 = 𝐺(𝑠)𝑢 The magnitude of the output is: 𝑦 𝐻 𝑦 = 𝑢𝐻 𝐺(𝑠)𝐻 𝐺(𝑠)𝑢 Where H denotes a complex conjugate transpose. This deals strictly with signal amplitudes, and is only meaningful if outputs and inputs are scaled by their limiting values, such that the resulting amplitude is a pure number. The matrix G(s)HG(s) is symmetric, so its eigenvectors are orthogonal. Furthermore, its eigenvalues are all either positive or zero. 390 Cybernetics for Systems Engineers – DRAFT F The eigenvalue decomposition becomes: 𝐺(𝑠)𝐻 𝐺(𝑠) = 𝑈 𝐻 Σ2 𝑈 Where Σ2 is a diagonal matrix of eigenvalues. If 𝑣 = 𝑈𝑢 𝑣 𝐻 𝑣 = 𝑢𝐻 𝑈 𝐻 𝑈𝑢 = 𝑢𝐻 𝑢 Since U is orthogonal. This means the relationship between the magnitudes of input and output vectors depends only on the eigenvalues of GHG. This is all very well, but controlling the vector magnitude of the output is going a bit too far in our quest to avoid using signal phase. We desire to control individual outputs. The elements of Σ are gains relating the inputs to the outputs. The degree to which each input is amplified by each gain depends on the rows of U. These may be thought of as input directions corresponding to the gains. We can find the directions of the output: 𝐺 𝐻 𝑦 = 𝐺 𝐻 𝐺𝑢 = 𝑈 𝐻 Σ2 𝑈𝑢 𝑦 𝐻 𝐺𝐺 𝐻 𝑦 = 𝑢𝐻 𝑈 𝐻 Σ4 𝑈𝑢 The magnitude of the output can only be invariant if: 𝐺𝐺 𝐻 = 𝑌Γ 2 𝑌 𝐻 Where Y is orthogonal. For consistency with the expression for GHG: Γ=Σ And G may be written in the form: 𝐺 = 𝑌Σ𝑈 The ith column of Y characterises the effect of the ith gain on the output, whilst the ith row of U determines the input which affects the ith output (all other inputs having no effect). The elements of Σ are called the principal gains, or the singular values, and the partitioning of a rectangular matrix into two orthogonal matrices and a diagonal matrix of singular values is called a singular value decomposition. The singular value decomposition of the system matrix is of interest in its own right, as it reveals how inputs affect outputs over the frequency range, revealing clues as to how the system can be partitioned into functionally independent modes. Note that even if G is square, the system response is determined by the principal gains (singular values) and not the eigenvalues. The eigenvectors corresponding to eigenvalue decomposition are not orthogonal, as is needed to associate specific input and output directions with each gain. 391 Cybernetics for Systems Engineers – DRAFT F Note that the singular value decomposition is not a similarity transform. The left hand matrix multiplying the diagonal matrix of singular values is not the inverse of the right hand matrix. For a matrix which is a function of frequency (i.e. a transfer function), the H∞ norm is the supremum of the maximum principal gain, taken over the frequency range, i.e. the highest resonance peak of all the possible system responses. 24.4 Closed Loop Characterisation We have introduced the idea of input and output directions for MIMO systems, which impose limitations on achievable performance and robustness. These are additional considerations which will be addressed once the issues of working just with signal magnitude have been considered in the SISO case. We saw in the construction of Nichols charts, that simply specifying the desired closed loop response was insufficient to enable compensators to be designed, because the closed loop zeros are the same as those for the open loop. Typically, we would expect the compensator to introduce additional zeros, so apparently this approach would require the compensation to be known, before it could be applied. ‘Tail wag dog’ methods often arise when approaches are not properly thought through. However, we recall from considering the tracking properties of a steerable sensor (e.g. for communicating with a distant space vehicle), that there exists a conflict between minimising the response to sensor noise whilst accurately tracking a moving target. In fact, this is an extremely common problem, and enables the closed loop to be described by a pair of transfer functions, one characterising response to the input, the other the response to noise. We have: 𝑇= 𝐾𝐺 1+𝐾𝐺 Where K is the compensator transfer function, and G is the open loop transfer function. T is the usual closed loop transfer function, or loop transmission. The response to disturbances (and the error response to the input) is characterised by: 1 𝑆 = 1+𝐾𝐺 This is called the sensitivity. We have two transfer functions characterising the closed loop, one of which has the open loop zeros, and one which doesn’t. By considering the conflict between accurate tracking of the input, and minimising the response to noise, we have a potential technique for designing the compensation. 392 Cybernetics for Systems Engineers – DRAFT F 24.4.1 Tracker Example Figure 24-2: Tracking Loop with Measurement Noise and Gust Disturbance The sensor loop is as shown. The size and weight will presumably have been determined from initial detection capability, volume and cost considerations. Presumably the operating environment would be specified, so that local gust loading on the structure would be known. The loop consists of the desired sight line direction θD, which is not measured explicitly. The sensor measures the boresight error between the desired direction and the actual sensor pointing direction. The system could be driven by an explicit boresight demand during initial acquisition, involving an outer loop, optimised to minimise the acquisition time. The signal from a distant target is expected to be weak, so the boresight error will be contaminated with noise. We expect there to be some filtering of the boresight error combined with compensation for the loop dynamics (Kf). This feeds into some form of servo (Gs) to steer the sensor. The servo applies torque to the sensor, as does the wind. The total torque is the servo plus wind gusts. The wind gust is represented by a white noise source filtered so as to represent its expected spectral content (Gd) . 24.4.2 Comment This is actually a very high level view of the tracker. We are only considering a single channel, and we may choose to feed back the angular velocity of the sensor, rather than rely on position alone. This is expected to introduce additional biases and noise, the consideration of which would be deferred to a later iteration. 393 Cybernetics for Systems Engineers – DRAFT F This represents the bare bones of what is needed to understand the trade-offs between servo requirements and sensor detectivity, and to determine realistic power, weight and volume requirements for the tracker. If we are tracking a distant space craft, the sight line direction is expected to move slowly but the signal will be weak, so we expect the compensator time constant to be large and the signal energy may be integrated over a significant period of time whilst the boresight is held nearly constant. If we are tracking an incoming missile, loop response may be more important. We can frame the requirements in terms of the signal to noise, or some other means of characterising the noise input, and an overall loop response time. More often than not, we need to find the best response time which can be achieved using candidate sensors, and analysis of the higher level process will determine whether the available performance is adequate. Reducing response to measurement noise also slows down the response to target motion. Since the sensor measures bore sight error, the reference signal (the true sight line direction) is not available as an independent measurement. It is therefore impossible to apply a pre-filter to modify the demand, as is frequently suggested. The problem is not resolvable using control methods alone. We must proceed to a finer level of detail. Typically, we expect the more sensitive sensor to be larger, e.g. a larger dish is needed for greater gain, which is more vulnerable to gusts. There may be a trade off between sensor noise and disturbance noise at a secondary level of analysis. By ensuring that disturbance noise is dominant by virtue of the sensor design, we can optimise the control to maximise disturbance rejection. However, there are cost, weight and power implications in increasing the size of the plant, and it is evident that a compromise is needed. Any ‘optimisation’ which does not take the dynamic behaviour of the entire loop into account isn’t worth the paper it is written on. 24.4.3 Input Output Relationships From our tracking loop we can determine the output in terms of the inputs and disturbances. 𝜃 = (𝜃𝐷 − 𝜃)𝐾𝐺𝑠 𝐺𝑝 + 𝑚𝐾𝐺𝑠 𝐺𝑝 + 𝑑𝐺𝑑 𝐺𝑝 We are also interested in the input to the servo, as this characterises saturation. 𝑢 = (𝜃𝐷 − 𝜃)𝐾 + 𝑚𝐾 Re-arranging: 𝐾𝐺 𝐺 𝐾𝐺 𝐺 𝐺 𝐺 𝑠 𝑝 𝑠 𝑝 𝑑 𝑝 𝜃 = 1+𝐾𝐺 𝜃 + 1+𝐾𝐺 𝑚 + 1+𝐾𝐺 𝑑 𝐺 𝐷 𝐺 𝐺 𝑠 𝑝 𝑠 𝑝 𝑠 𝑝 And: 𝐾 𝐾𝐺 𝐺 𝐾 𝑑 𝑝 𝑢 = 1+𝐾𝐺 𝐺 𝜃𝐷 + 1+𝐾𝐺 𝐺 𝑚 − 1+𝐾𝐺 𝑑 𝐺 𝑠 𝑝 𝑠 𝑝 𝑠 𝑝 Or in terms of the sensitivity, S and transmission, T: 394 Cybernetics for Systems Engineers – DRAFT F 𝜃 = 𝑇(𝑚 + 𝜃𝐷 ) + 𝑆𝐺𝑑 𝐺𝑝 𝑑 𝑢 = 𝐾𝑆(𝑚 + 𝜃𝐷 ) − 𝐾𝑆𝐺𝑑 𝐺𝑝 𝑑 The tracking error is: 𝑒 = 𝜃𝐷 − 𝜃 = 𝑆𝜃𝐷 − 𝑇𝑚 − 𝑆𝐺𝑑 𝐺𝑝 𝑑 In this case the measurement noise was added into the boresight error measurement, because this is what the sensor actually measures. More typically, the measurement error is added to the output θ. The sole effect of which is a reversal of sign, which is irrelevant to a zero mean white noise process. Evidently, accurate tracking and complete disturbance rejection implies S=0 and T=1, or K→∞. But suppression of measurement noise requires T→0, in direct conflict to the requirement for disturbance rejection. We must seek a compromise. More often than not, we design the plant such that one or other noise source is dominant. Hence radar dishes are invariably housed inside radomes, to ensure that performance is limited by detector sensitivity. The processing may involve long observation times which dominate the compensator response, such that the details of the motion of the antenna become secondary. Some sensors may be electronically steered, in which case the system delays are determined almost entirely by the processing delays. This compromise between measurement and disturbance noise was resolved in the Kalman filter by producing a minimum variance state estimate. We have considered explicitly examining the effect of feedback on loop stability using ‘classical’ methods. It was also found that stability could be addressed implicitly by minimising tracking error over time using the linear quadratic method to calculate the gain matrix. We may also address the stability of the closed loop in the frequency domain by using a measure of mismatch between the desired and actual frequency responses which is sensitive to resonance peaks, which occur as instability is approached. The appropriate measure is the H∞ norm. For the reasons discussed, the transmission T does not form a good basis for compensator design, but the sensitivity S does not share its zeros with the open loop transfer function, so could be used. We see that it determines the response to disturbances originating within the loop. These may arise not only from physical disturbances such as wind gusts or vibration, but when considering an ensemble of control loops, they may represent uncertainty in the plant parameters. In considering the sensitivity, we can not only synthesise controllers, we can also design them to achieve robustness to specified uncertainties in plant parameter values. 24.4.4 Weighted Sensitivity We specify the closed loop behaviour as a weighting function (wp), which is the reciprocal of the desired sensitivity, expressed as a function of frequency. (The subscript p denotes performance, as opposed to robustness). We would choose a function which reflects the requirements for sensitivity. The gain should be low at low frequency, which implies large feedback gain. However, 395 Cybernetics for Systems Engineers – DRAFT F this cannot be maintained in the region of gain crossover (the frequency which characterises the open loop bandwidth), without risking instability or control saturation. A typical weighting might be: 𝑤𝑃 = 𝑠⁄ +𝜔 𝐵 𝑀 𝑠+𝐴𝜔𝐵 Where ωB is the desired bandwidth, M is the high frequency gain, A the low frequency gain and the integrator is included in the denominator because we suspect integral of error is probably needed for accurate tracking. The need for integral control cannot be derivable from methods based on zero mean noise processes. If we are concerned with high frequency cut-off, we could try: 𝑁 𝑠 ⁄ 1 + 𝜔𝐵 𝑀𝑁 𝑤𝑃 = ( 1 ) 𝑠+𝐴𝑁 𝜔𝐵 To increase the slope below the specified bandwidth to 20N dB per decade. The problem reduces to finding the compensator K, for which the H∞ norm of the expression: |𝑤𝑝 𝑆| Is less than 1. The method is, more or less, directly applicable to the MIMO case, only dealing with a sensitivity matrix: 𝑆 = (𝐼 + 𝐺𝐾)−1 Rather than the magnitude, the maximum singular value is used. The sensitivity, at large values of gain is determined by the smallest singular value of the plant. If this is small, the control input will have little effect on the output, and there will be a risk of saturation, because high feedback gains will be needed. With scaled inputs and outputs, we need the lowest singular value of the plant to be greater than unity, otherwise it is unlikely to be controllable. 24.5 Stacked Requirements The sensitivity may yield controllers with phase advance, but there is no control over response above the specified bandwidth. In particular, there is no control over the high frequency cut-off needed to suppress measurement noise, nor is there any attempt to limit the control input. A more satisfactory approach would seek to bound the sensitivity, the control effort and the tracking error by using a matrix of requirements: 𝜔𝑃 𝑆 𝑁 = [ 𝜔𝑇 𝑇 ] 𝜔𝑢 𝐾𝑆 The problem then reduces to finding a transfer matrix K which minimises the H∞ norm (the highest peak of the largest singular value). The functions ωT and ωu characterise the desired closed loop response and control input respectively. 396 Cybernetics for Systems Engineers – DRAFT F 24.5.1 Robustness The use of H∞ as a means of compensator synthesis, on the face of it, does not appear to offer any advantage over linear-quadratic methods. It’s advantage lies in the fact that it can be adapted to account for plant uncertainty, which may be included in the compensator design. The uncertainties impose constraints on the singular values of both sensitivity and loop transmission, in many respects reminiscent of the Gershgorin bands encountered in diagonally dominant systems. A simple introduction to this subject would probably be misleading, and in any case, the author does not consider himself competent to explain it in detail. However sufficient has been presented to indicate where and when the methods are appropriate, and that is all that is required of an introductory text. As we have introduced the idea that it is the effect of plant delays (phase) which is the principal reason for the error signal failing to reflect the true state error, the idea of designing systems without reference to phase, represents a major departure from fundamental principles. It is generally the additional phase lag associated with neglected high frequency dynamics, which is of principal concern. Gain errors arise mainly from uncertainties in the plant parameters which are actually present in the system model. The H∞ approach effectively assumes the worst possible phase is associated with each gain uncertainty, so is truly robust, if probably excessively conservative. 24.5.2 Phase in MIMO Systems The phase reference is the input, so that a multi-input system has multiple phase references. There are as many harmonic loci as there are potential loops. We might expect a system to have up to, say 3 inputs and 15 outputs, yielding 45 possible harmonic loci. All 15 outputs could potentially be associated with each input, so we should expect three plots of 15 loci, from which the worst gain and phase margin associated with each input may be determined. The computation needed to generate the full population of potential loops is trivial by modern standards. However, the harmonic loci can only be produced once all the loops have been closed, and each loop for each input cut one at a time to see the consequences of additional phase loss (delays) in each feedback path. The phase margins may be found, but some other means is needed to generate the candidate design in the first place. Ultimately, one wonders whether it is a good idea to agglomerate the system description into a single matrix at all. It is the information flows between components which is of greatest interest. We are merely hiding the connectivity information within a single mathematical entity. This was highly successful in the time domain with state space methods, where the entire information content of the matrix is contained in its eigenvalue decomposition. The frequency domain description, although mathematically valid, does not appear to benefit as much from the matrix description. If anything it obscures, rather than clarifies the problem. We are probably better off sticking to scalar representations of the individual components interconnected in a network, so that it becomes possible to decide the phase lag associated with each path through the network, from which the margins may be deduced. 397 Cybernetics for Systems Engineers – DRAFT F Indeed, the classical control engineer (and the general systems engineer) is much happier dealing with networks of various kinds to characterise their systems, than with a matrix algebra description which completely omits the most important signal characteristic. We actually need to know how each individual input affects each output, so it is unwise to try and deal with inputs and outputs as vector quantities. This is the background understanding which underpins any attempt to automate the system. Singular values appear just the best way of describing an inherently poor system representation. 24.6 Concluding Comment Automatic control is an engineering discipline, not a pure science, and research budgets need to be predicated on the potential of the methods produced, not on awards which academics may give each other. For many years now, academia has seriously neglected its paymaster, yet continues to expect support for producing little which is presented in a form which line engineers can use. Despite the fact that this approach has been known for at least 30 years, the literature on the subject of MIMO control remains in a state of flux. This, in the author’s opinion, arises from the absence of practitioners in the loop to determine what is actually useful. The best textbooks are written by those academics whose expertise has been required in the unforgiving world of commercial engineering, although even here the target audience appears to be graduate mathematicians, rather than engineers. The commonest objective appears to be to make graduates aware of engineering, rather than helping engineers to understand the method. For these reasons, only a very sketchy outline of the method is presented, which it is hoped is sufficient to enable the reader to form an informed opinion as to whether the methods are appropriate to his/her task. Its application requires a deeper understanding than is possible to present in a single chapter. However, there are standard MATLAB algorithms available, which can be handle-cranked by the ignorant, and it must be admitted, they generally produce very good controllers, with nobody knowing exactly why. From the perspective of requisite entropy, MIMO systems introduce the additional problem that the inputs and outputs do not act in the same direction, so that one ceases to be an estimate of the other. This is in addition to the effect of delay. Many systems may appear decoupled at low frequency, but start to encounter cross-coupling as we try to increase the bandwidth. The input and output directions obtained from singular value decomposition help us to detect the onset of this particular limitation. When we ask too much of a SISO system we encounter instability. MIMO systems are far less forgiving, and we have the potential to screw up big time. On the positive side, these are the only methods which truly apply to MIMO systems, and address the issue of robustness which was overlooked in ‘optimal’ control. Also, the approach to robustness is inherently superior, and actually more in keeping with traditional engineering thinking. Nobody outside the control community has any idea what a gain margin or phase margin is, and indeed these terms are all but meaningless in the MIMO case. It must be better to base our stability margins on explicit awareness of the uncertainty in the plant. Admittedly, plots of singular values are probably even more obscure, but it should be possible to characterise parameter errors as 398 Cybernetics for Systems Engineers – DRAFT F statistical distributions and quote the stability and performance margins as probabilities that the uncertainty in the plant will lead to instability. The H∞ approach originated in the 1980s, when computer processing power was severely limited. Perhaps, with modern computing power, explicit Monte-Carlo methods could be applied to the characteristic equation, and appropriate distributions fitted to the resulting closed loop pole positions on the Argand diagram. It makes sense to specify a probability that the plant will misbehave, because this is what the wider community wishes to know. 399 Cybernetics for Systems Engineers – DRAFT F 25 Catastrophe and Chaos 25.1 Introduction This, and the following chapter, present material which is peripheral to the control engineer’s stock in trade, and will only be of interest to readers having a more philosophical interest in system science. It has already been mentioned that linear systems are actually quite rare. In cybernetics, we are fully aware that our equations are wrong. That is not the issue. The methods are concerned with finding how wrong we can be before our linearised analysis fails us. This boils down to how much performance we can get from our system before its behaviour becomes unpredictable. Feedback introduces inherent robustness, such that any interaction with an open set of disturbances must use feedback in some form. Compared with open loop control, feedback substantially reduces the accuracy requirements and cost of system components, and endows the system with substantial disturbance rejection, where the open loop has none. As part of this smokescreen, a considerable amount of pretentious nonsense has been written on the subjects of catastrophe theory and chaos theory, as alternatives to linear system theory. In fact both these areas of study actually complement linear theory. As in linear system theory, we seek to characterise and explain extremely complex behaviours by means of simple expressions. To those not versed in basic cybernetics, the idea that simple systems can exhibit complex behaviours appears anomalous. However, if we are familiar with the Principle of Requisite Entropy, it comes as no great surprise. Indeed, it is expected. 25.2 Catastrophe Both catastrophe theory and chaos theory deal with non-linear systems. Taking small perturbations about any single condition we see that the characteristic equation can have positive roots at some values of the state variables. In catastrophe theory, the system moves quickly to values of the state variable which result in negative roots for small perturbations around the new position. The system starts at one stable equilibrium point and a coefficient is varied until this equilibrium ceases to be stable, so that the system moves to a new stable equilibrium condition. This sudden transition between two stable equilibria is known as a ‘catastrophe’ because the interval the transition takes place in is very short, compared with the rate the controlling coefficient is varied. Catastrophe theory deals only in static equilibria, without reference to the nature of the transition between them, or whether the system will actually settle down at all. When the dynamics of the transitions are of interest, we find ourselves in the field of chaos. The idea is to characterise the state of the system as a potential energy which is a function of position coordinate (x). If a perturbation in x increases the potential energy, it represents a stable equilibrium, otherwise it is neutral or unstable. This was actually the idea behind Liapunov’s direct method, but in general, the ‘potential energy’ analogy is not very helpful. 400 Cybernetics for Systems Engineers – DRAFT F 25.3 The ‘Fold’ The simplest illustration of this is called the ‘fold’ catastrophe. It serves to illustrate the idea. But has little practical value. It has the ‘potential’ function: 𝑉 = 𝑥 3 + 𝑎𝑥 Where ‘a’ is the controlling parameter. This may be thought of as a surface in 3D space, each vertical cross section of which is a cubic curve. The roots of a cubic equation may be found by means of the following algorithm. Given: 𝑥 3 + 𝑎2 𝑥 2 + 𝑎1 𝑥 + 𝑎0 = 0 The nature of the roots can be determined by first evaluating: 𝑞= 𝑎1 3 − 𝑎22 9 𝑎3 1 𝑟 = 6 (𝑎1 𝑎2 − 3𝑎0 ) − 272 If 𝑞3 + 𝑟2 < 0 All roots are real, otherwise the curve has one real root and a pair of complex conjugate roots. The second case has no turning points, so there can be no equilibia. The first case has one peak (unstable equilibrium) and one trough (stable equilibrium). In this case a0=-V, a1=a and a2=0. 𝑎 𝑞 = 3, 𝑟 = 𝑉 2 The condition for three real roots is: 𝑎3 < − 27 2 𝑉 4 This requires a to be negative. As a is varied from an arbitrary large negative value, the system remains in the equilibrium position corresponding to the trough, until as a approaches 0, the equilibrium disappears. It has been argued that this could represent irreversible catastrophic failure. The point where the control variable a approaches zero is called the ‘tipping point’. The stable equilibrium line is: 𝑎 3 𝑥 = −√ 401 Cybernetics for Systems Engineers – DRAFT F 25.4 The ‘Cusp’ The most commonly used catastrophe model is the cusp. The author is dubious about representing 3D objects pictorially on 2D paper, because this introduces ambiguity. The idea of trying to represent higher dimensional objects in 2D really does cross the threshold of self-deception. Instead of representing the potential hypersurface ( i.e. 4D plot of V), we draw a 3D contour representing the set of potential equilibrium positions. The pictorial representation of the cusp catastrophe is usually the derivative of V with respect to the system variable x, plotted against two control variables (a and b). The potential function is: 𝑉 = 𝑥 4 + 𝑎𝑥 2 + 𝑏𝑥 The equilibrium surface is given by: 𝑑𝑉 𝑑𝑥 = 0 = 4𝑥 3 + 2𝑎𝑥 + 𝑏 This is similar in shape to the potential surface of the fold catastrophe. Figure 25-1 : Equilibrium Surface for the Cusp Catastrophe The fold in the equilibrium surface is called a cusp. Nothing interesting happens if ‘a’ is positive, as this corresponds to an equilibrium position beyond the apex of the fold. 402 Cybernetics for Systems Engineers – DRAFT F With negative ‘a’, increasing b from a negative value as far as the cusp doesn’t raise any problem. However, the only equilibrium position beyond this point is on the other side of the cusp. The state variable changes suddenly to accommodate this transition. Reducing the value of b does not cause an immediate transition back, but the system moves to the edge of the cusp before transitioning back. This effect is called hysteresis. Rather than characterising a truly catastrophic event, the cusp models the sudden transition between two levels in the state variable. It behaves as a binary switch with hysteresis. 25.5 Comment There are several catastrophe models, involving more control variables and states, but these two illustrate the principle. Since it was first formulated, catastrophe theory has acquired some strange bedfellows, and is often central to some pieces of pseudo-scientific nonsense. However, there are circumstances where it can be useful. The author has used it personally to explain the behaviour of a command to line of sight missile in multipath. In that particular case, the missile trajectory appeared to transition spontaneously to different nulls in the tracking radar polar diagram. The attraction of catastrophe theory is its ability to combine continuous signals with the transitions of a finite state machine, without using explicit logic. It has been mentioned in earlier chapters that finite state machines cannot deal with an open set of disturbances, and the Turing machine cannot adapt by random mutation. It appears that catastrophe theory might furnish a means of circumventing these limitations. In principle, the control variables of a catastrophic system could be adapted by feedback, and an organism which employs catastrophe, as opposed to Boolean logic, appears to have the potential to evolve by natural selection. The weakness of the Turing machine is its inherent sequential processing, which is like a chain; if one link breaks, so does the chain. If processing is done in parallel, with relatively short sequences, it may be possible to devise a processing machine which can adapt as an error-driven regulator. This seems a reasonable speculation in view of the incredible feats of processing achieved by nervous systems whose elements trigger in milliseconds, rather than the nanoseconds of electronic devices. It is difficult to see how this could be achieved with such slow elements without massive parallelism. We can imagine the processing power of a system which has both electronic switching times and biological parallelism. If catastrophe is used as the process of adapting a finite state machine process, it must follow that the finite state machine is subordinate to one employing continuous signals. This is at variance with current wisdom which appears to believe that information should become more quantised, the higher up the system hierarchy it is used. The fact that we don’t see finite state machines in nature, should raise concerns regarding the modern perception that they furnish a universal representation of system dynamics. 403 Cybernetics for Systems Engineers – DRAFT F 25.6 Chaos 25.6.1 Adding Some Dynamics We can treat the cusp catastrophe as a dynamic system by adding a kinetic energy term to the potential energy: 1 𝑉 = 2 (𝑥̇ )2 + 𝑥 4 + 𝑎𝑥 2 + 𝑏𝑥 Differentiating: 𝑥̈ 𝑥̇ = −(4𝑥 3 + 2𝑎𝑥 + 𝑏)𝑥̇ i.e: 𝑥̈ = −(4𝑥 3 + 2𝑎𝑥 + 𝑏) Everywhere on the equilibrium surface 𝑥̈ = 0. We shall denote the equilibrium value of the state on this surface as X, and use the lower case x to denote a perturbation with respect to equilibrium. 𝑥̈ = −4(𝑋 3 + 3𝑋 2 𝑥 ⋯ ) − 2𝑎(𝑋 + 𝑥) − 𝑏 Equilibrium requires: 4𝑋 3 + 2𝑎𝑋 + 𝑏 = 0 The small perturbation equation is: 𝑥̈ = −(12𝑋 2 + 2𝑎)𝑥 We can write this in terms of the control variable b, by solving the equilibrium equation for X. An analytical solution only exists for the case of a single real root, but we are interested in the region of the cusp, where there are three real roots. Instead, we shall calculate b for a range of values of X, resulting in the following plot of ‘stiffness’ as a function of b: Figure 25-2 : Stable and Unstable Regions for the Cusp Catastrophe 404 Cybernetics for Systems Engineers – DRAFT F The small perturbation equation is of the form: 𝑥̈ = −𝑐𝑥 If c is positive, the system is neutrally stable, and the state will oscillate about the equilibrium condition indefinitely. If negative it is unstable, and the system will diverge from the nominal equilibrium point. The unstable region corresponds to the region between the two turning points in the equilibrium surface. The resulting system will not actually settle down as predicted by catastrophe theory unless we introduce a damping term. 𝑥̈ = −𝑑𝑥̇ − 𝑐𝑥 Where the coefficient d characterises the damping. With this modification, the system should behave dynamically as predicted by catastrophe theory. For optimal damping, we shall choose: 𝑑 = 1.4√𝑐 i.e. a damping ratio of 0.707. We shall illustrate this by starting the system on the cusp with X=0.4, and plotting the transition to the stable equilibrium position at X=-0.8. Figure 25-3 : Transition from Unstable to Stable Equilibria The transition is illustrated by the above phase plane plot. Note that the system starts at the unstable equilibrium point on the right hand side of the plot. The motion is determined by the nature of the stability at the start and end points, and most descriptions of non-linear system call these points ‘attractors’. 405 Cybernetics for Systems Engineers – DRAFT F 25.6.2 Attractors Stability of non-linear systems cannot be quantified by a simple criterion. There is no simple extrapolation of the right half plane or unit circle criteria to non-linear systems. We can only talk about stability of perturbations about nominal equilibrium positions. The potential equilibrium positions are calculable from the steady state solutions of the non-linear equations, so our restriction to examining the perturbation of states around the various equilibria appears to cover all eventualities. In system theory, we deal with potential equilibria, i.e. the equilibrium positions which are determined by setting all time derivatives to zero. Catastrophe theory, and a lot of quasi-scientific theories, appear to assume that these equilibria are actually reached, and indeed characterise the problem. We have seen, however, that equilibrium is only reached and maintained for a particular set of system parameters. These potential equilibrium points are called ‘attractors’, to indicate that although there might be a tendency for the system to move towards them, usually they never get there. Linear theory yields three types of attractor; stable, neutral and unstable, as shown in the following figures. 406 Cybernetics for Systems Engineers – DRAFT F 25.7 The Lorenz Equation Linear theory indicates that there are three types of attractor, the system may converge on a stable equilibrium point and stay there, the system may cycle around the nominal equilibrium indefinitely, 407 Cybernetics for Systems Engineers – DRAFT F or the system may diverge away from the nominal equilibrium position. The reason for this is the fact that a characteristic equation can be factored only into simple and quadratic terms, and the only behaviours the corresponding modes can exhibit correspond to the three basic types of attractor. Consider now a system governed by the non-linear equations: 𝑥̇ = 𝜎(𝑦 − 𝑥) 𝑦̇ = 𝜌𝑥 − 𝑦 − 𝑥𝑧 𝑧̇ = −𝛽𝑧 + 𝑥𝑦 Our analysis starts by linearising about a reference point X,Y,Z: 𝑥̇ = 𝜎(𝑦 − 𝑥) 𝑦̇ = (𝜌 − 𝑍)𝑥 − 𝑦 − 𝑋𝑧 𝑧̇ = 𝑌𝑥 + 𝑋𝑦 − 𝛽𝑧 The system matrix is: −𝜎 (𝜌 𝐴 = ( − 𝑍) 𝑌 𝜎 0 −1 −𝑋) 𝑋 −𝛽 From which, the characteristic equation is: 𝑠 3 + (1 + 𝜎 + 𝛽)𝑠 2 + [𝜎(1 + 𝑍 − 𝜌 + 𝛽) + 𝛽 + 𝑋 2 ]𝑠 + 𝜎[𝛽 + 𝑋 2 + 𝑍𝛽 − 𝜌𝛽 + 𝑋𝑌] = 0 The Hurwitz test matrix is: (1 + 𝜎 + 𝛽) 𝜎[𝛽 + 𝑋 2 + 𝑍𝛽 − 𝜌𝛽 + 𝑋𝑌] 0 2 [𝜎(1 + 𝑍 − 𝜌 + 𝛽) + 𝛽 + 𝑋 ] 𝐻=( ) 1 0 (1 + 𝜎 + 𝛽) 0 𝜎[𝛽 + 𝑋 2 + 𝑍𝛽 − 𝜌𝛽 + 𝑋𝑌] The first test function is: 𝑇1 = 1 + 𝜎 + 𝛽 This depends only on the system parameters. The second and third test functions are polynomials in X, Y and Z. Since these can adopt arbitrary values, it is evident that there will be regions of state space where the system will be unstable. From a control perspective, that is the end of the matter. We detect the unstable regions and develop controllers to stabilise them. Furthermore, we consider the constraints on the size of perturbations and system bandwidth which arise from the non-linearities, in order to ensure robustness. 408 Cybernetics for Systems Engineers – DRAFT F However, the Lorenz equation was not formulated for a system over which control could be feasibly exercised. It was a simple model of the weather. Linear analysis indicates that the solution would just diverge. However, usually we want to know a bit more about the weather. We know that weather is not a divergent phenomenon, although the long term effects of global warming may force us to revise this perception. If anything it appears cyclic, but not predictable by simple harmonic models. The cycles appear aperiodic. 25.7.1 Explicit Solution Solving the Lorenz equation for σ=10, ρ=27 and β=2.666 results in the somewhat surprising plot of figures Figure 25-4, Figure 25-5 and Figure 25-6. Rather than a divergence, the system appears to orbit around a pair of attractors, but doesn’t conduct periodic orbits about either. Evidently the unstable mode near one attractor is a different mode to the one which is unstable about the other. The system appears both unstable and stable at the same time. The result from linear theory is therefore pessimistic, because the system does not simply diverge to infinity, as predicted. Figure 25-4 : Lorenz Equation Solution: x-y Plane Figure 25-5: Lorenz Equation Solution : x-z Plane 409 Cybernetics for Systems Engineers – DRAFT F Figure 25-6 : Lorenz Equation Solution : y-z Plane The neither stable nor unstable attractor has become known as a ‘strange’ attractor. It indicates that non linear systems are not necessarily unstable in certain regions, but many exhibit this ‘strange’ behaviour. Unlike artificial systems which must be designed for stability and robustness of stability, naturally occurring systems can and do evolve well into this region of aperiodic, unpredictable behaviour. In fact, linear dynamic systems are such a rarity outside the field of artificial systems, that linear theory is all but useless for studying naturally occurring dynamic phenomena. Actually, it is difficult to see how life could have evolved without chaos to stir the soup. However, if it is within our power to control a system, it would be the height of folly to allow it to operate in the unpredictable chaotic regime. Yet, this is precisely how human economies are managed. It is a case of accepting what we cannot control, but controlling everything we can. Cybernetics furnishes us with the wisdom to tell the difference. 25.8 The Logistics Equation The Lorenz equation is most frequently cited because it was allegedly the first to draw our attention to the phenomenon of chaos. Doubtless it had been encountered earlier, but rather than gaining the laurels of a new discovery, the programmer responsible was probably dismissed for incompetence. It really isn’t a surprise to discover the weather is difficult to predict. That was established many centuries ago. The surprise is that the unpredictability can be demonstrated with a deterministic model. The Lorenz equation requires three parameters, and is probably not the best option for studying the basics of chaos. A simpler model is one used to represent predator/prey interactions or population growth. We sample the population at equal intervals. If the population last year was x, we expect it to increase this year by an amount kx: 𝑥𝑛+1 = 𝑥𝑛 + 𝑘𝑥𝑛 410 Cybernetics for Systems Engineers – DRAFT F But this is a formula for exponential growth. We should expect a regulatory influence to exist, such as predation, overcrowding or starvation. A certain proportion of last years’ population will not survive to breed. We assume the proportion reduces the higher the population, so we have: 𝑥𝑛+1 = 𝛼(𝑥𝑛 + 𝑘𝑥𝑛 )(𝛽 − 𝑥𝑛 ) Where β is the proportion which can breed when the population is very small (αβ=1). Defining a new variable: 𝑦= 𝑥 𝛽 The equation for population growth becomes: 𝑦𝑛+1 = 𝛾𝑦𝑛 (1 − 𝑦𝑛 ) Where γ is a constant. The steady state is found by setting yn+1=yn.: 𝑦∞ = 1 − 1 𝛾 The linearised equation about a population Y becomes: 𝑦𝑛+1 = 𝛾𝑦𝑛 − 2𝛾𝑌𝑦𝑛 Taking the z transform: 𝑧 − (1 − 2𝑌)𝛾 = 0 Stability requires z to lie within the unit circle: |𝛾 − 2𝑌𝛾| < 1 Or: 𝛾 2 (1 − 2𝑌)2 < 1 Substituting Y=y∞: 𝛾<3 If the population were to breed faster, the equilibrium would not be stable. In practice, real populations move well into the unstable region, and oscillate between short-lived equilibria , or collapse altogether for no apparent reason. A naive analysis would delude us into thinking that the equilibrium population is given by y∞=1, whilst it is evident that we do not dare exceed y∞=2/3. 411 Cybernetics for Systems Engineers – DRAFT F 25.9 Bluffer’s Guide The original plan had been to expand on the logistics equation, to cover the basic methods used in chaos theory. I have decided against this, partly because I am not best qualified to cover the subject with any authority, but primarily because it is not really relevant to the subject of cybernetics. Chaos, like signal processing, is a close relative of cybernetics, and despite appearances (they all use mathematical formulation), the subjects are different. They have different objectives. However, a lot of chaos ‘buzz words’ have crept into the systems science vocabulary, so I shall limit the discussion to definitions of these terms. 25.9.1 Basin of Attraction If we move beyond the linear stability region, the system behaviour may become characterised by two or more attractors. A basin of attraction is associated with each of these attractors, and consists of the set of all starting points which end up converging towards that particular attractor. If we are trying to control a system, the basin of attraction should be the whole of state space. Frequently, the basins of attraction are fractals, and this is one reason for the sensitivity of the system to small changes in initial conditions. They also make cool T-shirts. 25.9.2 Bifurcation If we solve the equation of motion for different values of the control parameter, we expect the long term solution to converge on a single value, so a plot of final value against parameter is expected to be a monotonic curve. Beyond the stability region, for a fixed end time, the behaviour is expected to be oscillatory, and the plot splits into two branches. Increasing the control parameter still further, causes a doubling of the frequency of oscillation and each branch of the plot again splits into two branches. 25.9.3 Cobweb Plot For a sampled data system, such as the logistics equation, we plot the current value of the state against its previous value. The equation is of the form: 𝑥𝑛+1 = 𝑓(𝑥𝑛 ) Starting at point x0, a line is drawn vertically to intersect f(x). This is projected horizontally on to the line y=x, and vertically again on to f(x). This is repeated until the solution converges or the oscillation can be characterised. 25.9.4 Feigenbaum Number The bifurcation of the plot of final value against control parameter takes place at a definite ratio of the changes in the control parameter between successive bifurcations. As the number of bifurcations approaches infinity, this ratio approaches the value 4.669. This is independent of the actual form of the non-linear function. 25.9.5 Liapunov Exponent A Liapunov exponent characterises the divergence of solutions of a chaotic system. It is effectively the time constant of an exponential fit to the difference between the two solutions as time evolves. It characterises the maximum time span the model may be used over to predict future events. 412 Cybernetics for Systems Engineers – DRAFT F 25.9.6 Poincaré Section A system characterised by continuous rather than sampled-data dynamics may be investigated for features such as bifurcation and symmetry by converting it into a sampled data system using a sampling frequency which in some sense characterises the motion. (e,g, the period of stable oscillations before the system is driven into the unstable regime). 25.9.7 Symmetry Breaking In addition to bifurcation, operating the system beyond the stable region frequently introduces an asymmetric motion, even though the equations exhibit symmetry. 25.10 Concluding Comment Whatever else can be said about chaotic systems, from a cybernetics perspective they are bad news. It appears however, that they are detectable from the stability characteristics of the linearised systems. To attempt to control a non-linear system explicitly is to miss the point. We have a sound body of knowledge on which to base our control systems , central to which is a means of deciding whether such control is possible, and what its limitations are. The word ‘cybernetics’ comes from the Greek word for ‘helmsman’. It is about taking control of systems, rather than characterising the unfolding disaster of the careering train. As technology advances, the ability of mankind to acquire mastery over the environment ought to improve. The current ethos of cowering in the rear of the cave, too afraid to intervene, and allowing natural phenomena regularly to slaughter entire populations, simply isn’t acceptable. Managing fish stocks by simply stopping fishing may not address the problem of overfishing. Perhaps the elimination of a natural predator could have pushed the dynamics into the chaotic region. As long as we base our ideas on the naive direct cause-and-effect, we are going to be surprised at the results. In many ways the setting of prices in a ‘free’ market resembles the predator/prey system, and we should expect similar dynamic behaviour. The belief that equilibrium prices set using static ideas are stable, seems a very dangerous assumption. Much of human history indicates that they are nothing of the sort. Systems are not like cars, which we simply drive, they are like donkeys; if we push them too hard, they simply turn around and kick. 413 Cybernetics for Systems Engineers – DRAFT F 26 Artificial Intelligence and Cybernetics 26.1 Introduction This chapter is peripheral to the mathematically rigorous control theory covered in earlier chapters, and is presented for the sake of interest. The present author has had very little practical experience of artificial intelligence methods, so is not well placed to offer anything but a rudimentary summary of the established techniques. There is a conceptual problem, in that AI defines intelligence in terms of specific mental skills which humans exhibit, and which the discipline of AI seeks to emulate mechanically. I have already explained my objection to defining a quantity by specifying the manner in which it is implemented, rather than by its actual measurable effect. Indeed, the author tends to view the relationship between AI and automatic control as akin to that between astrology and astronomy. The former is needed to attract funds from a technically naïve establishment, whilst the latter undertakes real science. But that is a particularly harsh personal view. The reader is encouraged to acquire the facts and form his or her own opinion. It has been mentioned that the specific mental skills of human beings may be irrelevant to, for example a Blue Whale, which according to the principle of requisite entropy, would almost certainly require a different set of mental skills to survive in its environment. There is something chauvinistic and parochial about seeking to emulate human intelligence, as opposed to finding out what intelligence actually is. The following describes techniques which are established, or have been tried, with varying degrees of success. 26.2 Intelligent Knowledge Based Systems Intelligent knowledge-based systems or ‘expert systems’ (note the tendency to hyperbole of the AI world), are essentially databases with an automated search process attached. Knowledge differs from data in that it consists of inter-relationships between facts, rather than just the facts themselves. Recall the estate agent’s database of chapter 19, the estate agent looks for values for each of the attributes describing the properties which match the customer’s requirements. Much the same effect could be achieved by presenting the customer with a sequence of questions, the answers to which would trigger ‘if…then’ rules to present the next question. However, an expert system can use the relationships between its rules to find the questions which give rise to particular results. This gives it the ability to backtrack, if a sequence fails to reach the desired result. This feature is essential if the expert system is to be used within a closed loop. We can imagine that the attributes are descriptors of the state of the system, and a particular combination of attributes triggers an activity which the system performs on the environment, which 414 Cybernetics for Systems Engineers – DRAFT F hopefully will change the state to a desirable one. If the actual state is not as desired, the IKBS must be able to select an alternative. This conceptually yields a measure of how near different states are in terms of the number of rules that must be re-tried before the correct course of action is chosen. We might be able to organise the IKBS in hierarchical form, with highly quantised states at the highest levels corresponding to initial stages in a sequential set of operations. From a cybernetic perspective, we are not concerned with how to implement an IKBS. Our concern is with characterizing its dynamic range and time response, which are essential to understand its inclusion within a broader system. Needless to say, expert systems are never characterized in this way. The process is intelligent, in that it reduces entropy, it is knowledge-based in the sense that it deals with relationships between data items, but a ‘system’ it is not. Computer Science tends to call data ‘knowledge’, and knowledge (i.e. the organisation of data) as meta-knowledge. Since it is impossible to tell from a single sample whether it constitutes signal or noise, this appears an unwarranted promotion for what could be, and frequently is, meaningless nonsense. 26.3 Fuzzy Logic Expert systems have found widespread applications in diagnostics tools, where the user is not only presented with possible solutions, but also the line of reasoning which led to the result. They deal with highly-quantised, multi-dimensional information typically expressed as data fields, each of which contains a limited number of options. An expert system is conceptually a finite state machine, and as such is extremely difficult to design to adapt, as is required by requisite entropy. Although feedback might be applied as a random selection, if there are a large number of stages in a sequence of operations, the adaptation time is likely to be astronomic. A genetic algorithm, for example could be used to devise sequences of operations, by combining those which achieve a result closest to the objective. This raises the question as to what we mean by ‘close’. If the problem is one of guessing the winning lottery number, every possible choice is equally ‘close’ to the winner as all are potential winners. As Ross Ashby pointed out, Nature is not that perverse as to present us with a lottery each time we need to adapt. However, it is evident that if we can somehow characterise the difference between the desired outcome and the actual outcome by some metric, we should be in a better position to decide on a change in behaviour which will correct the error. In order to overcome the inherent brittleness of sequential processing and rigidly defined sets of data items, the idea of a fuzzy set was introduced in the 1970s. With fuzzy sets we try to accommodate the frequently occurring problem that real world entities simply don’t fit neatly into the artificial categories specified by human beings. We considered a few examples in the estate agents case. Rather than define an entity as a member of one set, fuzzy 415 Cybernetics for Systems Engineers – DRAFT F reasoning allows us to assign it a degree of membership in more than one, possibly all of the categories we choose to define. What we are trying to represent is the vagueness associated with our artificial categories. It must be emphasised that we are not trying to define the probability that a specific entity will find itself in one set or another. Probability could be used, but we would need to consider conditional probabilities and apply Bayes’ Theorem. This confusion of degree of set membership with probability has given fuzzy logic a bad name amongst statisticians. We define a set membership value by first defining the attributes an entity must have to be the quintessential member of a particular category. The membership value is the degree to which the current entity resembles the definitive member of each category. In practice, this is highly subjective, but is a way of capturing human perception of similarities and differences between objects. Indeed it is in mechanising tasks undertaken by human beings where fuzzy sets are claimed to have some merit over mathematical analysis of the system dynamics. Suppose we are heating a room with a heater whose output is variable from zero to a defined maximum power. We can define the heater response as high or low according to the degree of membership of the ‘low’ or high fuzzy sets of heater settings. Similarly we can describe the room temperature in terms of the degree of membership of the sets ‘cold’, ‘moderate’ and hot. 416 Cybernetics for Systems Engineers – DRAFT F The control of the room temperature takes the form of a set of fuzzy rules: ‘if the room is hot set the heating low’ ‘if the room is cold, set the heating high’ ‘if the temperature is moderate set the heating low’ This differs from conventional logic rules in that they are applied simultaneously, rather than sequentially, and the output of each is the set membership value of the associated state of the room. There are many ways of converting the result obtained from the rules into a heater setting. The simplest is to apply the rule which has the highest output and apply the value to the appropriate heater set. The heater setting is then the value corresponding to the value of membership function of the associated set. Whilst this overcomes the limitations of crisp sets, the plant behaviour is ignored, so that stability or robustness are matters of faith, rather than demonstrable. However, in mechanising activities undertaken by human beings we may be reasonably confident that the processes are likely to be stable. If nothing else, the fuzzy sets approach illustrates the liberties which can be taken with feedback control. 26.4 Neural Nets Fuzzy sets and expert systems are essential ly techniques for capturing human knowledge and skills and implementing them mechanically. Neural nets were introduced as an attempt to extend the emulation to the level of nerve cells. 417 Cybernetics for Systems Engineers – DRAFT F Each ‘neuron’ applies weightings to a set of inputs which are then combined and fed through a nonlinear function. Usually large numbers of these elements are arranged, usually in three layers denoted input, hidden and output layers. The weightings are determined by training the neural net with a number of examples representative of the task which the neural net is intended to perform. Whilst considerable success has been claimed, and some actually achieved, there is nothing to prove that a neural net will work for anything outside its training set. It is impossible to prove that it will without some indication as to how the neural net is actually performing its task. 26.5 Concluding Comment The AI community, like the Computer Science community, seem to set their system boundaries around their particular application, which they then proceed to call a ‘system’. This is a somewhat parochial view, yet it appears to dominate modern systems thinking. In order to use these elements within a cybernetic system they need to be characterised in some way in terms of their time evolution and errors, otherwise they can only be employed as stand-alone applications. Rather than produce homomorphs of behaviour, which are actually useful for proper systems analysis, we have no option but to implement the actual software explicitly. In order to use a component which employs AI techniques within a system, it is necessary to characterise it as a system element, in terms of its time evolution, input/output relationships and dynamic range. This requires insight and understanding beyond the current ‘suck it and see’, approaches which seem prevalent. The tendency to design controllers without reference to plant delay is a distinctly worrying development. The introduction of non-linearity into the system is not a virtue, but a vice, for the reasons which have been presented repeatedly in the earlier chapters of this book. It is bad enough for the plant to be non-linear, without compounding the problem by making the controller non-linear as well, particularly as there is no attempt to characterise the bounds of validity of the method employed. 418 Cybernetics for Systems Engineers – DRAFT F 27 Conclusions 27.1 Introduction I have tried to introduce ideas in order of importance, starting with the fundamental principle of cybernetics : the Principle of Requisite Entropy, from which the necessity for feedback is derived. Even in an apparently open loop controller, the parameters are determined by an outer loop involving the acquisition of the knowledge of system behaviour, which itself followed an iterative process. This process might be the general conjecture/experiment cycle of science, culminating in knowledge in the designer’s brain, or it may take the form of a higher level automatic process whose function is to acquire information on the system behaviour. For example, economic policy may be tested in a model, and the actions which appear to give the desired results are proposed for implementation. The loop is then closed by the actual response of the economy. Invariably, the predictions are wrong, so the model must be modified to take this into account. This approach indicates a belief that the system is in fact controllable, observable and stable, when all the indicators point to chaos. We are better off with a crude model of the dynamics of the system from which the systemic issue of stability may be addressed rather than concentrating on calculating the wrong answers to great precision. In general, we do not care how the error null is achieved, it may be implemented in ways which are limited only by ingenuity and imagination. Ross Ashby, and Nature use random selection. We can say nothing about the system behaviour until we have characterised its time evolution. When the system elements are represented as difference, or differential, equations, this is a direct mapping. More generally, a level of abstraction is required to identify the dynamic processes, and characterise them into differential (or difference) equations which can be linearised around the set of circumstances which interest us. At a minimum, we can represent the time evolution of a process as a pure delay. Typically, we should characterise the input and output in such a way that the duration of the delay may be related to the attributes of the signal. For example, we might reasonably assume that the greater the entropy change between input and output, the longer the processing will take. 27.2 Hierarchy The important point, which has been hinted at from time to time is that the state variables do not have the same meaning throughout the system . What are parameters to a lower level system may well be states of a higher level system. The functions are different in nature, to the extent that they cannot be lumped together in a single system matrix. We saw from the sensor scanning example that this does not matter, so long as we keep the processes response times widely separated. Indeed, our concerns for the effects of non linearity are ill-founded; we should never operate the higher level system any faster than between 1/3rd and 1/10th the speed of its slowest subordinate system. Also ,we have seen that our linear analysis will indicate whether we risk not only instability but chaos, and we can determine the bounds on the signals which will push the system into a chaotic regime. 419 Cybernetics for Systems Engineers – DRAFT F The dynamic model at any level in the system would typically treat lower level elements as infinitely stiff and higher level signals as constants. In view of the ratio of response times, a hierarchy consisting of elements of equal intelligence, must contain between a third and a tenth as many elements in each hierarchical layer as there are in the layer beneath it. A hierarchical pyramid emerges from purely cybernetic principles, based on an egalitarian assumption. In human hierarchies some myth of intrinsic superiority is usually invoked to justify the right to abuses of power characteristic of leaders. The fact that coercion and oppression tend to figure highly in establishing and maintaining this hierarchy, leads to doubts that our masters really are our ‘betters’. 27.3 Intelligence Despite appearances, we human beings aren’t really intelligent. Our brains are specialised to a particular set of intellectual activities which presumably were of benefit for survival, and most ‘intelligence’ tests are geared to measuring performance at those particular tasks. Those intellectual skills which lead to financial success figure highly in validation of the tests. In Western society, being white and middle class tends to lead to a higher probability of financial success in later life, ergo ‘intelligence’ is strongly genetic in origin. In order to handle life as we find it, we put together a set of myths, or heuristics, which we know are not true but assist in helping us understand the world in terms which are familiar. As long as we know we are dealing with convenient fictions, there is no problem. The difficulties begin when we treat them as truth. Science has sought , through the acid test of experimental verification, to distinguish speculation from fact. When formulating heuristics to describe the world, however, we can tolerate the odd counter example. This is legitimate, provided we don’t misrepresent the area of study as a ‘science’. Cybernetics lies above the level of this debate. We do not care how the models originate. Our definition of ‘intelligence’ is therefore independent of the specific intellectual feats of the human mind, but is in principle extendable to non-human, and mechanical minds. The more capable organism may well act out its options in an internal model before committing itself to action. We could ascribe some meaningless term, like ‘thinking’, or ‘imagination’ to this process, but to do so would miss the point. In defining a metric, we need to observe the phenomenon and quantify it, not dictate how it comes about. 27.4 Seeking Null In order to extend the linear theory analysis to higher levels in the hierarchy it is necessary to find a way of describing the data items in such a manner that it is possible to decide whether the achieved result is equal to the desired result, within the bounds of uncertainty characterising the process. If all the possible results can be defined as an enumerated type, all we have to do is search through the set of possible outcomes and compare them in turn with the actual outcome. For trivial systems with few alternatives, an exhaustive search is probably adequate. 420 Cybernetics for Systems Engineers – DRAFT F More generally, we need a faster search process. As an example, suppose our objective is to identify a target from a high resolution image. The target may be at any orientation with respect to the observer, and it may come from a large set of potential targets. Furthermore, it is not good enough simply to match the best fit to the known target set, sufficient knowledge of the noise should be available to return an ‘unidentified’ classification. A search matching all possible targets at all resolutions is probably prohibitive in the time available. We need some metric of similarity before we can determine how well a target resembles a template. The correlation integral introduced in Chapter 19 fulfils this requirement, but doesn’t help as far as run time is concerned. Supposing a rectangle is fitted to cover the extremities of the target in the image. The aspect ratio of this rectangle should immediately yield information about the target, but this will be ignored. We fit the image with a raster initially of 2×2 pixels, producing a sequence of four samples characterising the image. Only a proportion of the targets will match this low resolution image. We then repeat the process with a 4×4 pixel image. This is repeated until the full image resolution is achieved, rejecting all but the set of targets which ‘look like’ the actual target at each resolution level. In practice, the correlation would be done with a Fourier transform, with upper frequency bound set to the reciprocal of the pixel width. In essence, the image would be sampled at ever increasing spatial bandwidths. The algorithm could be arranged, such that the lower bandwidth results could be used to shorten the calculation of the higher bandwidth results. In characterising the targets in this way, we have a means of quantifying how similar two targets are. This information might itself be useful in generating an error signal to drive the system to its result. Producing metrics which characterise the data such that their values indicate how close together the corresponding data items are requires considerable ingenuity, but is central to establishing a cybernetic representation of the process from which performance and stability may be assessed. The alternative is the extremely expensive and unreliable process of explicitly constructing the system and testing it. 27.5 And Finally... There is a common misapprehension, particularly amongst the class of incompetent manager who view mathematics as a ‘specialisation’, that the methods of cybernetics apply only to the lowest levels of system operation. Hopefully, the material presented here will go some way to dispelling this myth and exposing it as the fig leaf for ignorance which it is. To quote Lord Kelvin: ‘When you cannot measure it, when you cannot express it in numbers, your knowledge is of a meagre and unsatisfactory kind; it may be the beginnings of knowledge, but you have scarcely, in your thoughts, advanced to the stage of science’ 421