BEYOND THE INDUSTRIAL WEB:

ECONOMIC SYNERGIES

AND

TARGETING METHODOLOGIES

 

 

 

 

BY

STEVEN M. RINALDI, MAJOR, USAF

 

 

A THESIS PRESENTED TO THE FACULTY OF

THE SCHOOL OF ADVANCED AIRPOWER STUDIES

FOR COMPLETION OF GRADUATION REQUIREMENTS

 

 

 

SCHOOL OF ADVANCED AIRPOWER STUDIES

AIR COMMAND AND STAFF COLLEGE

AIR UNIVERSITY

MAXWELL AFB, ALABAMA

JUNE 1994

DISCLAIMER

 

This publication was produced in the Department of Defense school environment in the interest of academic freedom and the advancement of national defense-related concepts. The views expressed in this publication are those of the author and do not reflect the official policy or position of the Department of Defense or the United States government.

 

ABOUT THE AUTHOR

 

Major Steven M. Rinaldi (BSEE, University of Oklahoma; MSEE, Air Force Institute of Technology; PhD, Air Force Institute of Technology) is a developmental engineer and currently a student at the School of Advanced Airpower Studies, Maxwell AFB, Alabama. Major Rinaldi’s initial assignment was to the Air Force Institute of Technology. After receiving his masters and doctoral degrees, he was assigned to the Air Force Weapons Laboratory, Kirtland AFB, New Mexico. He directed several advanced laser programs, and led the coupled laser device technology program. In 1989, he was selected for the Scientist and Engineer Exchange Program. As part of the program, he served two years as an exchange scientist at the Ecole Polytechnique, Palaiseau, France. During this tour, he performed research on laser light scattering and nonlinear optics.

 

ACKNOWLEDGMENTS

 

As with any major project, this study could not have been undertaken without the support and aid of many individuals. I would like to offer my sincere appreciation to Keith Almquist, William M. Arkin, Edwin L. Averill, Mark D. Beisemeyer, Gil Broyles, Jon Caswell, Max Crocker, Dennis Engi, Brad Godfrey, William F. Hartman, Kirk Hilbig, Doug Lawson, W. John Light, Bill Ling, Len Malczynski, Andy Martin, John Milloy, G. Alan Petzet, Stephen C. Roehrig, Russell Staggs, Col John A. Warden III, and James S. Wineinger. Bob Cover deserves special thanks for arranging tours of Sandia National Laboratories and interviews with many laboratory personnel. Roger E. Rinaldi has my sincere thanks for arranging tours of and interviews in industrial facilities in Tulsa, OK, and for answering innumerable questions about the oil and gas industries. Finally, I am indebted to Maj Pete Faber for his guidance and encouragement throughout this project.

I alone am responsible for any technical errors or inadequacies in this report.

 

Steven M. Rinaldi

8 June 1994

ABSTRACT

Economies are complex systems, composed of a number of infrastructure elements. These elements, such as electrical grids, petroleum-oil-lubricants (POL) distribution networks, and telecommunications system, are interconnected in a myriad of ways. As a result of this connectivity, an attack on one infrastructure element influences the others to varying degrees. When targeting an economy, an air planner must account for this connectivity and the downstream effects that naturally occur. Historically, however, air planners have overlooked the interrelated nature of a nation’s infrastructure and employed reductionist targeting techniques. Typically, they split an economy into individual target sets. Then, they select targets in each set in isolation from other targets, without anticipating the holistic effect of air bombardment. This is an inappropriate technique for targeting, as it overlooks the complex behaviors and characteristics of economies.

In this thesis, we propose a new manner of targeting economies—a holistic approach that accounts for the linkages between infrastructure elements and their resultant synergies. We first establish a theoretical foundation for targeting based on complexity science. This discipline examines the nature of complex, interconnected systems such as economies. Next, we demonstrate that economies are indeed highly interconnected systems. These linkages cannot be ignored in the targeting process. Finally, we tentatively propose a computer algorithm capable of targeting multiple, interacting infrastructure elements. The technique employs a genetic algorithm coupled to standard industrial analysis programs. When implemented, this computer technique should dramatically improve the effectiveness of economic targeting.

TABLE OF CONTENTS

Approval.........................................................................................................................ii

Disclaimer......................................................................................................................iii

Acknowledgments..........................................................................................................iv

List of Figures ..............................................................................................................vii

List of Tables ..............................................................................................................viii

Abstract .........................................................................................................................ix

About the Author.............................................................................................................x

1. LOOKING BEYOND THE WEB............................................................................1

Objectives and Scope...........................................................................................2

A Convergence of Technologies..........................................................................3

Thesis Organization.............................................................................................5

2. COMPLEXITY THEORY AND ECONOMIC TARGETING................................10

Complex Systems Defined.................................................................................11

Characteristics of Complex Systems...................................................................14

Emergent Behavior.................................................................................15

Adaptive Self-Organization....................................................................18

Evolution to the Edge of Chaos..............................................................23

Information Processing...........................................................................26

Summary............................................................................................................27

3. SYNERGIES AND INFRASTRUCTURE ELEMENTS........................................36

Targeting Issues.................................................................................................36

Dissecting an Economy......................................................................................38

Nodes and Couplings..............................................................................41

Summary............................................................................................................45

4. MODELING AND SIMULATION TECHNIQUES...............................................49

Targeting Philosophies.......................................................................................51

Numerical Techniques........................................................................................55

Iterative Improvement............................................................................57

Simulated Annealing..............................................................................60

Genetic Algorithms.................................................................................64

Targeting Economic Sectors: A Proposed Numerical Simulation......................69

Problem Description...............................................................................69

Electrical Grid Load-Flow Analysis........................................................70

POL Network Hydraulic Analysis...........................................................71

The Database..........................................................................................73

The Genotype.........................................................................................75

The Fitness Function...............................................................................75

Program Logic........................................................................................78

Discussion..............................................................................................80

Philosophical Aspects of Computer-Based Targeting.........................................81

Summary............................................................................................................82

5. CONCLUSIONS....................................................................................................91

Recommendations..............................................................................................92

Final Remarks....................................................................................................93

Appendix: SYNERGY TABLES..................................................................................95

Bibliography................................................................................................................109

LIST OF FIGURES

1. The hierarchical structure of complex systems........................................................12

2. Local interactions lead to emergent global behavior................................................15

3. The three regimes of dynamic systems....................................................................24

4. The model economy...............................................................................................39

5. Simplified view of the input-based targeting process..............................................52

6. Simplified view of the output-based targeting process............................................54

7. Outline of the iterative improvement algorithm......................................................57

8. A multipeaked cost function...................................................................................59

9. Outline of the simulated annealing algorithm..........................................................62

10. Outline of a genetic algorithm.................................................................................66

11. Simplified flowchart for the targeting program.......................................................79

LIST OF TABLES

1. Typical Couplings Between Elements in the Model Economy................................42

2. Typical Load-Flow Analysis Inputs........................................................................71

3. Typical Hydraulic Analysis Inputs..........................................................................72

4. Fitness Function Rewards and Penalties..................................................................77

5. Electrical Grids.......................................................................................................97

6. POL Distribution Networks..................................................................................100

7. Communication Networks....................................................................................105

8. Transportation Systems.........................................................................................106

CHAPTER 1

LOOKING BEYOND THE WEB

 

Science today stands on something of a divide. For two centuries it has been exploring systems that are either intrinsically simple or that are capable of being analyzed into simple components. The fact that such a dogma as "vary the factors one at a time" could be accepted for a century, shows that scientists were largely concerned in investigating such systems as allowed this method; for this method is often fundamentally impossible in the complex systems.

—Ross W. Ashby

Since the dawn of powered flight, air planners have recognized that economies are complex, interconnected systems. As early as 1911, French Lieutenant Poutrin wrote in Revue Générale de l’Aéronautique Militaire that German aerial attacks on key ministries, transportation networks, and communication centers in Paris would shut down essential public services, thus preventing the country from mobilizing. However, the notion of economies as intertwined entities took on particular meaning at the Air Corps Tactical School (ACTS) in the 1930s, with the development of the "industrial web" theory by the ACTS "bomber mafia." Since modern theorists have likewise recognized the importance of connections between economic sectors, we would logically expect air planners to identify and exploit the linkages present in modern economies. Paradoxically, however, most economic targeting has proceeded as though the interconnections between elements of an economy were of secondary importance.

In practice, air planners have behaved much as Ashby’s scientists. They have dissected economies into their component parts and targeted each part in isolation. During WWII, the industrial web theory influenced planners to search for bottlenecks—those critical industries upon which significant portions of an enemy war economy relied. But once such industries were pinpointed, the planning tended to focus on destroying individual target sets rather than attacking key points in different sets. Consequently, when one system failed to yield the desired results, the priorities shifted to another target. Planners "sought, partly and inescapably through trial and error, to find good target systems." Even though they only partially succeeded, this reductionist approach survives today. During the Persian Gulf War, for example, planners did not perform detailed systems analyses that highlighted the intertwined nature of the Iraqi economy. As in prior conflicts, airmen recognized the complex interconnections of the Iraqi economy, but did not fully exploit them in the planning process.

Targeting science today stands on something of a divide. We have reached the point where we can retire the old reductionist targeting methods. Rather than separating the different sectors of an economy and targeting each one in isolation, we can now approach targeting from a synergistic or holistic viewpoint. Economies are complex systems and must be targeted as such. The purpose of this research is to move us across the targeting divide, from the old reductionist ways to newer holistic methods.

OBJECTIVES AND SCOPE

The focus of this study is economic targeting. It has two principle objectives. First, it will demonstrate that economies are complex systems that do not readily submit to reductionist analyses. This point has important ramifications. As noted above, planners following the traditional approach first separate the economy into its principle infrastructure elements, such as electricity, POL, telecommunications, and transportation. Next, they examine each element in isolation from the other elements and select appropriate targets. This approach ignores the linkages that exist between infrastructure elements, and that they cannot be deduced from individual analysis.

If the traditional method is inappropriate, how should a planner approach economic targeting? This question lies at the heart of the second objective of the study. It develops a new technique for analyzing economies that preserves their complex nature. The technique draws from several recent developments in computer technology as well as standard industrial analyses. The result is a proposed computer tool capable of performing targeting analyses on the interconnected elements of an economy. This computer algorithm should help overcome the current dearth of economic modeling and simulation tools available for targeting.

Several limitations apply to this study. First, it does not address how economic targeting brings about desired political objectives. The study is strictly limited to examining the ties between target sets, not the political mechanisms triggered by synergistic attacks. Second, the study presumes that our national authorities ordered economic targeting. It does not debate the relative merits and drawbacks of targeting economies. Third, much of the data originates from interviews with U.S. industrial experts. Many of the examples, therefore, apply to the U.S. infrastructure and may not be valid for foreign nations. The usual caveats against mirror-imaging strongly apply to this report. However, the computer analysis technique developed below is applicable to any economy. Finally, this project is primarily theoretical—it develops a framework for analyzing economies. The "experimental verification" of the theory awaits future research efforts.

A CONVERGENCE OF TECHNOLOGIES

Why are we only now beginning to take a hard look at the interrelationships that exist between infrastructure elements? To a large degree, the answer lies in scientific advances. The recent convergence of several technologies provides us with the ability to examine the question in detail. These technologies include stealth, precision-guided munitions (PGMs), day-night aircraft with improved weather capabilities, and computer algorithms and hardware.

During WWII, sequential air operations were the general rule. Air planners selected target sets and attacked them in series. To a large degree, aircraft limitations and weather were responsible for this mode of operation. High bombing circular error probables (CEP) meant that huge numbers of aircraft had to mass together in order to mathematically assure mission success. Consequently, the number of targets that could be attacked simultaneously was severely limited. Relatively low bomb loads (by today’s standards) also drove up the number of bombers required for a mission. Distant targets, which required more fuel and smaller bomb payloads, exasperated the problem. Finally, the weather constrained the number of bombing days per month. These factors strictly limited the number of targets that could be attacked simultaneously or within a short time period. As a result, planners sought bottleneck targets that could be attacked sequentially. Out of necessity, synergies created by near simultaneous attacks in different target sets were pushed aside in favor of seemingly lucrative bottlenecks.

Fifty years later, technological revolutions allowed a radically different form of airwar during the Persian Gulf War. PGMs decreased the number of sorties required to destroy targets by orders of magnitude. Increased payloads further reduced the aircraft requirements per target. Stealth and cruise missiles contributed as well; targets in highly defended areas were suddenly vulnerable to attack. Day and night operations were possible due to improvements in sensors and aircraft. In sum, technological advances allowed a fundamental shift in operations: parallel- and hyper-warfare were born. Instead of serial attacks against bottleneck targets, planners during the Gulf War attacked significant numbers of targets in different sets simultaneously. The goal was strategic paralysis, a condition in which the Iraqi authorities were unable to react to Coalition actions. Consequently, conditions were right for the air planners to seek synergistic effects during operations. Unfortunately, a crucial element was missing: the capability to perform nodal analyses of infrastructure elements.

Today, the final piece of technology is appearing that will allow a major shift in the way planners target economies. Engineering and nodal analyses are beginning to appear in military writings. Even so, an analysis of a single element, such as POL or telecommunications, is simply an improved version of the reductionist approach. What is still required is a method that allows analyses of multiple economic elements that preserves the complex interconnections between them. The development of such analysis tools will allow a further revolution in operations: we will be able to leap across Ashby’s divide.

Concentrating on synergistic effects in analyses will do far more than just lead to elegant target sets. Rather, it will allow a fundamental shift in operations. Consider for the moment a conflict in which the number of available aircraft or munitions is severely restricted. In this type of conflict, every sortie must count and every bomb must hit a significant target. "Rules-of-thumb" or "back-of-the-envelop" techniques are inadequate to determine the appropriate targets. Furthermore, reductionist analyses of isolated target sets, even if performed with tools such as load-flow programs, will not produce the best target sets—in fact, inappropriate targets may result. We will examine these propositions in detail in the ensuing chapters.

THESIS ORGANIZATION

Chapter 2 provides the theoretical framework for targeting studies. This framework draws heavily from dynamical systems analysis, or complexity theory as it is more popularly known. Complexity theory deals with the behaviors of complex, interconnected systems. As we shall see, economies are examples of the type of systems that complexity theory treats.

In Chapter 3, we move from abstract theory to a concrete discussion of the complex nature of economies. The chapter begins with a discussion of centers of gravity (COG). Determining whether a given infrastructure element is a COG is a crucial step in planning. Next, the chapter looks at a detailed notional economy that contains four elements (an electrical grid, natural gas and oil distribution networks, and a telecommunications system). The example illustrates the complex nature of economies, and highlights the inadequacies of reductionist targeting.

Chapter 4 then explores computer techniques applicable to targeting complex economies. It begins with a discussion of targeting philosophies, then presents several computer algorithms that a planner could employ in a targeting program, and examines at length a proposed computer program that produces target sets for complex economies. The program uses standard industrial analysis programs (such as a load-flow program) to model the effects of an attack on given infrastructure elements. However, it preserves the complex linkages between the different elements. In this respect, the program represents a considerable step forward in targeting.

Chapter 5 recapitulates the general points made earlier and recommends additional research. The Appendix, in turn, contains several tables that list typical linkages between the electrical, POL, telecommunications, and transportation elements. The tables provide a final, compelling argument for the holistic analysis techniques advanced in this study.

CHAPTER 2

COMPLEXITY THEORY AND ECONOMIC TARGETING

Economies are vital, dynamic entities that constantly react to the influences of their environments. A typical economy consists of different infrastructure elements, such as electrical power grids, transportation networks, POL distribution systems, and communications systems. These elements intertwine in a myriad of nonlinear linkages and feedback loops. Examining each element in greater detail yields a hierarchy of parts, or key nodes and linkages. To understand how these nodes and linkages function, we must turn to dynamical systems analysis, or complexity theory as it is popularly known.

Dynamic systems analysis seeks to explain how time-dependent systems operate. It encompasses complexity and chaos theories. Complex dynamic systems such as economies can exhibit a rich variety of behaviors, including many that are counterintuitive. An essential truth of complexity theory is that such systems must be analyzed from a holistic rather than reductionist point of view. In recent years, several papers have appeared examining warfare in light of chaos theory. However, analysis of economies for warfighting purposes goes beyond the bounds of chaos theory; the proper framework actually lies within complexity theory.

As a result, this chapter explores complexity theory and economic targeting. More specifically, it examines three premises. First, economies are complex systems. We can employ complexity theory to understand economic infrastructures and their behaviors. Second, given that economies are complex systems, air planners must account for their dynamics when targeting them. Finally, and perhaps most importantly, the reductionist methodology followed in traditional economic targeting is invalid. A planner must instead approach the problem from a synergistic viewpoint. These three contentions will become clear as we further explore the nature and the behavior of complex systems.

COMPLEX SYSTEMS DEFINED

Dynamic systems analysis has its roots deep in modern mathematics. Although some of it traces back to the work of the French mathematician Henri Poincaré in the early 1900s, the field has grown enormously during the past 15 to 20 years. Computers are largely responsible for the recent growth, since they have allowed scientists and mathematicians to study the temporal dynamics of complex, nonlinear systems. Today, complexity theory encompasses a wide variety of disciplines ranging from physics, chemistry, and biology to sociology, cultural anthropology, and economics.

A complex system is a large collection of interacting parts or entities. In the language of complexity theory, the parts are commonly referred to as agents. These interacting parts create the environment in which they exist. Further, by constantly acting and reacting with one another, the parts continually perturb and modify their environment. For example, in a chemical system, the physical interactions between the reagents form linkages. As the chemical reactions consume the reagents and create new products, the environment each molecule "sees" changes. The interaction rules for the agents can be quite simple. Even so, systems with simple interaction rules frequently give rise to extremely intricate behaviors. In sum, collections of interacting agents form the basis of complex systems.

Complex systems often contain hierarchical structures. Figure 1 sketches the concept. Separate groups of agents form within the system. Each group is in its own right an agent, and interacts with the other groups. In general, the form of the agents, environment, and interaction rules vary at each hierarchical level. A government bureaucracy illustrates the principle. Offices interact to form divisions, divisions cluster to create agencies, and agencies work with one another to create the overall governmental structure.

Agents and their linkages can be characterized in three ways. First, we can broadly classify the couplings between the agents as tight and loose. Tight coupling refers to agents that are strongly dependent upon one another. Disturbances in the system may be highly correlated to each other when the system is tightly coupled. Time-dependent processes, with little give or slack, characterize tightly coupled systems. Additionally, disturbances tend to propagate throughout a tightly coupled system. A natural gas-fired electrical generator and the gas pipeline system form a tightly coupled pair. The generator provides the electricity for the compressor stations along the pipeline, which in turn assure a constant supply of fuel to the generators. Since gas generators generally have no local storage, this particular system is very tightly coupled. Generator disturbances will directly affect the pipeline and vice-versa. Loose coupling, on the other hand implies that the agents are relatively independent of each other. Events in the system are usually either weakly correlated or independent. Give and slack exists in the system, since the processes are not nearly as time-dependent as in a tightly coupled system. A coal-fired electrical generator and the diesel powered railroad network supplying its coal are weakly coupled. Coal-fired generators often have 90 day supplies of coal available. Consequently, short term disturbances in the rail supply rarely affect power generation. Tight and loose coupling, then, refers to the degree of dependencies between the agents.

Second, the couplings between the agents may be further classified as linear or nonlinear. These terms describe the mathematical forms of the linkages. A linear linkage satisfies two conditions. First, it obeys the law of proportionality. If an input x to some linkage results in an output y, then linearity requires that an input of ax produces an output of ay, where a is some arbitrary constant. Second, the linkage must allow the superposition or additivity of inputs. That is, if the inputs x1 and x2 give rise to outputs y1 and y2, respectively, then the input (x1 + x2) yields (y1 + y2). Nonlinear linkages do not display proportionality or superposition. Slight changes in the inputs may result in disproportionately large differences in the outputs or vice-versa, depending upon the mathematical relationships between the inputs and the outputs. In a nonlinear system, the whole is not necessarily equal to the sum of the parts. Clearly, linear linkages are simpler to understand and model than nonlinear linkages.

Lastly, agents or collections of agents may carry out sequential or branching processes. This refers to the manner in which the agents carry out their processes. Sequential processes, as the name implies, proceed in series: event A occurs before event B which occurs before event C. The process occurs in steps, much like an assembly line. Most systems designed by man are sequential. Branching operations, however, have a more involved structure: they contain feedback loops, branches and bifurcations, and jumps from one linear sequence to another. The consequences of perturbations in a branching operation are much less clear or comprehensible than in a sequential operation. Communication systems contain this type of branching, as calls can be routed in a myriad of directions, through many different switching stations. The German war economy during WWII is a further example of a branching process. Production was characterized by a geographic division of labor and supplies, linked by the Reichsbahn. As industry dispersed in response to the Allied bombing attacks, the number of supply and subassembly paths between facilities proliferated. Consequently, the economic processes became intrinsically more intricate in the manner denoted above.

In summary, complex systems consist of large collections of interacting agents. Hierarchies of agents may be present, each with its own particular dynamics. Agents are tightly or loosely coupled to one another, depending upon their degrees of interdependency. Furthermore, the linkages themselves are either linear or more typically nonlinear. Finally, the agents can carry out sequential or branching operations. With this abstract background on the nature of complex systems, let us turn now to their characteristics and behaviors.

CHARACTERISTICS OF COMPLEX SYSTEMS

Complex adaptive systems appear in a wide variety of settings. Examples readily arise in the scientific and engineering disciplines. However, we also find complex behavior within the social and human sciences. Cultural, social, and political systems all display attributes of complexity. An economy is an "example par excellence" of a complex adaptive system. A considerable body of research has been amassed relating to complexity in economies. Herein lies the motivation behind this section: targeting complex adaptive systems requires an understanding of their dynamics.

Despite their diversity, complex systems share certain fundamental behaviors. These include emergent behavior, adaptive self-organization, evolution to the edge of chaos, and the ability to process information. Unfortunately, valid mathematical laws do not exist that describe the conditions under which these behaviors appear. Nevertheless, they are general patterns that arise in a wide variety of systems. The following subsections examine each of these properties in detail. Interestingly, many of the system properties commonly discussed in military circles, such as Allison’s organizational and political models, economic substitution, and Boyd’s observe-orient-decide-act (OODA) loop, arise naturally as a result of complex behavior.

Emergent Behavior

Emergent behavior is an important distinguishing characteristic of complex systems. The interactions of the agents may lead to emerging global properties that are strikingly different from the behaviors of the individual agents. These properties cannot be predicted from prior knowledge of the agents. The global properties in turn affect the environment that each agent "sees," influencing its behavior. A synergistic, feedback loop is thus created—the interactions between agents determine the emerging global properties which in turn influence the agents. Figure 2 diagrams the process. Furthermore, emergent structures of agents can arise through their mutual competition and rivalries. Each structure exhibits its own emergent behavior, and in turn influences the global behavior of the system. The emergence of coherent, global behavior in a large collection of agents is one of the hallmarks of complex behavior.

Emergent behavior implies that reductionist analysis methods have limits when applied to complex systems. The traditional scientific method is the embodiment of the reductionist philosophy. The researcher decomposes a system into progressively smaller and smaller parts. His analysis centers on the properties of the pieces, rather than the dynamics of the entire system. Particle physics, the study of the basic building blocks of matter, is perhaps the ultimate limit of this philosophy. However, by focusing on the parts instead of the system as a whole, the emergent properties are lost. The blurring of the global behavior occurs because the emergent properties are functions of the interactions between agents and their effects. Although important, the properties of the agents do not exclusively determine the global system properties. As P. W. Anderson notes:

The ability to reduce everything to simple fundamental laws does not imply the ability to start from those laws and reconstruct the universe. In fact, the more the elementary particle physicists tell us about the nature of the fundamental laws, the less relevance they seem to have to the very real problems of the rest of science, much less to those of society.

An analysis of complex systems requires a holistic or constructionist approach rather than a reductionist one. Consider for the moment a complex system composed of hierarchies of agents. The structures and the rules of behavior change with each level of organization. New emergent behaviors appear for each layer in the hierarchy. To understand the global properties of the system, one must move progressively higher and higher in the organization, incorporating the interactions between each level of agents in the analysis. Allison’s Governmental (Bureaucratic) Politics Model III illustrates these principles quite well. He describes this model as:

[T]he Governmental (or Bureaucratic) Politics Model sees no unitary actor but rather many actors as players—players who focus not on a single strategic issue but on many diverse intra-national problems as well; players who act in terms of no consistent set of strategic objectives but rather according to various conceptions of national, organizational, and personal goals; players who make government decisions not by a single, rational choice but by the pulling and hauling that is politics.

According to this model, the global emergent properties (the strategic decisions) of the government come about not because of the personal, organizational, and national goals of the agents (players), but rather because of the interactions (political maneuvering) between the agents within the governmental hierarchy. An a priori knowledge of the agents does not suffice in comprehending the emergent decisions of a government. A holistic, rather than reductionist, methodology that includes the interactions between agents and various hierarchical levels is necessary to understand global dynamics.

The reductionist philosophy likewise fails to properly analyze economic infrastructure elements for targeting purposes. It wrongly perceives economic structures as isolated entities. Rather, they are intertwined in complex, nonlinear ways and continuously interact with and influence each other. The usual method of targeting the elements in isolation overlooks the synergistic effects and the emergent properties of the economy. An appropriate targeting philosophy must incorporate these interactions and examine the global properties of the system. The overall effects of an attack can only be determined in this manner.

This point is crucial in an armed conflict, as "war is...an act of force to compel our enemy to do our will." The target(s) and timing of an attack should be designed to trigger a mechanism that precipitates a desired political outcome. If the synergies and emergent properties of the enemy economic system are not considered during the planning process, then the specific results of an air attack cannot be adequately assessed. As a result, the inputs to the mechanism blur or become indeterminate. In the worst case, the connection between the attacks and the desired political outcomes breaks; in the best case, the level of fog and friction in the targets-timing-mechanism-political outcome paradigm rises. Incorporation of synergies and emergent behaviors in targeting analyses will not assure that the air attacks produce the desired political outcomes. However, it will provide the planner with a clearer view of the effects of air attacks and hopefully reduce the level of uncertainty associated with targeting.

Emergent behavior is a hallmark of complex systems. It demands that analyses proceed from a synergistic rather than reductionist viewpoint. The implications for targeting are clear: interactions between target sets must enter the decision making process if the global effects of air attacks are to be determined.

Adaptive Self-Organization

Adaptive self-organization is the second fundamental characteristic of complex systems. As Kauffman notes, "contrary to our deepest intuitions, massively disordered systems can spontaneously ‘crystallize’ a very high degree of order." This appears to be an innate property of complex systems. Self-organization arises as the system reacts and adapts to its externally-imposed environment. Such order occurs in a wide variety of systems, including for example convective fluids, chemical reactions, certain animal species, and societies. Although the specific mechanisms of self-organization are difficult to determine, some valuable insights come from considering the Second Law of Thermodynamics and the concept of "attractors."

At first glance, self-organization appears to be in direct conflict with the Second Law. The law states that for isolated systems that cannot interact with their environment, entropy will remain constant or increase in time. The increase in entropy is commonly interpreted to mean that the isolated system will become increasingly disordered, since higher entropy levels are associated with an increased number of states or configurations in which the system can exist. This interpretation has been incorrectly applied to nation-states and military systems. A more detailed examination of the Second Law helps clear up the discrepancy.

For a system of weakly- or non-interacting particles, the implications of the Second Law are fairly clear. If such systems are isolated (that is, they cannot exchange energy, matter, etc., with their environments), then the usual interpretation of the law holds. For nonisolated systems, the situation changes considerably. In this case, exchanges with the environment take place. The internal entropy of the system will still increase or remain constant in time. However, entropy can now be exchanged with the environment. The flux of entropy may be such that the overall entropy of the system decreases. For the noninteracting particles, this decrease in entropy is often associated with a decrease in the accessible states, or an increase in order.

The situation is much less clear for nonisolated systems of strongly interacting particles or agents. In the presence of nonequilibrium constraints, the system can evolve toward multiple ordered states via phase transitions. Kauffman notes that "the second law really states that any system will tend to the maximum disorder possible, within the constraints due to the dynamics of the system." In sum, for nonisolated systems, whether weakly or strongly interacting, there is no conflict with the Second Law. Systems can move toward states of higher order under the influence of exchanges with the environment and external, nonequilibrium constraints.

Attractors provide a further means of visualizing the increases in order. Every system has an associated phase space. In the simplest terms, the phase space is the set of all possible states in which the system may exist. The dynamics and external constraints imposed on the system will generally limit it to certain regions of phase space. These regions are called attractors. An undisturbed system remains within the confines of its attractor. In general, the attractors contain only a fraction of the total volume of phase space. In this sense, then, attractors represent an increase in order of the system: the system is not free to wander through all regions of phase space.

Attractors themselves are dynamic. The environment of the system is not necessarily static. It may change in time, for example, if some of the constraints are altered. As the environment changes, the attractors shift. The system correspondingly modifies its behavior, as it is constrained to its attractors.

Attractors are not just abstract notions that apply to systems from physics and mechanics. Attractors exist for all dynamic systems, whether they are immune systems, neural networks, bodily organs, communities, or ecosystems. Economies self-organize around their attractors, as they are complex dynamic systems. The attractors define the types and limits of behavior that the system can exhibit. In short, attractors may be thought of as imposing order upon the complex systems.

What is the applicability of adaptive self-organization to warfare, and in particular, to economic targeting? Since economies are complex entities, the ties are direct. Two examples, one abstract and the other concrete, will help illustrate the connections between adaptive self-organization and warfare.

The abstract example draws from Clausewitz’s notion of a reactive enemy. In discussing the differences between warfare and the arts and sciences, Clausewitz states:

The essential difference is that war is not an exercise of the will directed at inanimate matter, as is the case with the mechanical arts, or at matter which is animate but passive and yielding, as is the case with the human mind and emotions in the fine arts. In war, the will is directed at an animate object that reacts.

Furthermore, the two parties in the conflict mutually interact and affect one another:

War, however, is not the action of a living force upon a lifeless mass (total nonresistance would be no war at all) but always the collision of two living forces. The ultimate aim of waging war, as formulated here, must be taken as applying to both sides. Once again, there is interaction.

Clausewitz implies later that interactions occur not only between the combatants, but also in the processes of war itself. He portrays war as an interactive undertaking in which each adversary reacts to changes in himself originating in enemy actions, and changes in the environment arising from the actions of both combatants.

These concepts parallel the notions of adaptation to environmental changes and self-organization. The conflict changes the environment in which the combatants exist. The environmental shifts modify the phase space and the attractors of the adversaries. In order to survive, the adversaries must adapt to the changes. An adversary who cannot adapt or whose attractors shift too radically will end up broken and defeated on the battlefield.

Wartime economic substitutions and workarounds provide a second, more concrete connection between adaptation theories and warfare. Economies are organic and adjust to the disruptions of war, in manners consistent with Clausewitz’s observations. An examination of the WWII wartime economies and Iraq in the aftermath of the Persian Gulf War reveals several possible adjustments:

• elimination of excessive or luxurious uses of items;

• diversion of resources and the means of production from nonessential to militarily essential items;

• rationing or reduced distribution of essential goods and services;

• substitution of different raw materials for scarce ones required in manufacturing;

• increasing the number of shifts per day;

• operation of equipment in hazardous manners that would normally not be employed in peacetime;

• removal of cost reducing practices that hinder the production rate (practices that reduce the costs of fuel, coal, electricity, water, etc.); and

• dispersal of critical industries and other measures to reduce the vulnerability to attack.

Although the direct and indirect costs of these adjustments may be high, wartime necessity can justify their implementation. Furthermore, Olson makes three observations about the degree of substitution that an economy can make. First, when a small portion of any industry is destroyed, the economy can adjust to make up the relatively small shortfalls. Items are shifted from nonessential uses to the critical needs. As a result, he argues that an economy can weather small losses to even a large number of industries (such as those created by the British area bombing of Germany). Second, if a large part of an industry is destroyed, the economy has a much more difficult time making up the losses. Finally, an economy can accommodate the loss of most of any industry provided the industry is small enough. The small, critical industry upon which the entire economy hinges simply does not exist. These ideas closely track with some of the concepts of phase spaces and attractors.

If small perturbations occur in a system, the attractors shift by relatively small amounts. If the shifts are not too great, the system transitions to the modified attractors. This particular case roughly corresponds to Olson’s first and third points. Economic adaptation accommodates relatively small disturbances. However, if the perturbations are large, the qualitative nature of the phase space may change. Alternatively, a large disturbance could so modify the attractors that the system shifts to a different attractor altogether. In either case, the behavior of the system changes markedly. Olson’s second observation corresponds to this case. The changes in the attractors thus qualitatively explain the modifications observed in the economies.

Adaptive self-organization is the second characteristic of complex systems. The systems tend to crystallize into some ordered state, from which the emergent behaviors become apparent. The exact origins of the order are uncertain. Economic systems are subject to self-organization, and display many of its properties. In particular, the adjustments of economies under the duress of war are manifestations of the dynamics of adaptive self-organization.

Evolution to the Edge of Chaos

Dynamic systems occupy a "universe" composed of three regimes. The first is an ordered, stable region. Perturbations to the systems tend to die out rapidly, creating only local damage to the system. Additionally, information does not flow readily between the agents. In the second regime, chaotic behavior is the rule. Disturbances propagate rapidly throughout the system, often leading to destructive effects. The final regime is the boundary between the stable and chaotic regions. Known as the complex regime, it is a phase transition region between the stable and chaotic regions. According to Kauffman, systems poised in this boundary regime are optimized to evolve, adapt, and process information about their environments. Disturbances propagate throughout the system, yet damages are limited, allowing further system evolution. Figure 3 shows the relationship between these three regimes.

As complex systems evolve, they appear to move toward the edge of the chaotic regime. As time progresses, they generally increase in complexity. Competition and environmental disturbances force the systems to discover new "functions" that allow them to perform new tasks and better survive in a competitive world. These phenomena parallel natural selection. The same type of growth in complexity occurs in manmade objects, such as jet engines, microprocessors and microelectronics, and software. Social, economic, and bureaucratic structures display similar evolution patterns as well.

In the case of bureaucratic structures, Allison’s Organization Process Model II illustrates the evolution toward the edge of chaos. The essence of this model is that a government (or by implication, an organization) is "a loose conglomeration of semi-feudal, loosely allied organizations...Governmental behavior can therefore be understood, according to [this] model, less as deliberate choices and more as outputs of large organizations functioning according to standard patterns of behavior." According to this model, a government approaches a problem by first factoring it into manageable pieces. It distributes the pieces to the appropriate agencies for resolution. The agencies solve their parts of the problem by applying standard operating procedures (SOPs). Normally, the agencies have repertoires of SOPs from which to choose. An agency will usually choose the first solution to the problem that it finds in its repertoire, rather than looking for the best solution. Thus, the individual solutions generally are satisficing rather then optimizing. The overall solution to the problem is a conglomeration of individual solutions, tailored by governmental leaders. Problem resolution is essentially the output of organizational routines.

Occasionally, a problem confronts an agency for which no SOP is applicable. Then, the agency uses a simple-minded procedure to develop an applicable solution. The agency serarches its existing routines for one that can be adapted to the problem at hand. It is more likely to adapt an existing SOP than generate an entirely new procedure. Nevertheless, the agency learns from the experience and adds the new SOP to the repertoire. The evolutionary growth is slow and incremental, although major disasters may precipitate large organizational changes.

In many respects, the Model II evolutionary growth parallels that of a complex system. The system has a set of internal "models" or analogies that drive its behavior. The models are used to interpret the environment and guide the system’s reactions. The models are not static; they are modified to accommodate new situations. Feedback about the performances of the models further tunes them. The models are the counterparts of Allison’s SOPs; both fulfill the same role in problem resolution and system evolution.

By extension, the same processes occur in an economy during a conflict. Destruction of materiel, the means of production and distribution, and other infrastructure elements pose significant problems that the economic system must overcome. The damage inflicted by the adversary drives the system away from the edge of chaos and toward either the chaotic or stable regions. The system attempts to return to the complex region by invoking a set of emergency SOPs, such as those detailed by Olson and Arkin or employed by Speer. Although not evolution per se, the movement back toward the edge of chaos is an attempt to limit or overcome the damages and return to the status quo.

Evolution toward the edge of chaos is a natural behavior of complex systems. Allison’s organizational process models closely tracks the major themes of this evolution. There appear to be similarities between the reactions of economies to wartime destruction and the dynamics of organizations and complex systems. Although only advanced here as a hypothesis to explain the actions of an economy under the pressures of war, the use of complexity theory to frame the problem merits further examination.

Information Processing

In general terms, complex systems possess the ability to process information. The systems sense their environments and collect information about surrounding conditions. They then respond to this information by using a set of internal models to guide their actions. The systems may also encode data about new situations for use at a later date. This characteristic is closely related to the adaptation that occurs near the edge of chaos.

Information processing is similar to Boyd’s OODA loop. In his model, a system observes some event of interest, decides how to resolve a problem posed by the event, and finally acts upon its decision. This process frequently arises during military operations, where the commander’s objective is to "get inside the enemy’s OODA loop." He does this by simultaneously destroying the enemy’s capability to sense, process, and act on information while preserving his own ability to do so. Once reaching this point, the commander can force the enemy to constantly react rather than take the initiative. Both the friendly and enemy sides cycle through the OODA process; the friendly objective is to do so more rapidly than the adversary. In doing so, the enemy’s actions lose coherence with the changing environment.

In an exceptionally close paraphrase of Boyd’s theory, Kauffman describes how information is processed by complex systems:

But it is also plausible that systems poised at the boundary of chaos have the proper structure to interact with and internally represent other entities of their environment...organisms sense, classify, and act upon their worlds. In a phrase, organisms have internal models of their worlds which compress information and allow action...Such action requires that the world be sufficiently stable that the organism is able to adapt to it. Were worlds chaotic on the time scale of practical action, organisms would be hard pressed to cope.

By sensing, classifying, and acting upon their worlds, organisms cycle through Boyd’s OODA loop. Chaos on the time scale of practical action implies that some organisms are unable to assimilate environmental conditions fast enough. Here, the environmental conditions change within a single period of an organism’s OODA loop, precluding it from acting coherently. In essence, the environment has "gotten within the organism’s OODA loop." The correspondence between Boyd’s model and the information processing capabilities of complex systems is striking.

Information processing is the final major characteristic that complex systems possess. It relates to the system’s ability to sense its environment and act in a manner consistent with a set of internal models. The frameworks erected by complexity theory, Allison’s organizational processes model, and Boyd’s OODA loop possess a remarkable degree of consistency with respect to information processing.

SUMMARY

In this chapter, we have explored the nature of complex systems. These systems are large collections of interacting agents. The agents may be tightly or loosely coupled to one another via linear or nonlinear linkages. The interactions between the agents define and influence the environment in which they exist. Complex systems are not simply abstract creations of mathematicians and scientists; real-world examples abound in such diverse fields as physics, chemistry, biology, sociology, anthropology, and most importantly for this study, economics.

Further, complex systems display several virtually universal characteristics. First, the local interactions of agents give rise to emergent, global behavior. This behavior cannot be predicted a priori, even given perfect knowledge of the individual agents. Consequently, reductionist analyses tend to miss the important macroscopic features of systems while concentrating on the microscopic ones. The analyst must use holistic or constructionist approaches to determine emergent, global behavior. Second, systems tend to adapt to their environments and self-organize. Instead of tending toward maximum disorder as do weakly interacting isolated systems, complex systems spontaneously crystallize into highly ordered states. Third, systems evolve toward the edge of chaos and increase in complexity as time passes. They "learn" from their environments and add new functions to cope with previously unknown conditions. In doing so, they increase their complexity and adapt along the edge of chaos. Finally, systems exhibit the ability to process information. Simply put, they can sense their environment and react to it based on internalized models. Information processing is closely related to a system’s ability to learn near the edge of chaos. Although these properties may appear to be associated with living organisms only, they apply to complex systems in general, ranging from the man-made to the organic to the sociological and cultural.

Since an economy is perhaps the "example par excellence" of complex systems, economic targeting should take into consideration the above characteristics. In particular, planners must bear in mind that reductionist approaches to targeting will miss the global properties of economies—and consequently, the global effects of an attack. A holistic approach, in which the synergies between target sets are properly included, is the correct one to follow. Other considerations that play into (economic) targeting, such as Olson’s theories of substitution and Allison’s models of organizational dynamics, fall naturally under the framework of complexity theory. In short, economies are complex, adaptive systems and they should be targeted as such.

In the next chapter, we will look at four economic infrastructure elements from a nonreductive viewpoint. The elements are national electric grids, natural gas distribution networks, oil distribution networks, and communication systems. Of particular interest are the ties between each of these elements. Although limiting the study to only four elements is in its own right reductionist, it is a first step in demonstrating the types of synergies that can be expected within economic structures. It is furthermore a step away from the traditional method of targeting each element in isolation.

CHAPTER 3

SYNERGIES AND INFRASTRUCTURE ELEMENTS

In the previous chapter, we saw that economies are complex systems. However, the text treated the subject in a rather abstract manner, employing the language of complexity theory. For practical purposes, the planner needs to appreciate in more solid terms the synergies that exist between economic infrastructure elements. In this chapter, we address this point.

A recurring theme throughout the chapter will be the shortcomings of the reductionist approach. In the past, air planners have generally followed the reductionist path, primarily due to a lack of suitable analysis tools. Nevertheless, this traditional approach must yield to a synthesis-based methodology.

This chapter explores in detail the interlinked nature of economic infrastructure elements. The goals are to deepen our appreciation of the complex nature of these elements, and demonstrate the pressing requirement for a nonreductionist approach. The chapter first contemplates several issues associated with targeting in general and synergies in particular. The balance of the chapter then examines a simplified infrastructure, composed of an electrical grid, a natural gas pipeline system, an oil pipeline system, and a communications network. The subtle problems posed by targeting interrelated systems should become clear as we study this example.

TARGETING ISSUES

When examining an enemy economy, a fundamental issue the air planner must address is whether or not its infrastructure elements are also COGs. The value of the economic targets must be considered in light of political and military strategies and the commander’s objectives. A Douhet-style campaign aimed at the enemy population might place a high value on economic elements under certain circumstances. However, a strategy based on isolating enemy leadership may place an entirely different emphasis on an economy. The political objectives and the nature of the conflict are important determinants of COGs.

In addition, an infrastructure element may not be important just in economic terms. Many elements have military as well as societal or economic significance. National railroad systems may provide the bulk of long-distance freight shipping capacity for a nation; yet the same railroads, locomotives, and rolling stock may also serve as the primary means of national mobilization. Similar examples can be found in other infrastructure elements. Thus, each element must be examined from a military and economic viewpoint.

Further, the value of each element varies on a country-by-country basis, as well as with the particular contingency. The cultural and societal dimension cannot be overemphasized. For example, American society is highly organized around plentiful energy sources (electricity, oil, coal, etc.). Likewise, energy sources are the driving force behind much of U.S. economic power. Without a continuous, reliable energy supply, the U.S. economic base would be in dire straits. However, in an agrarian society, such as Vietnam in the 1960s, the same elements may be of negligible significance. Societies are individualistic, and each must be regarded as such.

Synergistic effects between different target sets also play into COG considerations. As with any complex system, surface or outward appearances of certain elements may be deceiving. Nonlinear, branching linkages may place greater weights upon certain elements than are initially evident, based upon a cursory examination. As a hypothetical example, we assume that some Country X produces all of its electricity from oil- and gas-fired generators. Furthermore, the gas compressors, oil pumps, controls, valves, and central dispatch station along the pipeline require electricity to function. A feedback loop is present in the electric grid-POL distribution complex. Loss of electricity will affect the pipeline operations. Similarly, a reduction or shutdown of oil and gas flow will impact electrical generation. The analysis must not stop at this simple level of determining synergies; the importance of electricity, oil, and natural gas must be evaluated for the society as a whole. To illustrate this point, we further assume that the society has few requirements for POL. At first glance, POL may have little significance as a COG. However, if the society or military is critically dependent upon the national electric grid, then oil and gas assume greater importance as economic COGs. Bottlenecks and synergies may in themselves be deceiving; they too must be examined from a holistic viewpoint.

In a related vein, the existence of substitutes, workarounds, emergency plans, and repair capabilities may obviate some seemingly promising economic targets. Certain elements may possess extensive contingency plans or backup capabilities. Other elements may be highly redundant and flexible. Such considerations will complicate targeting, and an air planner must consequently factor them into COG analyses as well. An economic infrastructure element may be a valid COG, but it may also be very difficult to destroy in a direct manner.

A network analysis is an important part, then, in evaluating economic COGs and target sets. This is all the more important in light of complexity theory, where emergent behaviors may be more critical than the properties of the individual agents. With these ideas in mind, we turn now to an example that further illustrates the intricacies of complex, interconnected economic systems.

DISSECTING AN ECONOMY

In this section, we will investigate the interrelationships that exist in a simple economy. For our purposes, it will consist of four primary infrastructure elements: an

electrical grid, a natural gas distribution network, an oil pipeline system, and a communications network. Figure 4 sketches the appropriate elements.

The model economy is purposely simple so that the principle connections within it are clear. The elements themselves are notional, with many of their details omitted. In the electrical system, for example, the generator step-up transformers and distribution network layouts within the cities are deleted for clarity. (The Appendix contains more detailed information on electrical grids, POL distribution networks, communications systems, and transportation networks.)

The electrical grid has five main generators. Two generators are gas-fired. They connect directly to the gas pipeline and have no local gas storage. The other generators are oil-fired. These generation stations stock an average of two months fuel on site. A dispatch center with a backup site controls the electrical flows throughout the network. The control centers have backup electrical generators with three-week fuel supplies. A satellite communications network transmits data and control information between the elements of the electrical grid. Landline data and voice communications backup the satellite system. Note that Figure 4 only depicts the long-haul transmission system and generators. The notional distribution networks show the grid connections to the other infrastructure elements (such as gas compressors and oil pumps) that require electrical power.

The natural gas pipeline network consists of several main pipelines, numerous spur lines, two gas conditioning plants, and fifteen compressor stations. The compressors and related controls are electrically powered. Each compressor station connects directly to the electrical grid. The gas conditioning plants also tie into the electrical grid. A computerized control center with a backup control site runs the pipeline. Data communications between the control centers and compressor stations rely primarily upon satellite communications with landline data and voice backups. A computerized SCADA system runs the entire pipeline; safe manual operation is not feasible. Both control sites have backup electrical generators, with two-week supplies of fuel.

The oil pipeline system is similar to the gas pipeline network. The system has several main lines, two refineries, twenty eight pump stations, a centralized dispatch control center, and a backup dispatch control center. The pump stations and refineries require electricity in order to operate. The dispatch control center uses a SCADA system to manage network operations. Satellite communications relay data between the pump stations and the dispatch control center. Landline data and voice communications backup the satellite system. Each control center has a backup electrical generator, with a two-week fuel supply. As with the gas pipeline, the network cannot operate safely without its computerized SCADA system.

The communication network ties all of the other systems together. The system consists of a satellite and terrestrial network. It draws power directly from the electrical grid. However, the system has numerous backup generators, each with a two-week supply of fuel. Each major switching station, for example, has a backup generator. A computerized, central control site with a backup site manages all network operations. The network furnishes vital services to the gas and oil pipelines and electrical grid.

Figure 4 portrays a simplified picture of the real world. The model economy has only four elements; these systems are themselves simplified representations of their real counterparts. Nevertheless, the figure serves to underline the complex, highly interconnected nature of economic infrastructure elements. We shall now examine in detail the nature of this interconnectivity.

Nodes and Couplings

The systems of the model economy are coupled to one another in intricate manners. Table 1 lists several representative couplings between the elements. Tight coupling is characterized as having no slack, give, or buffering—it is highly time dependent. Loose coupling, on the other hand, has some degree of slack or give. It is not as time dependent as tight coupling. Note that tight, moderate, and loose couplings permeate our notional model.

Degree of Coupling

Agents

   

Tight

• Satellite terminals—electrical grid

 

• Gas-fired electrical generators—natural gas pipeline

 

• Natural gas pipeline compressors—electrical grid

 

• Oil pipeline pumps—electrical grid

   

Moderate

• Electrical grid dispatch control centers—communications network

 

• Natural gas pipeline dispatch control centers—communications network

 

• Oil pipeline dispatch control centers—communications network

 

• Natural gas pipeline dispatch control centers—electrical grid

 

• Oil pipeline dispatch control centers—electrical grid

 

• Communication system control centers—electrical grid

   

Loose

• Oil-fired electrical generators—oil pipelines

By examining Table 1, we see that tight, moderate and loose couplings permeate the economy. The gas-fired electrical generators–gas pipeline couplings fall under the tight category, as gas is not stored locally at the generators. If the gas pipelines shut down, then the generators drop off line—there is no slack or buffer in the couplings. On the other hand, if the fuel oil supplies to the oil-fired generators are interrupted for several weeks, the generators can continue operations by drawing down local fuel reserves. In this case, the coupling is loose. The various control centers have moderate couplings to the electrical grid, as each center has a backup generator. As long as the fuel supplies to the backup generators is uninterrupted, the control centers continue to operate indefinitely. Similarly, the control centers are moderately coupled to the communication systems. Each control center relies primarily upon satellite communications with landlines for backups. If either the satellite or terrestrial communication networks fails, the control centers will continue to operate. Only the loss of both communication networks will isolate the control centers. In these examples, the coupling degree indicates the time criticality of the linkages and the availability of backup systems.

The couplings listed in Table 1 thus link different economic infrastructure elements together. However, by examining the model further, we see that it has a hierarchical structure. The economy is composed of four elements (electricity, natural gas, oil, and communications), each element is composed of subsystems (the electrical network contains generators, control centers, substations, etc.), each subsystem is composed of further subsystems (electrical generators have fuel burners, boilers, steam turbines, etc.), and so on. Tight, intermediate, and loose couplings of agents exist within and between each level of the hierarchy. Consider the electrical grid. The dispatch control centers are moderately coupled to the generators, transformers, and substations. The loss of the control center and its backup would disrupt the operation of the grid. This illustrates that each agent, then, has links to other agents of the same hierarchical level, as well as possible links to agents at other levels. Clearly, the couplings between infrastructure elements of the model economy are complex. What makes them even more so are the hierarchies of couplings.

The couplings examined so far directly link elements (and agents) to each other. This form of coupling is direct, or first order. Indirect, or higher order, couplings are also present in the network. The gas pipeline network, for example, has a direct linkage to the communication network through its SCADA system. An indirect linkage connects the gas lines and communication system through the electrical network: natural gas fuels two generators which in turn provide electricity to the communication system. (Of course, the oil-fired generators also supply electricity to the electrical grid. This further complicates the example.) Additionally, these systems have several feedback loops. A major feedback loop connects gas-fired generators and the gas pipelines. If the gas supplies to the generators are interrupted, the generators drop off line. However, the generators provide electricity to the pipelines and compressors. In this manner, a feedback loop forms between the two elements. Even in this simple example, the interconnected nature of the agents is intricate.

All four elements have branching rather than sequential topologies. The branching layouts add redundancies to each of the networks. For example, if a given compressor station fails in the natural gas distribution system, the dispatch control center might be able to reroute gas flows through other pipelines. In this manner, the natural gas system adapts to internal changes or component failures. However, a highly interconnected system also tends to propagate perturbations throughout the entire system. Disturbances in one part of the system thus affect all other parts. If one of the five electrical generators fails, the effects will be felt throughout the entire electrical grid. The branching nature of the layouts further complicates the analysis of the four infrastructure elements.

The branching nature, higher order linkages, and feedback loops that comprise the model make it difficult to intuitively analyze the economic elements. Some of the interactions between the elements are hidden, or at least not obvious. Additionally, the higher order linkages and branches are nontrivial. Depending upon the configuration of a network and its environment, the links may have major impacts upon overall system operation. Consequently, the effects of losing or destroying an agent or a link are not clear. What is clear is that even our simple model requires some form of nodal analysis to determine the results of an attack. Furthermore, a reductionist analysis is inadequate to describe the synergistic effects that arise from the couplings. An analysis based on a nonreductive methodology is the only way to proceed.

A "real" economy is vastly more complicated than the simple example of this section. More than four elements will be present, a myriad of interconnections will tie the agents to one another, and the emergent behaviors will preclude a reductionist analysis. "Back of the envelope" calculations or "rules of thumb" may be sufficient to determine an initial, rough target list. However, these methods will fail when refining the target set required to obtain the commander’s objectives, or when projecting the overall results of an attack. As this paper will later illustrate, constructionist nodal analyses are necessary for a deeper understanding of the economy and the targeting process.

SUMMARY

Couplings and synergies abound in an economy. The interconnectivity goes far beyond the bottlenecks and chokepoints sought by the member of the ACTS faculty and WWII planners. The intertwined nature of an economy leads to emergent behaviors, which in turn demand a holistic approach to targeting. Every nation is a unique entity, and its particular socio-cultural character determines how it values its particular economy. These considerations must all factor into the nodal analyses, and ultimately, the master target list and campaign plan.

The notional economy in this chapter demonstrated that systems consisting of only a few interacting agents can be difficult to intuitively comprehend. Fortunately, a variety of software tools exist that allow a planner to explore the behavior of various economic infrastructure elements. These codes, mated with several recently developed numerical algorithms, hold significant promise as planning tools for the targeteer. We will delve into these possibilities in the next chapter.

CHAPTER 4

MODELING AND SIMULATION TECHNIQUES

Economic infrastructures are complex, interacting systems that continuously react and adapt to their environment. As we have seen, they may be composed of thousands of units, nonlinear linkages, and feedback loops. Consequently, intuitive estimates of the system-wide effects of an attack upon a given unit may contain significant error margins. "Rules of thumb" or "back of the envelope" analyses developed for individual, isolated target systems may be of little use, as they are generally derived from a reductionist rather than holistic view of an economy. In short, understanding and estimating the effects of attacks upon economic infrastructures is not a simple task due to the complex, interactive character of the problem.

Systems analysis techniques can contribute significantly to the understanding of the inner functioning of an economy. A variety of detailed numerical simulations exist that can be adapted to modeling attacks upon various infrastructure elements. These programs require databases containing information about a particular element. In principle, a planner could numerically simulate the effects of an attack upon an economy provided he or she had a sufficiently detailed database and the appropriate computer codes. Master attack plans, targets lists, and ultimately the air tasking order could be "fine tuned" with the aid of a computer.

At first glance, the use of a computer to determine an attack plan may appear to be risky at best. First, high fidelity physical models of the economy might be difficult to develop. (High fidelity generally equates to significant amounts of detailed data about the economy or its infrastructure.) An incomplete database will lead to approximations and possibly inaccuracies in a simulation. Even if the data is available, anchoring the model to a foreign economy may be impossible: one cannot attack a foreign state simply to validate a model. Second, if the infrastructure elements are tightly coupled, the computer models must be capable of analyzing all of the important elements. A reductionist rather than holistic approach would at best only approximate the results of an attack. Finally, by their very nature, modern economies present a considerable number of possible targets. With potentially thousands of targets, the task of computing a target list that accomplishes a particular military objective appears daunting. It also appears to circumvent human intuition, a critical element in war. Political and military constraints further complicate the numerical problem. Nevertheless, these problems are not insurmountable, especially in light of several recent developments in numerical techniques.

Two different computer algorithms developed during the past decade hold promise in overcoming computational problems in targeting. The first computer method, simulated annealing (SA), draws from statistical mechanics. As its name implies, it is patterned after the molecular processes of annealing and freezing. Genetics and natural selection provide the basis for the second numerical technique, the genetic algorithm (GA). Either of these methods could be readily coupled to existing economic models, thus providing powerful new tools for an air planner.

In this chapter, we will explore numerical simulations of economic infrastructures in more detail. The focus is on targeting, while the objective is to propose a new class of numerical tools that aid targeteers in their systems analyses without replacing their human judgment and intuition. In the following section, we will examine two diametrically different targeting philosophies. One philosophy, based on desired outcomes, is clearly superior to an input-based approach, which is narrow and short-term in focus. The remainder of the chapter develops numerical algorithms suited to the output-based philosophy. In particular, it presents details on three numerical techniques that are applicable to air targeting. We will pay particular attention to the relative merits and drawbacks of each technique, two of which (SA and GA) are suitable for targeting purposes. We will tie together all of the theoretical and practical aspects of targeting complex economies with a detailed numerical targeting algorithm of our own design. In this algorithm, a GA coupled to load-flow and hydraulic analysis programs provides specific targeting information for electrical grids and POL distribution networks. The couplings between these two networks complicate the targeting problem. The chapter then concludes with a philosophical discussion on the use of computers versus humans in the targeting process.

TARGETING PHILOSOPHIES

Two general targeting methodologies exist, each approaching the problem from opposite points of view. The first focuses upon inputs to the battle; it concentrates mechanically on the number of sorties and the ordnance delivered. The second approach is based upon outputs. In this technique, the selected targets flow from the commander’s intent and desired objectives. The input approach is easier and is perhaps employed more frequently. However, the output approach directly addresses the commander’s desires, and thus is the preferred method.

The input method is a relatively simpler approach to targeting. Given a particular scenario, this technique seeks the answers to how and how many questions: how should the enemy forces be attacked, and how many targets should be selected? In schematic form, the planner surveys the available targets, and selects some set to attack. If the results of the attack fulfill the commander’s objective, the campaign shifts its emphasis to the next objective. If the results do not meet the objective, the planner can either continue the attacks on the initial target set or abandon it in favor of another set. Figure 5 depicts the input method. In a sense, this method approaches the commander’s objectives from a backwards direction: the effort is focused on the inputs (the how and how many questions) with the goal being outputs that support the campaign objectives.

Computer modeling tools can aid the input-based targeting method in a straight-forward manner. Given an adequate computer model of the enemy target system, the planner numerically simulates the attack and examines the results. If the results support the commander’s objectives, the planner orders the attacks. Otherwise, he uses the simulation to modify the attacks. This approach parallels one type of engineering failure analyses, in which the consequences of the failure of some component or group of parts are determined. The engineer may modify the design or incorporate backup systems depending upon the analysis results. For example, examining a bridge design might reveal that the collapse of a strut would lead to the catastrophic loss of the bridge. The engineer might strengthen the strut or add additional structural members to prevent the collapse of the bridge. Similarly, systems analyses provide important information to the planner, and help him fine tune his attacks.

The input-based philosophy is limited in its utility, however. It is primarily tactical, as it starts from the how and how many questions, rather than addressing the larger question of what is accomplished by the attacks. In many cases, it is driven by the immediate battlefield situation, making it a reactive rather than proactive approach. The measures of merit become the numbers of sorties and bombs dropped, rather than what is achieved on the battlefield. Servicing targets becomes the watchword, instead of achieving results. To a large extent, the Army corps commanders pursued this philosophy during Operation Desert Storm. They wanted a fixed number of sorties (inputs) per day in their areas of responsibility. The important measure of merit became the number of sorties flown for battlefield preparation, rather than whether each corps could perform its scheme of maneuver when the ground war began. In sum, this targeting philosophy is narrow in focus and short-sighted in perspective.

The output-based philosophy avoids many of these problems. It essentially flows in the opposite direction of the input-based approach. In this method, the commander’s objectives direct certain results—the outputs—that must be accomplished. The desired results then determine the requisite attacks—the inputs. Figure 6 outlines this process. This approach places the primary emphasis on what questions: what must be achieved in order to meet the commander’s objectives? The strategic air campaign in Operation Desert Storm was designed with this philosophy, and became one of the major sources of disagreement with the army corps commanders, who were thinking in terms of the input-based paradigm. This output approach is superior to the input technique, as operational art is now at the forefront rather than tactics.

As with the input-based philosophy, computerized tools can provide invaluable information to the output-oriented planner. As will be shown below, SA and GA algorithms (coupled with system models and databases) can determine targets that yield the required outputs. The algorithms can be roughly thought of as performing a fault tree failure analysis: given a desired system-wide failure, what components will provide the necessary failure when destroyed? The process is analogous to determining what will cause a bridge to collapse, for example, rather than asking about the effects of destroying a single supporting pier. Clearly, tools that aid the planner in obtaining a desired output are a tremendous boost to the targeting process.

We must note that there is a difference in the way that targets are selected and the manner in which they are attacked. An air force can attack targets in either a serial fashion as in WWII, or in a parallel manner as in the Persian Gulf War. Notice, however, that the target selection procedure for serial or parallel attacks can be either input-based or output-based. The target selection procedure will not automatically determine how the attacks are executed.

In summary, a planner may follow one of two general targeting philosophies. The first is based on inputs, and is limited in scope and utility. The superior method focuses on the commander’s objectives and required results. In the following sections, we will develop several computer techniques that will aid the planner in executing this second targeting philosophy.

NUMERICAL TECHNIQUES

Targeting belongs to a class of problems known as combinatorial optimization. In this class, there is some function that must be minimized or maximized. It is alternatively known as the cost, objective, or fitness function. The class differs from the classical minimization and maximization problems in that there are no continuous variables to optimize upon. Rather, the problems contain a large number of discrete elements that may be varied. The number of different combinations of the variables is factorially large; hence, an exhaustive search for the best combination for a given cost function is clearly impossible. For example, if the problem has 100 discrete elements that can be in either of two states (100 targets that are either attacked or left unharmed), there are 2100 = 1.3 x 1030 different combinations or state configurations. A random or exhaustive search for even this small problem is impractical. Fortunately, a variety of techniques exist that efficiently determine an acceptable if not optimal solution.

The traveling salesman problem is an example of combinatorial optimization that has been extensively studied. The salesman must visit a number of cities while following the shortest path possible. The cost function is simply the total distance traveled by the salesman. Since he wishes to follow the shortest path possible, the salesman must minimize his cost function. He does this by varying the order in which he visits the cities. Hence, the cities themselves form the discrete elements in this problem; these elements are varied by changing their visitation order in the salesman’s itinerary. Furthermore, we can add constraints to the problem. For example, if there are toll bridges in the salesman’s region, we can require that the salesman avoid them if possible (in order to pay the minimum toll). The cost function is now the distance the salesman travels plus a penalty that is assessed for each toll bridge crossing. The constrained problem now seeks to minimize simultaneously the distance traveled by the salesman and the number of toll bridge crossings.

Targeting is a further example of combinatorial optimization. Here, the cost function indicates how well the targeteer meets the commander’s objectives, and includes any constraints, restraints, or rules of engagement (ROE). The individual targets are the discrete variables. Given potentially hundreds of targets, the planner must determine some combination thereof that will produce the desired results and hence the commander’s objectives. In general, the final target list should be minimized so that sorties are not wasted on redundant or frivolous strikes. Then, the air planner could use several of the computer algorithms developed to solve combinatorial optimization problems for targeting purposes.

The following subsections examine three numerical algorithms commonly used to solve combinatorial optimization problems. The methods are iterative improvement, SA, and GA. A detailed example which uses a GA to solve a targeting problem appears in the next section. Although any of the three methods could be applied to the targeting problem, SA and GA presently hold the highest potential.

 

Iterative Improvement

Iterative improvement is the simplest of the three algorithms. Figure 7 illustrates the technique for the minimization problem. The algorithm starts from an initial set of values for the discrete elements. These initial values are called the base configuration. The base configuration can be chosen at random; however, a judicious choice of the base near the minimum will speed convergence. The routine evaluates the cost function for the base configuration. Next, it applies a rearrangement operator to the base configuration and reevaluates the cost function. The routine iterates the procedure until some configuration appears with a lower cost. This new configuration and its cost replace the current base, and the process starts over. Eventually, the algorithm locates some optimum configuration literally by "walking" through the various rearrangements.

We can use the traveling salesman problem to illustrate the iterative improvement method. For this example, we will assume that there are no constraints (e.g., there are no toll bridges or toll roads to avoid, etc.). The cost function is then the total miles the salesman travels while visiting the cities. The base configuration is an initial itinerary for the salesman—an initial ordered list of the cities to be visited. A variety of rearrangement operators have been discussed in the literature. One such operator randomly removes a section of the path and replaces it with the order of the cities reversed. Another version removes a path section and inserts it between two other cities. Note that these two rearrangement operators do not completely randomize the order of the cities; some of the past history is retained after each rearrangement. Several different convergence criteria are possible. For example, if the average change of the cost function per iteration drops below some threshold, the routine will stop. Alternatively, if the routine applies the rearrangement operator more than some preset number of times without finding a lower cost configuration, the routine can declare convergence. The routine for the traveling salesman problem follows the flowchart in Figure 7.

Iterative improvement suffers from a serious drawback. Many cost functions are multipeaked, resembling a mountainous landscape with many crests and valleys. Figure 8 depicts a multipeaked cost function. A number of local minima exist, such as points A and B in the figure. However, only one global minimum exists; it is the particular configuration with the overall lowest cost (point C). In the mountainscape analogy, the global minimum corresponds to the deepest point in the lowest valley, and the local minima correspond to the floors of the other valleys. The iterative improvement technique tends to lock into a specific valley and determine its minimal value. This point may or may not be the global minimum. For some problems, this valley may contain a sufficient although not best solution. Consequently, although the iterative improvement scheme can locate local minima, it may not solve the particular problem at hand. For the traveling salesman problem, suppose that the dotted line in Figure 8 represents the highest acceptable cost. Then, point A is unacceptable whereas B and C are viable solutions. If the cost function is extremely rugged with many peaks and valleys, the iterative improvement technique will be useless unless it starts from an extremely good initial base configuration.

Several modifications to iterative improvement have been implemented. For example, the algorithm itself can be iterated over a range of initial configurations until it locates some "best" local minimum. Nevertheless, if the cost function is very rugged, even a modified version of the algorithm may not perform well.

In sum, iterative improvement is a simple method capable of converging to a local optimum configuration. It works well if the cost function has a single minimum. However, if several local minima exist, there is no a priori guarantee that the routine will converge to the global or even a near-global optimum.

In principle, we can develop a routine based on iterative improvement that solves the targeting problem. The routine would determine a target list that optimizes the cost function. Here, the cost function directly reflects the commander’s objectives. (We will delve into this point in depth in the next section.) However, since iterative improvement locks into a local but not necessarily global optimum, this technique does not hold significant promise for the targeting problem. A local optimum point may be insufficient to meet the commander’s requirements; it may in fact be worse than a solution generated "by hand." Therefore, we must look to other methods that can locate global (or near-global) optima for the targeting process.

Simulated Annealing

Simulated annealing was first proposed by Kirkpatrick et al. in 1983. Since then, it has undergone rapid development. SA has successfully solved a number of problems, including the traveling salesman problem. Press et al. examine the traveling salesman problem in detail, and provide a heavily documented FORTRAN source code for the problem with river crossing constraints. SA is an interesting technique that is analogous to annealing and freezing processes.

SA has its roots deep in statistical physics. The principle algorithm, termed the Metropolis method, was developed in 1953 for calculations of the material properties of collections of interacting molecules. Kirkpatrick and coworkers modified the Metropolis method in the first published description of SA. To understand the physical basis of the algorithm, consider a collection of molecules in a liquid. If the temperature decreases, the molecules slow down and begin to freeze or crystallize. While doing so, the molecules seek an arrangement with the minimum energy. Since the molecules are in constant motion just above the freezing point, the system can "explore" its entire configuration space and search out the minimum energy configuration. If the system is cooled slowly enough, the system selects and freezes itself into this arrangement. In other words, the system has optimized (minimized) its energy state. SA works in an almost identical fashion to optimize the cost function.

SA is essentially a version of the iterative improvement method that borrows the Metropolis algorithm. Figure 9 is a simplified flowchart of the SA algorithm for the minimization problem we just described. The method starts with a random initial base configuration at an initial "temperature" T. The algorithm computes the cost of the base configuration. The procedure applies a rearrangement operator to the configuration, and calculates the cost of the new configuration. If the change in cost DC < 0 (the new configuration has a lower cost), the algorithm accepts the new configuration and its cost as the base. However, if DC > 0, the new configuration is accepted with a Boltzmannian probability. That is, the algorithm selects a random number z between 0 and 1. The algorithm accepts the new configuration and its cost as its base if z < e-DC/kT where k is the Boltzmann constant, even though the new base cost is greater than the original. Otherwise, the routine retains the original base configuration. The algorithm applies the rearrangement operator to the configuration, and a new iteration begins. In this manner, the algorithm explores a range of states in the configuration space, as in the annealing process.

The ability to accept higher cost configurations is an important advantage of SA over iterative improvement. As noted above, iterative improvement routines will often lock into local minima. By accepting positive DC values, the SA routine can hop out of a local minimum and continue the search for the global minimum. This feature considerably increases the power of SA over iterative improvement.

The temperature T plays a crucial role in the algorithm. T is actually a control parameter, but it is analogous to the temperature in freezing and annealing processes. The SA method initially starts at a relatively high temperature T. Here, the routine can accept relatively large positive values of DC. It will not get trapped in valleys as does the iterative improvement technique. In this manner, it is free to explore large regions of the configuration space as it searches for the global minimum. This is precisely what our collection of molecules does in its liquid phase: it tests a large variety of arrangements as it searches for the minimum energy state. Upon discovering some promising region of configuration space, the algorithm decreases the temperature. As T decreases, progressively smaller positive values of DC are accepted with a high probability. In this stage, the algorithm tends to lock into a valley with either a near global or the actual global minimum. Additional reductions in T force the algorithm to seek the valley floor. Convergence to the solution occurs when further reductions in T result in insignificant reductions of the cost function. Essentially, the algorithm mimics the behavior of molecules undergoing a phase transition. In sum, varying the control parameter T is analogous to varying the temperature of our collection of molecules.

The initial value of T and the schedule for lowering it are critical to the performance of the algorithm. If T decreases too rapidly, the algorithm may lock into a region with a local but not global optimum. This is similar to rapidly freezing a liquid, which results in a glassy solid rather than a minimum-energy crystal. On the other hand, lowering T too slowly wastes computer time. Determining a good initial value of T and an appropriate schedule requires insight into the problem and/or trial-and-error experimentation.

SA has several additional advantages and disadvantages. On the positive side, the algorithm readily adapts to parallel processing. For a given T, a parallel version can analyze a number of configurations simultaneously. SA uses its past configuration history to improve the search for the global optimum, since the rearrangement operator does not completely randomize the test configurations. On the negative side, the temperature variation schedule is crucial to the proper operation of the algorithm. Numerical experimentation is probably necessary to determine good initial values of T and temperature schedules.

At this juncture, SA annealing appears to have promise for the targeting problem. Since SA does not lock into the first valley (and thus local minimum) that it locates, it is more robust than iterative improvement. Since the cost function for the targeting problem is likely to be quite rugged (especially if the economic infrastructure elements are very complex), the robustness of SA is a definite advantage. We shall explore this aspect in more detail in the example of the next section.

Genetic Algorithms

Practical genetic algorithms date back to 1975, when John Holland of the University of Michigan developed classifier routines that "evolved" solutions to optimization problems. As with SA, GAs have enjoyed almost explosive growth in recent years in both theoretical and practical terms. The method is particularly suited to applications in which there is no simple relationship between the system configuration and the cost function or where the solution literally must be discovered rather than calculated.

GAs belong to a subclass of artificial intelligence known as artificial life. In the GA literature, the cost function is called the fitness function or fitness landscape. In the natural world, the fitness function of an organism is a measure of its ability to survive in a given environment. Reproduction, exchange of genetic material, mutations, and natural selection change the genetic code of successive generations of the organism, either improving their positions on the fitness landscape or not. A GA uses the same basic processes to evolve optimal solutions to problems inside a computer. Like its organic counterparts, the GA creates "generations" of solutions that progressively move toward the global maximum of the fitness function. In solving the problem, the GA mimics naturally occurring biological processes—hence its inclusion in the artificial life class of algorithms.

A principle element of a GA is the gene string or genotype. The simplest and most general genotype occurs in the binary combinatorial optimization problem. Here, there are n discrete elements or variables, such as n potential targets. Let ai represent the ith element. Since the elements are binary, they can take only one of two values, 0 or 1 (on or off, attacked or not attacked, etc.). Concatenating the elements in a string yields a binary variable a1a2a3...an. Thus, every value of a1a2a3...an corresponds to a point in the configuration space. For example, 0111001011 and 1101010011 are two points in the configuration space of an n = 10 binary problem. In an analogy with the biological case, the string a1a2a3...an represents a chromosome (genotype), each bit position of the genotype corresponds to a gene, and each gene represents the state of a particular discrete element. Consequently, the genotype represents all of the possible system configurations.

A second key element of the GA is the fitness function. As noted above, the fitness function is merely the cost function. In essence, the fitness function is the embodiment of the problem at hand. As with iterative improvement and SA, care must be taken when designing a fitness function. In the targeting problem, the value of the fitness function denotes how well a given targeting solution (an individual) meets the commander’s requirements. If a solution comes close to satisfying his requirements, the fitness of the solution is high. On the other hand, a low fitness solution fails to meet the requirements. Clearly, the fitness function will change for every targeting scenario. In all scenarios, the GA will attempt to "evolve" high fitness targeting solutions.

The operation of a GA parallels the biological processes of selection, reproduction, and genetics. Figure 10 outlines the procedure. The algorithm begins by

creating an initial population of individuals. That is, the routine generates m values of a1a2a3...an, where m is some integer. Each individual is a trial solution to the optimization problem. The values may be randomly chosen, although this is not required. Any insight into the problem that is used when assigning initial values to the m individuals will probably help convergence. Once the program has created the population, it is ready to pass to the reproduction step. As the name implies, the reproduction step creates the next generation of individuals. First, the routine evaluates the fitness function for each individual. The fitnesses determine whether an individual survives to the next generation or dies out. Low fitness individuals have a low probability of survival; selection may preclude their replication in the next generation. On the other hand, high fitness individuals are more likely to survive. In fact, the routine may create multiple copies of high fitness individuals in the next generation. In this manner, selection probabilistically eliminates low fitness individuals and ensures that those with the highest fitness form the basis of the next generation. The average fitness of the successive generation is generally higher than that of the previous one.

Following reproduction and selection, the algorithm performs two crucial operations. The first is crossover, in which two individuals swap blocks of genetic material. In the simplest form, the routine arbitrarily selects two individuals for crossover. Next, a point in the genotype is randomly chosen, and all genes following this point are exchanged between the two individuals. This operation adds diversity to the population, and also ensures that the best gene combinations are preserved in successive generations. In the second operation, a mutation operator selects an individual at large and then randomly chooses and flips one of its bits. As a result, the operator prevents the algorithm from becoming stuck in some local optimum of the fitness function. Crossover and mutation are important steps that maintain the diversity of the population as well as allow the algorithm to sample large regions of the configuration space.

The above description sketches a highly simplified picture of the main steps of a GA. Researchers have modified this simple routine in many ways, adapting it to a variety of problems. The key concept is that a properly operating GA converges to a high average fitness population after a sufficient number of generations. The high fitness individuals represent points in configuration space that are close to the optimal solution of the fitness function. In Figure 10, after some number of generations, the GA would determine values of the sextuple (a1, a2,..., a6) that produce high fitness values. In this manner, the properly designed and tuned GA "evolves" a solution to the problem represented by the fitness function.

As with any numerical routine, GAs have their particular advantages and disadvantages. The positive aspects include relative robustness and applicability to a wide range of problems. GAs readily adapt to parallel processing. In fact, the algorithm is inherently parallel; even on serial computers a number of solutions are processed in parallel due to the nature of the algorithm. One of the principle drawbacks is the tuning requirements. A GA has a considerable number of parameters, such as the number m of individuals in the population, the selection probabilities, and the crossover and mutation rates. Often, successful convergence of a numerical routine depends upon the choice of such parameters. As with SA, the designer frequently has to vary the parameters in order to obtain a well-behaved, convergent GA. This may require trial-and-error tuning of the GA for each targeting scenario as well as the use of any insight into the problem. Computer time requirements are another drawback to GAs, as the routine must evaluate the fitness function of each individual during each generation. Many of the unanswered questions about GAs are the subjects of current theoretical and simulation projects.

On the whole, GAs appear to be promising for output-based targeting. Targeting itself adapts well to the GA, as it is an example of binary combinatorial optimization. Furthermore, the complex nature of economies will lead to emergent and often counterintuitive behavior. There may be no simple relationships between the destruction of specific targets in the economy and the global results. Consequently, the overall effects of attacking the economy cannot be adequately guessed a priori. From this viewpoint, the GA is particularly suited for the targeting problem and merits further exploration. The following section examines this proposition further, outlining in detail a GA that determines the optimal targets for an electrical grid coupled to a POL distribution network.

TARGETING ECONOMIC SECTORS: A PROPOSED NUMERICAL SIMULATION

GAs and SA have potential for computerized routines that aid the output-based targeting process. To illustrate their potential, we will now develop in detail a notional computerized targeting routine. The program uses a GA coupled to load-flow and hydraulic analyses of an electrical grid and POL network, respectively. In principle, the program could use an SA driver routine instead of a GA. Note that the program described below has not yet been coded and tested; however, an air planner could adapt existing computer tools to create a targeting program based on the example.

Problem Description

The problem centers on the notional electrical and POL networks of some hypothetical country. Following the output-based targeting philosophy, the friendly commander has decided that certain sectors of the electrical grid and POL networks must be destroyed. Their elimination will hamper enemy efforts: integrated air defense systems and communication networks will suffer from power outages, electrified rail transportation for mobilization will shut down, motorized transportation will be hindered from the loss of POL resources, and so forth. The adversary can use backup electrical power generation and stockpiles of POL to overcome some of the immediate losses of economic resources. However, we are also interested in the synergistic effects that arise from the couplings between the networks.

In more concrete terms, we assume that the commander has decided that the electricity and POL pipelines must be shut down in the eastern half of the adversary nation. To facilitate reconstruction efforts after the conflict, those elements targeted for destruction must be reparable within six months. This restriction eliminates certain potential targets, such as generators and their step-up transformers. Furthermore, we assume that the ROEs constrain the attack sorties to the eastern third of the nation.

The problem thus poses objectives as well as several constraints. The fitness function must include all of these considerations. Before turning to the GA and its implementation, a description of the electrical grid and POL network analysis programs is in order.

Electrical Grid Load-Flow Analysis

The electrical power industry has developed a variety of computer tools for analyzing electrical grids. The type of study most applicable to targeting is the load-flow analysis. The main inputs to the studies include line impedances and normal operating conditions. The primary outputs of load-flow studies are the real and reactive power flows through the grid and the bus voltages and phase angles. Table 2 lists typical program inputs.

A typical contingency analysis employs a load-flow program to determine the effects of the loss of one or more grid elements. By direct analogy, a planner could estimate the effects of air attacks upon certain components of an electrical network with a load-flow program. By selectively eliminating components from a grid database, the planner would simulate the effects of the attack. Any grid components that exceed their emergency power ratings would in practice drop off the grid due to open breakers and other protective circuitry. If the load-flow program displays the components exceeding their emergency ratings as well as the targeted elements, the planner can directly observe the effects of the attack. This type of analysis yields important first-order information for targeting an electrical grid.

Grid Component

Required Data

   

Power Lines

• ID numbers of the buses to which the lines connect

 

• Line resistance and reactance

 

• Total line charging

   

Transformers

• ID numbers of the buses to which the transformers are connected

 

• Transformer resistance and reactance

 

• Total charging

   

Buses

• Bus names and ID numbers

 

• Bus type (swing or regulated)

 

• Bus-voltage magnitudes

 

• Bus phase angles

 

• Desired generation in megawatts and megavars

 

• Upper and lower limits of megavar generation

 

• Load in megawatts and megavars

 

• Total megavars of static capacitors and reactors on the bus

   

All

• Emergency power ratings

POL Network Hydraulic Analysis

To assess the effects of the attacks on the POL network, we must perform a hydraulic analysis. The pipeline industry uses hydraulic analyses to design pipeline systems. Given the network layout and the geographic elevations along the pipeline, a hydraulic analysis computes fluid flow rates through the system. Table 3 lists typical inputs for a hydraulic analysis. In particular, the key input required by the analysis is the elevation data along the entire length of the pipeline.

Network Component

Required Data

   

Terrain Geography

• Geographic elevations along the entire pipeline network

   

Network Layout

• Locations and layouts of pipes, pumps, fittings, etc.

   

Pipes

• Diameters

 

• Wall friction head losses

 

• Type of construction material (i.e., steel or iron)

 

• Safe and maximum working pressures

   

Fittings

• Head losses

   

Pump (Compressor) Stations

• Number of available pumps (compressors)—number of pumps (compressors) in use

 

• Net available head (pressure) per pump (compressor)—normal and emergency values

 

• Pumping rates (i.e., barrels per hour or cubic feet per hour)—normal and emergency values

 

• Suction pressures

 

• Pump (compressor) power source (i.e., electric motor or diesel engine)

   

Fluid or Gas

• Types of fluids (gases) normally pumped through the system

 

• Specific gravities (fluids)

 

• Pressures and temperatures (gases)

For the air planner, this analysis indicates the extent to which air attacks would reduce POL flow rates in the network. As with a load-flow analysis, the planner selectively eliminates pipeline elements from the network database. These elements represent the attacked targets. As components of the pipeline are destroyed, the pipeline operators will reroute the POL flows through undamaged sections. The analysis of the post-attack pipeline network indicates the new flow routes and rates that the damaged system can safely sustain. Thus, the planner directly observes the effects of his planned strikes in the outputs of the hydraulic analysis. It provides invaluable information for developing air attacks against a POL network.

The Database

Tables 2 and 3 indicate the types of physical information required by load-flow and hydraulic analysis programs. Before a planner can analyze any network, he or she must have a database containing the required inputs to the programs. The database is an extremely important part of the simulation. The more detailed the database information, the closer the simulation results will match reality. There is a balance point, however, at which the amount of database detail is sufficient for the planner. The amount of intelligence required to obtain higher detail may not be justified by marginal improvements in the accuracy of the results. Determining the balance point requires physical insight into the systems being modeled. In any case, successful use of the targeting routine relies upon the existence of an adequately detailed database.

By examining the two tables further, we note that a significant amount of the data requires geographic information. For example, the elevations along the pipelines are essential inputs to the hydraulic analysis. Transmission lines link geographically separated buses; hence, they form a tie between two spatial points. More importantly, displaying the results of the load-flow and hydraulic analyses on a map requires the geographic location of each item in the database. This form of geographic display is crucial to the commander’s situational awareness of the impacts of a proposed attack. The requirements for geographic data motivate us to employ a Geographic Information System (GIS) and geographic database.

GIS is a computer technology that links objects in a database via their geographic coordinates. It is much more, though, than a set of map coordinates appended to each database object. The power of GIS lies in its ability to perform spatial queries of the database. That is, a user may pose questions based upon geography to the GIS. For our example, such questions include:

• What infrastructure elements require electrical power within a 50 mile radius of a given electrical generator?

• Which customers will lose their oil supply with the destruction of a given section of POL pipeline?

• Which electrical generators do not lie in the restricted flight zone, and what happens to the POL network if they are destroyed?

Furthermore, a GIS capability allows us to model and simulate different elements by using the geographically linked database objects. This is precisely what our load-flow and hydraulic analyses perform. In short, a GIS is an enabling technology which we should employ with our database.

Although database development might at first appear to be overwhelming, there are considerations that may simplify the task to some degree. First, much of the data necessary to develop the initial network topologies is commercially available. For example, commercial databases exist that give general network layouts for the worldwide oil industry. Second, in some industries, there are few suppliers of critical heavy equipment. Consequently, some equipment may be standardized to a certain degree. (Electrical generators, for example, are manufactured by only a half dozen firms worldwide.) Finally, developing nations are following the U.S. lead in some infrastructure technologies. The telecommunications networks of some nations are based upon those in the U.S.—their equipment and protocols parallel those used in the U.S. in some instances. Nevertheless, the development and maintenance of adequately detailed databases is a large project that cannot be neglected.

 

The Genotype

Our notional targeting problem is an example of binary combinatorial optimization. The electrical grid and POL network consist of n components (lines, buses, transformers, generators, pipeline segments, compressor and pump stations, etc.), where n is some large integer. Each component can be in one of two states: targeted (and assumed destroyed during an attack) or untargeted. If the variable ai is the state of the ith component, then it takes one of two values, 0 for untargeted (undamaged) and 1 for targeted (destroyed). The state of the entire electrical grid and POL network is then represented by the genotype a1a2a3...an. The genotype takes the particularly simple form of a binary variable.

The genotype will be long if the number of components n is large. However, if the number of targets that can be attacked is limited (by total available aircraft, munitions, etc.) and is much less than n, then the genotype will be sparse. Compression of the genotype information will reduce the storage requirements, especially if the population size m is large. For example, in a very sparse genotype, it is only necessary to store a set of pointers that indicate which components are targeted, rather than storing information about each component.

The Fitness Function

The fitness function f is arguably the most important part of the routine. It is the embodiment of the targeting problem, and as such must incorporate the commander’s objectives and all constraints and restraints. As mentioned above, considerable care must go into its development.

Each electrical grid and POL network component has an associated set of rewards and penalties. In keeping with the commander’s desires, every electrical grid and POL network component in the eastern half of the country that shuts down as a result of the attack accrues a positive reward. Likewise, every eastern transmission line or pipeline that is still operational after the attack incurs a negative penalty. Any targeted facilities in the western two-thirds of the nation will also incur a negative penalty. Note that there is no penalty for components still running in the western half of the nation. Some components may be weighted more heavily than others, depending upon the commander’s wishes. For example, if the commander determines that destroying the electrical grid is more important than shutting off the POL flow, the electrical grid rewards would be correspondingly higher than those for the POL network. Note that the values of the penalties and rewards may require tuning to improve the convergence of the GA. However, any tuning of the parameters must be in consonance with the commander’s desires. Furthermore, any change in the commander’s desires will force changes in the weights. Table 4 lists the rewards and penalties for our particular problem.

Each component, then, has an associated set of weights. The weights form a vector; for this example, the weight vector is (ri, si, ti, ui, vi, wi). Using the weights from the table, a destroyed electrical grid component on the eastern border of the country with a repair time of 2 years (such as a generator step-up transformer) would have (100, 0, 0, 0, -25, 0) as its vector. A destroyed pump station with a 4 month repair time located in the restricted flight zone but in the eastern half of the country would be characterized by (0, 50, 0, 0, 0, -100). If the same pump station is unattacked but nevertheless shut down, its vector becomes (0, 50, 0, 0, 0, 0). Similarly, every component in the database has a weight vector.

Variable

Type

Weight

Description

       

ri

reward

100

Electrical grid component in the eastern region that is shut down

si

reward

50

POL component in the eastern region that is shut down

ti

penalty

-80

Electrical grid component in the eastern region that is running

ui

penalty

-40

POL component in the eastern region that is running

vi

penalty

-25

Attacked component has a repair time > 6 months

wi

penalty

-100

Attacked component is in the restricted flight zone

The fitness function is given by the sum of the rewards and penalties over all grid elements:

The maximum value fmax is simply the sum of the rewards, (ri + si), which occurs when all components in the eastern half of the country are down, and no constraints have been violated. With this particular fitness function, the program must attempt to find a target set that maximizes f. A reasonable criteria that a good solution must satisfy is f = afmax, where a is somewhat less than 1.

Program Logic

The program employs a genetic algorithm to evolve targeting solutions that meet the commander's requirements. Figure 11 is a block diagram of the program logic. The program follows the general flow of a GA. The synergies are incorporated during the load-flow and hydraulic analyses. The manner in which the synergies are incorporated merits deeper discussion.

The nodal analysis begins after the routine generates the population. Each individual in the population pool is, in essence, an attack plan. The value of the genotype indicates which components are attacked or bypassed. The program uses an iterative technique to determine the synergistic results of the attack. First, the routine simulates the attack by "removing" any targeted components from the database. The result is a "post-attack" database used in the ensuing nodal analyses. This database reflects the state of the economy after the attack.

Second, the routine performs separate nodal analyses of the two elements. In this step, the program analyzes each element in isolation from the other. In essence, the program calculates the effects of the damage on each element without regard for any synergistic couplings. The analyses determine the components that shut down due to the attacks. Any such electrical grid or POL pipeline component is removed from the post-attack database. The database now contains only those components that are still functioning in the isolated economic elements.

Third, the routine reconciles the effects of the couplings between the two elements. For example, if electricity is lost to a substation that feeds a pipeline pump, the pump ceases to function. Although the pump was not directly attacked, the loss of electricity causes the pump failure. The routine then removes the pump from the post-attack database. Similarly, if the natural gas pipeline feeding a gas-fired electrical generator shuts down, the electrical generator drops off line. The program reflects the loss of the generator by removing it from the post-attack database. At this point, the program has removed any components that either were destroyed in the attack, "failed" during the isolated nodal analyses, or shut down due to synergistic couplings.

Fourth, the program repeats the nodal analysis-reconciliation steps. We noted above that the routine removed components from the database during the nodal analysis and reconciliation steps. The program must now include these losses in assessing the results of the attacks. Therefore, it iterates the nodal analysis-reconciliation steps. In essence, the program calculates the cascading failures within and between the two elements of the model during this step.

Eventually, the routine will converge to a post-attack database that undergoes no further changes. This database represents the final operating state of the model after the attacks. It includes the results of the synergistic couplings and cascading failures. The last nodal analysis yields the final state to which the coupled economic elements deteriorate. The GA evaluates the fitness of this final state.

The preceding steps describe the nodal analysis performed on a single individual in the population pool. The GA repeats the analysis for the other m - 1 individuals in the population. The result is a set of fitness values that indicate how well the targeting solutions represented by the m individuals meet the commander’s desires. The GA then determines the next population iteration in the manner described in the previous section.

Discussion

This section has taken a top-level view of one form of a targeting program. Numerous modifications to the proposed program are possible: use of SA instead of a GA as the main driver routine, different fitness functions, different stopping criteria, and so forth. Indeed, a considerable amount of numerical experimentation might be required to determine the driver routine with the best convergence properties, reasonable fitness functions, the required level of database detail, and so forth. In any case, at this juncture, there appears to be several plausible approaches to solving the output-based targeting problem numerically with readily available modeling tools.

This particular example has focused on an electrical grid coupled to a POL network. Although the example contains only two coupled elements, it is a step up from the traditional reductionist approach. We have seen that systems do not exist in isolation and must not be treated as such. This becomes particularly important if the coupling between different systems is tight. The addition of other infrastructure elements may require a reformulation of the fitness function, changes in the fitness function weights, increases in the database size and genotype length, and more computations to determine the results of the attacks. In principle, the extension of our model to include other target sets is straightforward, with a program logic parallel to that described in this section. This type of multiple system simulation and planning tool would be extremely valuable to the planner.

PHILOSOPHICAL ASPECTS OF COMPUTER-BASED TARGETING

A philosophical dimension of computer-based targeting is the wisdom of turning the targeting function over to a machine. Basic Air Force doctrine states that "war is not an engineering project and must not be treated as such." However, in this instance, the machine is augmenting the commander—it is not replacing his judgment and intuition.

In any targeting problem, the planner must analyze his opponent. If a physical system such as a POL distribution network is the object of the attack, an engineering analysis is clearly called for. Whether done by a human or machine is immaterial; a well-designed program follows the same engineering analysis steps as a human with pencil and paper. The advantage of the machine is that it can examine more attack options at a much faster rate, and quite possibly find solutions that the analyst would miss. Furthermore, since infrastructure targets are complex, dynamical systems, their behavior under attack might be counterintuitive. What the planner might dismiss as an unreasonable system behavior might be the actual dynamical response to the attack. A program would not have the same reaction to strange dynamic behavior. In short, the computer performs the same analysis as its human counterpart, but at a faster rate, covering more options, and remaining unbiased by unusual results.

In all cases, the commander makes the final decision to execute, modify, or discard the proposed attack. The computer is simply a tool, neither designed nor destined to replace the human element. The commander is the final authority and must remain so, as war is truly a human enterprise.

 

SUMMARY

In this chapter, we have explored modeling and simulation tools applicable to the targeting problem. Two general approaches to targeting are possible. One is input-based, concerned more with the how and how many questions. It is tactical in focus, and of limited scope and utility. The more appropriate methodology is output-based, where the emphasis is placed on what must be done to obtain the commander’s objectives. This approach brings operational art to the forefront. It is the preferred approach to the targeting problem.

A variety of modeling and systems analysis tools exist that can be adapted to targeting. Under the approximation that each target on the list is destroyed in the attack, targeting becomes a binary combinatorial optimization problem. Two recently developed techniques, simulated annealing and genetic algorithms, appear applicable to numerical targeting programs. A GA formed the basis of a proposed targeting program described in this chapter, although SA could probably be used as well. Until such programs are developed and tested, the better algorithm choice will remain uncertain.

Numerical simulations and automated targeting programs do not remain far in the future. The tools exist today to develop such programs. In light of the recent Desert Storm experiences, we would be wise to do so.

CHAPTER 5

CONCLUSIONS

Complex systems abound in the world. They consist of collections of interconnected parts that interact with each other and their environment. We find examples of complex systems in many diverse disciplines, such as physics, chemistry, biology, and sociology. Of particular relevance to this study is that economies are complex systems. A little reflection shows that various economic infrastructure elements (such as electricity, POL, telecommunications, and transportation) are linked to one another. In many ways, each element is dependent upon the others and cannot exist in isolation. Communication systems cannot operate without electricity, POL distribution networks require communications to function properly, and so forth.

All complex systems exhibit several common traits. These characteristics include emergent behavior, adaptive self-organization, evolution toward the edge of chaos, and the ability to process and act upon information. The behavior of a complex system depends critically upon the linkages between its individual components. In fact, system-wide behavior generally cannot be deduced from an analysis of the component parts. Consequently, understanding system behavior requires more than a reductionist analysis of the individual parts—it requires a holistic analysis that incorporates the interactions between the parts. As economies are complex systems, we expect the same to be true of them.

Economies should exhibit the properties of complex systems. They contain various interconnections that act synergistically upon each other. In fact, the overall behavior of an economy may be difficult to predict due to the branching, nonlinear linkages between its elements. Most importantly, perfect knowledge of the various elements in isolation will not be sufficient to understand and predict the overall behavior of the economy. This reductionist approach overlooks the interconnections between elements and their influences upon economic behavior.

Because of the complex nature of economies, air planners require new forms of analyses for targeting. Nodal analyses, such as load-flow and hydraulic studies of electrical grids and POL networks, will greatly aid the planners. However, their utility is limited if the planners perform the analyses on isolated infrastructure elements. Rather, a targeting methodology that incorporates the linkages between the elements is needed.

Such holistic planning tools are at hand today. Indeed, we saw in Chapter 4 that simulated annealing or genetic algorithms can form the basis of a complex targeting algorithm. In particular, we examined a genetic algorithm coupled to load-flow and hydraulic analyses that should be capable of analyzing electrical grids linked to POL networks. The particular example preserves the ties between the two elements, and thus avoids the problems of reductionism. Although this new technique was only applied to electrical grids and POL networks, it is in principle extendible to any number of coupled infrastructure elements. Such analysis tools are sorely lacking today—yet the technology exists to begin their development.

RECOMMENDATIONS

1. This thesis is principally theoretical. It builds a framework for economic targeting upon complexity theory. However, "experimental verification" of the theory is lacking. Not only must we show that synergies and interconnections exist in economies, we must demonstrate that they have significant impacts on how economies function. Natural disasters provide one means of verifying the effects of the couplings. For example, the 17 January 1994 Los Angeles earthquake caused widespread power outages in central and southern California. However, the highly interconnected nature of the electrical grid led to power disruptions as far away as Montana, Washington, Oregon, Idaho, and Utah. Strikes and other disturbances may hold key data as well. Finally, a detailed holistic analysis of wartime economies impacted by aerial bombardment will provide the best evidence for the validity of the theory.

2. The GA–load-flow–hydraulic analysis program described in Chapter 4 has not been written. It currently is a paper concept. If coded and validated, the program would provide an invaluable tool for air planners. Furthermore, a modified version of the code could dramatically increase the realism of wargames, as the "actual" effects of bombing strikes could be incorporated into the games. The program should be written, validated, and incorporated into Air Force planning processes and wargames.

3. Finally, if the military is to pursue nodal analysis, it must possess sufficiently detailed databases on the infrastructure elements of potential enemies. As noted in the previous chapters, the databases will require a significant amount of effort to develop and validate. The time to start developing the databases is now, during peacetime, rather than after the commencement of hostilities with some aggressor. Without adequate databases, the best nodal analysis tools are useless.

FINAL REMARKS

Targeting science does indeed stand at the edge of a divide. In light of the complex nature of economies, the old reductionist approaches to targeting must give way to holistic methods. Within reach are nodal analysis tools that will allow planners to make this transition. We need a concerted effort to develop the computer programs and requisite databases. The payoffs are great: smarter targeting, greater system-wide effects for the same number of weapons, and fewer aircrews and aircraft at risk. As these tools were notably absent in our last major conflict, the Persian Gulf War, we would be sorely remiss to enter the next war without them.

APPENDIX

SYNERGY TABLES

In Chapter 3, we developed a simple model of a coupled economy. The economy consisted of four interconnected infrastructure elements: an electrical grid, a natural gas and an oil distribution system, and a telecommunications network. The example was deliberately simple, so that the interconnections between the elements would be readily apparent. Even so, many of the characteristics of a "real" economy were present: feedback loops, higher order connections, tight and loose couplings, branching processes, and so forth.

This appendix further explores the interconnected nature of economies. It focuses on four elements: electric grids, POL distribution networks, telecommunications systems, and transportation networks. Before examining the interfaces between the elements in the following tables, several notes are required:

1. The tables are not intended to be either checklists or exhaustive summaries. Rather, their purpose is to illustrate the type of synergies that are present in economies, and demonstrate their complex, interconnected nature.

2. The majority of the data relates to U.S. systems. Foreign systems may be completely different in structure, organization, and function. Consequently, the data below are only illustrative of the types of synergies that could exist. They do not reflect the nature of the synergies that will exist in every country.

3. Each country must be examined individually to determine the relevant synergies. Furthermore, these analyses must take into consideration the socio-cultural character of the nation, which will influence the forms, processes, and linkages of the economic infrastructure elements. The planner must avoid mirror-imaging U.S. synergies onto foreign nations.

4. The synergies in the tables are generic. Not every power plant is designed and built alike; each pipeline will have its own set of characteristics. As a result, the planner must tailor his analyses to the exact system in question.

5. Many foreign nations lag behind the U.S. technologically. As they upgrade their infrastructures, many use the U.S. as a model or purchase U.S. equipment. In these cases, there will be direct correlations between the U.S. and foreign infrastructures. Nevertheless, the usual caveats against mirror-imaging apply.

6. Only economic rather than military synergies and effects are addressed.

The tables focus on the connections between the four target sets. For example, Table 5, Electrical Grids, concentrates on how the electrical grid is tied to the POL, telecommunications, and transportation networks. More specifically, the first entry in this table shows how and why natural gas-fired generators are tightly coupled to the natural gas distribution network. There is no attempt in the tables to show the connections within a given target set (i.e., the fact that communication control centers rely upon communication networks is not spelled out in Table 7). Thus, the tables examine the interfaces between the elements and display the complex intertwined nature of economies. As a natural consequence, the tables further the compelling arguments for holistic rather than reductionist approaches to economic targeting.

 

 

Table 5. Electrical Grids

Infrastructure Elements

Linkages

   

POL

Natural gas-fired generators

• The generators require connections to the natural gas pipeline

• Gas regulating yards connect the pipeline to the plant

• Local storage of the gas is generally nonexistent ("storage is in the pipeline")

   
 

Oil-fired generators

• The generators require connections to the oil pipeline

• Fuel oil is stored in tanks on site

• 1-3 months of fuel is normally stocked at a generator

   
 

Gas turbine generators

• May use either natural gas or some other liquid/gas fuel

• The generators require connections to the pipelines

• Site may have no local fuel storage

   
 

Fuel for backup generators

• Electrical generator sites themselves have smaller backup generators for restarts, safely spinning down turbines in an emergency, etc.

• Diesel, gasoline, or propane engines may power the backup generators

• Dispatch control centers have backup generators to guard against power failures

• Backup generators require fuel to operate

• Local fuel storage is usually from a few days to a few weeks

   

Communications

Dispatch control centers

• Provide system-level control functions

• Require communication links to the generators and field sites

• Centers are important for balancing power interchanges between areas

• Centers are important for "blackout restarts"

• Centers will have redundant communication systems (i.e., satcom, fiber optic networks, voice communications, radios)

• Backup dispatch centers require comm links to field sites and main dispatch control centers

   

Transportation

Coal-fired generators

• Railroads supply the coal to the generators—vulnerabilities include rolling stock, tracks, bridges, etc.

• 3 months coal supplies are typically kept at sites in open storage areas

   
 

Nuclear generators

• Operations involve the shipment of nuclear fuels and waste materials

• Fuel is not stockpiled on the generation sites; it is kept in the reactors

   
 

Generators consume additional products (H2, water treatment chemicals, etc.)—the sites rely upon the transportation network for deliveries

   
 

Site Repairs

• Transportation network delivers equipment, material, and crews to damaged sites

• Transportation modes can be critical in remote, inaccessible areas

• For certain equipment, transportation is critical. (For example, generator step-up transformers are large and very heavy; only about 20 railcars exist in the U.S. that can transport these devices. In addition, delivery routes for these transformers must be carefully planned due to the high weights of the devices.)

   
 

Fuel deliveries for backup generators (at generator sites, dispatch control centers, etc.)

   

 

Table 6. POL Distribution Networks

Infrastructure Elements

Linkages

   

Electrical Grid

Oil fields

• Electrical pumps and motors

• Electrical controls on the machinery

   
 

Refineries

• Refineries employ electrical control systems, some electrically-powered rotating machinery (pumps, compressors, etc.)

• Refinery electrical requirements vary from site to site; may be in the range of 10-100 MW

• Some refineries use on-site electrical generation; others are directly connected to the electrical grid

• Attacking power would shut down the refinery, but the outage duration would vary from one refinery to the next

• Refinery could use emergency generators in some cases (i.e., 2-3 MW locomotives)

• Power transformers and substations on plant sites are easy to identify and might be very difficult to replace if destroyed

   
 

Pump and Compressor Stations

• Electricity powers pumps, compressors, valves, manifolds, controls, etc. (Note: some sites use diesel or natural gas engines to power the compressors and pumps.)

• Sites may have backup generators available

• Loss of one or two pumps or compressors will reduce flow rates through a pipeline, but perhaps not cripple it

• Pump and compressor stations maintain inventories of critical parts, including those with long lead times

   
 

Storage Sites

• Electricity powers pumps, valves, manifolds, controls, etc.

• Workarounds to power loss may be possible, depending upon power requirements and availability of backup generation

   
 

Distribution (Bulk) Terminals

• Terminals typically use electrically operated pumps, valves, manifolds, controls, etc.

• Workarounds to power loss may be possible, depending upon power requirements and availability of backup generation

   
 

Natural Gas Liquid (NGL) Plants

• These plants convert natural gas from gas to liquid or vice-versa

• Require electricity for compressors, coolers, controls, etc.

   
 

Pipeline Controls

• Controls are electromechanical (relays) or solid state

• The control network ties together all the elements of the pipeline system

• Pipelines rely upon computerized SCADA system for control and management functions

•• SCADA systems used to control all operations

•• SCADA system transmits information between dispatch control centers and remote terminal units (RTU) at pipeline facilities

• Manual workarounds to loss of SCADA, electromechanical controls might be extremely difficult to carry out

   
 

Dispatch Control Center

• Control centers and backup sites are heavily computerized

• Centers rely upon SCADA to control all operations

• Backups to power losses include standby generators, alternate control site

• Control centers cannot function without electricity; pipelines (in the U.S.) generally cannot function (or function safely) without their computerized controls

   

Communications

Pipeline Controls

• SCADA requires continuous communications between the RTUs and the dispatch control centers

• Control systems will use satcom and/or landlines for communications between RTUs and control centers

• Control systems will have communication backups (i.e., satcom primary system with landline and/or voice backups)

• Controls and communications are essential for the safe and efficient operation of a pipeline

   
 

Dispatch Control Center

• Linked to remote sites via SCADA and communications systems

• Controls and communications are essential for the safe and efficient operation of the pipeline

   
 

Production Control

• Communications systems are needed in the day-to-day management operations of the pipeline firms: placements of orders, dispatch control, coordination of repairs, etc.

   

Transportation

Water Transportation of POL Products

• Ocean shipment of crude is the primary means that many countries use to obtain their oil. Ocean shipment requires:

•• Port facilities

•• Ocean terminals

•• Oil tankers

•• Offshore unloading sites

• Port facilities are vulnerable to attack; such attacks could cripple a country if it imports the majority of its oil

• Inland water transport is often used; it requires port facilities, barges, offloading sites, etc.

   
 

Rail Shipment of POL Products

• Relatively small loads shipped via rail (vis-à-vis pipelines)

•• Much NGL is shipped via rail in the eastern U.S.

• Importance of rail shipments must be weighed against other transport methods when targeting oil

   
 

Truck Shipments of POL Products

• Forms the bulk of the local distribution system

• Trucking gets POL products from the "end of the pipeline" to the retail distributors and users (i.e., gasoline stations, aircraft at airports, home heating oil, etc.)

• Trucking is the primary means of supplying fuel to backup generators

   
 

Repairs and Inspections

• Transportation networks are critical for the delivery of equipment, material, and crews to damaged sites

• Pipelines are often sited near transportation networks to facilitate repairs

• However, pipelines may also be sited in remote, difficult to reach areas (jungles, deserts, mountains, etc.)

• Availability of transportation be critical in remote areas or for shipment of large equipment

•• Remote repairs may be impossible if adequate transportation does not exist

•• Workarounds to attacks in remote areas may be very difficult due to lack of transportation facilities

• Pipelines are inspected at regular intervals for leaks, damage, etc. Inspection means include:

•• Overflight

•• Walking the line

• Inspections become more important during a war

   

Table 7. Communication Networks

Infrastructure Elements

Linkages

   

Electrical Grid

Primary source of power for communications and computer equipment

• Includes switching sites, points of presence, regeneration stations, etc.

• Important sites will generally have backup generators and battery systems

   
 

Control Centers

• Control center is essential for network operations

• Centers are highly computerized

• They require electricity for operation

• Centers generally have a backup control site and standby generators

• Backup control sites are also highly computerized and require electricity to operate

   

POL

Fuels for the backup generators

   

Transportation

Repairs and Inspections of Facilities

• Transportation networks are critical for the delivery of equipment, material, and crews to damaged sites

• Availability of transportation becomes critical in remote areas or for shipment of large equipment

•• Remote repairs may be impossible if adequate transportation does not exist

•• Workarounds to attacks in remote areas may be very difficult due to lack of transportation facilities

• Inspection of lines and facilities

•• Aerial inspections are often used

   
 

Fuel deliveries to backup generators—relies upon local truck distribution of POL products

   

Table 8. Transportation Systems

Infrastructure Elements

Linkages

   

Electrical Grid

Airports—require electrical power for many facilities and systems, such as:

• Air traffic control, radars, communication equipment

• Runway lighting

• Power for support facilities

   
 

Railroads

• Primary source of power for some types of rail systems:

•• Electrified railways

•• Subways

•• Streetcars

• Electricity is required to operate many of the components of a railway, such as:

•• Signals

•• Switches, marshaling yards

•• Computerized controls

•• Control centers and their backup sites

   
 

Motor traffic (trucking, automobiles, etc.)—electrical power is required for:

• Signalization

• Fuel pumps at service stations

   
 

Miscellaneous "transportation" modes that require electricity in order to operate include:

• Elevators

• Escalators

   

POL

The fuel and lubricants for aircraft, diesel locomotives, boats, motor transport, etc.

   
 

The fuel for backup generators at control sites, radar sites, etc.

   

Communications

Operations at centralized control sites depend upon communication networks

• Sites include railroad control centers, air traffic control centers, etc.

• Integration of operations over large transportation networks demand reliable communications (such as air traffic control)

   
 

Signals control

• Railroad signals, especially if they are operated from a centralized control station

• Highway and traffic signals

   
 

Railroad switch controls, especially if they are operated from a centralized control station

   

Transportation

Intermodal ties

• These ties are the connection points between different means of transportation; they include:

•• Ports: links between water transportation and rail/trucking/pipelines

•• Marshaling yards: links between rail and trucking

•• Airports: links between air and rail/trucking

• Note that there are synergies between different modes of transportation

   
 

Damage to or destruction of one transportation mode may aggravate overuse problems on other modes, due to substitution

   

BIBLIOGRAPHY

Books and Published Sources

Ackley, David H. "An Empirical Study of Bit Vector Function Optimization," in Genetic Algorithms and Simulated Annealing, edited by Lawrence Davis (Los Altos: Morgan Kaufmann Publishers, Inc., 1987), 170-204.

Allison, Graham T. Essence of Decision: Explaining the Cuban Missile Crisis, (HarperCollins Publishers, 1971).

Ashby, Ross W. An Introduction to Cybernetics, (London: Chapman & Hall, 1965).

Berger, Bill D. and Kenneth E. Anderson. Modern Petroleum: A Basic Primer of the Industry, 2nd Edition, (Tulsa: PennWell Books, 1981).

Booker, Lashon. "Improving Search in Genetic Algorithms," in Genetic Algorithms and Simulated Annealing, edited by Lawrence Davis (Los Altos: Morgan Kaufmann Publishers, Inc., 1987), 61-73.

Brodie, Bernard. Strategy in the Missile Age, (Princeton: Princeton University Press, 1959).

Casti, John L. Complexification: Explaining a Paradoxical World Through the Science of Surprise, (New York: HarperCollins Publishers, Inc., 1994).

Clausewitz, Carl von. On War, edited and translated by Michael Howard and Peter Paret, (Princeton: Princeton University Press, 1976).

Davis, Lawrence and Martha Steenstrup. "Genetic Algorithms and Simulated Annealing: An Overview," in Genetic Algorithms and Simulated Annealing, edited by Lawrence Davis (Los Altos: Morgan Kaufmann Publishers, Inc., 1987), 1-11.

Douhet, Giulio. The Command of the Air, translated by Dino Ferrari, (Washington, D.C.: Office of Air Force History, 1983).

Finney, Robert T. History of the Air Corps Tactical School 1920-1940, (Washington, D.C.: Center for Air Force History, 1992).

Goldberg, David E. Genetic Algorithms in Search, Optimization, and Machine Learning, (Reading: Addison-Wesley Publishing Company, Inc., 1989).

Hallion, Richard P. Storm Over Iraq: Air Power and the Gulf War, (Washington, D.C.: Smithsonian Institute Press, 1992).

Hansell, Haywood S. Jr. The Air Plan That Defeated Hitler, (Atlanta: Higgins-McArthur/ Longino & Porter, Inc., 1972).

Iklé, Fred C. The Social Impact of Bomb Destruction, (Norman: University of Oklahoma Press, 1958).

Kauffman, Stuart A. The Origins of Order: Self-Organization and Selection in Evolution, (New York: Oxford University Press, 1993).

Kennett, Lee. The First Air War: 1914-1918, (New York: The Free Press, 1991).

Lewin, Roger. Complexity: Life at the Edge of Chaos, (New York: Macmillan Publishing Company, 1992).

May, Ernest. Lessons of the Past, (Oxford: Oxford University Press, 1976).

McFarland, Stephen L. and Wesley Phillips Newton. To Command the Sky: The Battle for Air Superiority over Germany, 1942-1944, (Washington, D.C.: Smithsonian Institute Press, 1991).

Mierzejewski, Alfred C. The Collapse of the German War Economy, 1944-1945, (Chapel Hill: The University of North Carolina Press, 1988).

Nicolis, Grégoire and Ilya Prigogine. Exploring Complexity: An Introduction, (New York: W. H. Freeman and Company, 1989).

Olson, Mancur Jr. The Economics of the Wartime Shortage: A History of British Food Supplies in the Napoleonic War and in World Wars I and II, (Durham: Duke University Press, 1963).

Olson, Reuben M. Essentials of Engineering Fluid Mechanics, Third Edition, (New York: Intext Educational Publishers, 1973).

Ott, Edward. Chaos in Dynamical Systems, (Cambridge: Cambridge University Press, 1993).

Perrow, Charles. Normal Accidents: Living with High-Risk Technologies, (New York: Basic Books, Inc., 1984).

Press, William H. et al. Numerical Recipes: The Art of Scientific Computing (Fortran Version), (Cambridge: Cambridge University Press, 1989).

Rawlins, Gregory J. E. "Introduction," in Foundations of Genetic Algorithms, edited by Gregory J. E. Rawlings, (San Mateo: Morgan Kaufmann Publishers, 1991), 1-10.

Reif, F. Fundamentals of Statistical and Thermal Physics, (New York: McGraw-Hill Book Company, 1965).

Stevenson, William D. Jr. Elements of Power System Analysis, 3rd Edition, (New York: McGraw-Hill, Inc., 1975).

The United States Strategic Bombing Survey, Summary Report (European War), 30 September 1945, (reprinted edition, Maxwell AFB, AL: Air University Press, October 1987).

Waldrop, M. Mitchell. Complexity: The Emerging Science at the Edge of Order and Chaos, (New York: Simon & Schuster, 1992).

Warden, John A. III, Col. "Employing Air Power in the Twenty-first Century," in The Future of Air Power in the Aftermath of the Gulf War, edited by Richard H. Shultz, Jr. and Robert L. Pfaltzgraff, Jr., (Maxwell AFB, AL: Air University Press, July 1992).

Articles

Abramson, D. "Constructing School Timetables Using Simulated Annealing: Sequential and Parallel Algorithms," Management Science, Vol. 37, No. 1 (January 1991), 98-113.

Anderson, P. W. "More is Different," Science, Vol. 177, No. 4047 (4 August 1972), 393-396.

Antonoff, Michael. "Genetic Algorithms: Software by Natural Selection," Popular Science, Vol. 239, No. 4 (October 1991), 70-74.

Arthur, W. Brian. "Why Do Things Become More Complex?" Scientific American, Vol. 268, No. 5 (May 1993), 144.

. "Pandora’s Marketplace," New Scientist (Supplement), Vol. 137, No. 1859 (6 February 1993), S6-S8.

. "Positive Feedbacks in the Economy," Scientific American, Vol. 262, No. 2 (February 1990), 92-99.

Austin, Scott. "Metamorph: A Genetic Algorithmic Tool," AI Expert, Vol. 5, No. 8 (August 1990), 48-55.

. "Genetic Solutions to XOR Problems," AI Expert, Vol. 5, No. 12 (December 1990), 52-57.

Bak, Per and Kan Chen. "Self-Organized Criticality," Scientific American, Vol. 264, No. 1 (January 1991), 46-53.

Basu, Atanu and L. Neil Frazer. "Rapid Determination of the Critical Temperature in Simulated Annealing Inversion," Science, Vol. 249, No. 4975 (21 September 1990), 1409-1412.

Beard, Nick. "Evolution and Computers," New Scientist, Vol. 125, No. 1699 (13 January 1990), 68.

Beyerchen, Alan. "Clausewitz, Nonlinearity, and the Unpredictability of War," International Security, Vol. 17, No. 3 (Winter 1992/93), 59-90.

Broyles, Gil. "The First Cut is the Deepest," In Perspective: The WilTel Magazine, (Summer 1992), 30-35.

Bylinsky, Gene. "Managing with Electronic Maps," Fortune, Vol. 119, No. 9 (24 April 1989), 237-254.

Conley, Stephen. "Quake darkens the West," USA Today (18 January 1994), 1A.

Crutchfield, James P. et al. "Chaos," Scientific American, Vol. 255, No. 6 (December 1986), 46-57.

Denning, Peter J. "Genetic Algorithms," American Scientist, Vol. 80, No. 1 (January-February 1992), 12-14.

Ellis, Steve and Paul Ginther. "Case history: GIS developed for Kern River gas system—Part 1," Pipe Line Industry (April 1993), 35-41.

. "Case history: GIS developed for Kern River gas system—Part 2," Pipe Line Industry (May 1993), 45-50.

Forrest, Stephanie. "Genetic Algorithms: Principles of Natural Selection Applied to Computation," Science, Vol. 261, No. 5123 (13 Aug 93), 872-878.

Gabris, William. "SCADA, MIS systems served by new hardware and software," Pipe Line Industry (April 1993), 21-25.

Grossmann, Siegfried and Gottfried Mayer-Kress. "Chaos in the international arms race," Nature, Vol. 337, No. 6209 (23 February 1989), 701-704.

Holland, John H. "Genetic Algorithms," Scientific American, Vol. 267, No. 1 (July 1992), 66-72.

Karr, Chuck. "Applying Genetics to Fuzzy Logic," AI Expert, Vol. 6, No. 3 (March 1991), 38-43.

. "Genetic Algorithms for Fuzzy Controllers," AI Expert, Vol. 6, No. 2 (February 1991), 26-33.

Kauffman, Stuart A. "Antichaos and Adaptation," Scientific American, Vol. 265, No. 2 (August 1991), 78-84.

Kennedy, Scott A. "Five Ways to a Smarter Genetic Algorithm," AI Expert, Vol. 8, No. 12 (December 1993), 35-38.

Kirkpatrick, S. et al. "Optimization by Simulated Annealing," Science, Vol. 220, No. 4598 (13 May 1983), 671-680.

Kirkpatrick, Scott. "Optimization by Simulated Annealing: Quantitative Studies," Journal of Statistical Physics, Vol. 34, Nos. 5/6 (1984), 975-986.

Korte, George B. "GIS Industry and Software Overview," P.O.B., Vol. 18, No. 3 (February-March 1993), 24-32.

Kwok, Andrew. "Fundamentals of SCADA and Automated Meter Reading," Pipeline & Gas Journal (February 1992), 22-28.

Lane, Alex. "Programming with Genes," AI Expert, Vol. 8, No. 12 (December 1993), 16-19.

Lewin, Roger. "The Right Connections," New Scientist (Supplement), Vol. 137, No. 1859 (6 February 1993), S4-S5.

Lewis, Rick, Lt Col. "DESERT STORM-JFACC Problems Associated With Battlefield Preparation," to be published in Air Power Journal.

Lin, Shen. "Computer Solutions of the Traveling Salesman Problem," The Bell System Technical Journal, Vol. 44, No. 10 (December 1965), 2245-2269.

Marshall, Steve. "Details on damage, delays and the dead," USA Today (18 January 1994), 3A.

Mann, Steven R. "Chaos Theory and Strategic Thought," Parameters, Vol. XII, No. 2 (Autumn 1992), 54-68.

McCready, Arthur K. "Client/server technology in SCADA systems," Pipe Line Industry (June 1992), 30-38.

Metropolis, Nicholas et al. "Equation of State Calculations by Fast Computing Machines," The Journal of Chemical Physics, Vol. 21, No. 6 (June 1953), 1087-1092.

Metz, Steven and Lt Col Frederick M. Downey. "Centers of Gravity and Strategic Planning," Military Review (April 1988), 23-33.

O’Brien, Larry. "Walking at the Edge of Chaos," AI Expert, Vol. 8, No. 12 (December 1993), 13-15.

Olson, Mancur Jr., 1st Lt. "The Economics of Target Selection for the Combined Bomber Offensive," RUSI Journal, Vol. CVII (November 1962), 308-314.

Rock, Denny and Joel Hirsh. "Will GAs Breed with Aerospace?" AI Expert, Vol. 8, No. 12 (December 1993), 29-34.

Ross, Bruce A. "The Case for Targeting Leadership in War," Naval War College Review, Vol. 46, No. 1 (Winter 1994), 73-93.

Ruelle, David. "Les attracteurs étranges," La Recherche, No. 108 (February 1980), 132-144.

Singleton, Andy. "Genetic Programming with C++," Byte, Vol. 19, No. 2 (February 1994), 171-176.

Sterba, Vernon J. "Choosing the right technology for integrated SCADA communications," Pipe Line Industry (May 1992), 25-28.

Stewart, Ian. "A New Order," New Scientist (Supplement), Vol. 137, No. 1859 (6 February 1993), S2-S3.

Vecchi, Mario P. and Scott Kirkpatrick. "Global Wiring by Simulated Annealing," IEEE Transactions on Computer-Aided Design, Vol. CAD-2, No. 4 (October 1983), 215-222.

Wayner, Peter. "Genetic Algorithms," Byte, Vol. 16, No. 1 (January 1991), 361-368.

"What is a GIS? (And What It Isn’t)," Design Management (July 1991), 26-28.

Wike, Andy. "Operating companies looking at new SCADA applications–Part 1," Pipe Line Industry (May 1992), 25-38.

. "Operating companies looking at new SCADA applications–Part 2," Pipe Line Industry (June 1992), 45-49.

Wilson, Steve. "How to Grow a Starship Pilot," AI Expert, Vol. 8, No. 12 (December 1993), 21-26.

Interviews

Almquist, Keith. Strategic Studies Department I, Sandia National Laboratories (hereafter abbreviated Sandia). Albuquerque, NM, 6 April 1994.

Arkin, William M. Greenpeace. SAAS, 29 April 1994.

Averill, Edwin (Bud) L. Senior Electrical Engineer, Production Engineering, Public Service Company of Oklahoma. Tulsa, OK, 26 January 1994.

Beisemeyer, Mark D. Engineer, Willbros Butler Engineers, Inc. Tulsa, OK, 28 January 1994.

Broyles, Gil. Assistant to the President, WilTel. Tulsa, OK, 27 January 1994.

Caswell, Jon. Sandia. Albuquerque, NM, 6 April 1994.

Cover, Bob. Starbase Laboratory, Sandia. Albuquerque, NM, 6 April 1994.

Crocker, Max. Project Manager, Willbros Butler Engineers, Inc. Tulsa, OK, 15-16 July 1993.

Engi, Dennis. Supervisor, Strategic Technologies Division, Sandia. Albuquerque, NM, 6 April 1994.

Godfrey, Brad. Sandia. Albuquerque, NM, 6 April 1994.

Hartman, William F. Special Projects Div 9614, Safeguards Program, Sandia. Albuquerque, NM, 6 April 1994.

Hilbig, Kirk. Shift Supervisor, Williams Pipe Line Company. Tulsa, OK, 27 January 1994.

Lawson, Doug. Strategic Offense Studies Department, Sandia. Albuquerque, NM, 6 April 1994.

Light, W. John. Director, Research & Development, Excel Energy Technologies, Ltd. Tulsa, OK, 27 January 1994.

Ling, Bill. Manager, Strategic Offense Studies Department, Sandia. Albuquerque, NM, 6 April 1994.

Malczynski, Len. Sandia. Albuquerque, NM, 6 April 1994.

Martin, Andy. Engineer, Willbros Butler Engineers, Inc. Tulsa, OK, 28 January 1994.

Milloy, John. Sandia. Albuquerque, NM, 6 April 1994.

Petzet, G. Alan. Exploration Editor, Oil & Gas Journal. Tulsa, OK, 26 January 1994.

Rinaldi, Roger E. Engineer, Willbros Butler Engineers, Inc. Tulsa, OK, 15-16 July 1993 and 26-29 January 1994; numerous telephone interviews from July 1993-May 1994.

Roehrig, Stephen C. Chief, Advanced Transportation Programs, Sandia. Albuquerque, NM, 9 December 1993, 6 April 1994.

Staggs, Russell. Engineer, Willbros Butler Engineers, Inc. Telephone interview, 21 April 1994.

Warden, John A. III, Col. Commandant, Air Command and Staff College. ACSC, 28 March 1994.

Wineinger, James S. Engineer, Network Development, WilTel Business Networks. Tulsa, OK, 27 January 1994.

Unpublished Sources

Air Force Manual 1-1, Basic Aerospace Doctrine of the United States Air Force (March 1992).

A-WPD/1: Munitions Requirements of the Army Air Forces to Defeat Our Potential Enemies (12 August 1941), HRA file 145.82-1.

AWPD-42: Requirements for Air Ascendancy, 1942, HRA file 145.82-42.

Bjorkman, Eileen, Maj., et al. "Chaos Primer," in Air Campaign Course 1993: Research Projects, edited by Dr. Richard Muller, Lt Col Larry Weaver, and Lt Col Albert Mitchum, (Maxwell AFB, AL: Air Command and Staff College, 1993).

Boyd, John R. "A Discourse on Winning and Losing," unpublished briefing and essays, Air University Library, Document No. MU 43947 (August 1987).

Catalog, Oil & Gas Journal Energy Database / Electric Light & Power Electric Power Database, Tulsa, OK, Fall/Winter 1993.

DeBlois, Bruce M., Maj, et al. "Dropping the Electric Grid: An Option for the Military Planner," unpublished research report (Maxwell AFB, AL: Air Command and Staff College, 1994).

DiJulio, Paul, Maj., et al. "Communications-Computer Systems: Critical Centers of Gravity," in Air Campaign Course 1993: Research Projects, edited by Dr. Richard Muller, Lt Col Larry Weaver, and Lt Col Albert Mitchum, (Maxwell AFB, AL: Air Command and Staff College, 1993).

Energy Catalog, PennWell Books, Tulsa, OK, Winter 1994.

Explorer Pipeline Company. Company Brochure (September 1992).

Felker, Edward J., Maj. Does the Air Force Practice Its Doctrine? A Limited and Focused Air Campaign Concept, unpublished thesis, (Fort Leavenworth, KA: Command and General Staff College, 1991).

Graper, Mark W., Maj., et al. "Petroleum as a Center of Gravity," in Air Campaign Course 1993: Research Projects, edited by Dr. Richard Muller, Lt Col Larry Weaver, and Lt Col Albert Mitchum, (Maxwell AFB, AL: Air Command and Staff College, 1993).

Griffith, Thomas E. Jr., Maj. Strategic Attack of National Electrical Systems, unpublished thesis (Maxwell AFB, AL: School of Advanced Airpower Studies, 1993).

Hust, Gerald R., Maj. Taking Down Telecommunications, unpublished thesis, (Maxwell AFB, AL: School of Advanced Airpower Studies, 1993).

Kaysen, Carl. "Note on Some Historic Principles of Target Selection," RAND, RM-189 (15 July 1949).

Layton, Cliff. "When to Use AI Techniques," unpublished briefing slides, presentation to the Oklahoma Artificial Intelligence Special Interest Group.

MAPSearch Services. Company brochure, MAPSearch Services, Durango, CO, 1994.

PCLF: Personal Computer Load Flow User Manual (Norman: University of Oklahoma, School of Electrical Engineering and Computer Science, 1986).

Pentland, Pat A., Lt Col. Center of Gravity Analysis and Chaos Theory or How Societies Form, Function, and Fail, unpublished thesis (Maxwell AFB, AL: Air War College, 1993).

Perera, Guido R. History of the Organization and Operations of the Committee of Operations Analysts, 16 November 1942 - October 1944, HRA file 118.01.

Plantation Pipe Line Company. Public Service Brochure (1993).

Shlapak, David A. "Electrical Power as a Target System," Project AIR FORCE, RAND Memorandum PM-187-AF (December 1993).

U.S. Army TM 5-343, Military Petroleum Pipeline Systems, Revision C-1, (Headquarters, Dept. of the Army, 1973).

Warden, John A. III, Col. "Strategic Warfare: The Enemy as a System," unpublished manuscript (3 January 1993).

WilTel. Various media information brochures, 1992-1993.

Wuesthoff, Scott E., Maj. The Utility of Targeting the Petroleum-Based Sector of a Nation's Economic Infrastructure, unpublished thesis, (Maxwell AFB, AL: School of Advanced Airpower Studies, 1992).