INTERDEPENDENCE, INFORMATION PROCESSING AND ORGANIZATION DESIGN: AN EPISTEMIC PERSPECTIVE Phanish Puranam ppuranam@london.edu London Business School University of London Marlo Goetting mgoetting.phd2007@london.edu London Business School University of London Thorbjorn Knudsen tok@sam.sdu.dk Strategic Organization Design Unit University of S. Denmark This Draft: Jan 13, 2010 Abstract1 We develop a novel analytical framework to study epistemic dependence: “who needs to know what about whom” as a basis for understanding information processing requirements in organizations, and the resulting implications for organization design. The framework we develop helps to describe and compare the nature of the underlying coordination problems generated by different patterns of interdependence, and the resulting knowledge requirements for the design and implementation of appropriate organizational structures. The framework offers a formal language that may prove useful for parsimoniously integrating what we know, as well as for building new theory. 1 The authors thank Yang Fan, Mike Ryall, Mihaela Stan and Bart Vanneste for helpful comments on earlier drafts. Puranam acknowledges funding from the European Research Council under Grant # 241132 for “The Foundations of Organization Design” project. 1. INTRODUCTION An extensive literature treats the design and structure of organizations as a response to the information processing needs generated by interdependence between its constituent agents (e.g. Simon, 1945; March and Simon, 1958; Lawrence and Lorsch, 1967; Thompson, 1967; Van de Ven, Delbeq, Koenig, 1976; Galbraith, 1977; Tushman and Nadler, 1978; Burton and Obel, 1984). The goal of this paper is to offer a novel analytical framework that can potentially refine and elaborate on the foundational concept of “interdependence” in this literature. Specifically, we develop an approach to systematically analyze the epistemic implications of different patterns of interdependence. We explicitly describe the nature of the underlying coordination problem (Schelling, 1960) generated by different patterns of interdependence, and the resulting knowledge requirements for the design and implementation of appropriate organizational structures.2 In doing so, we develop a formal language that may prove useful for parsimoniously integrating what we know, as well as for building new theory. Interdependence is a concept that applies at multiple levels of analysis- between individuals, groups and organizations- but always refers to a relationship between two decision-making entities. Given two tasks undertaken by actors A and B, there is dependence of A on B if A’s outcomes depend partly on the actions taken by B. Interdependence exists when A’s outcomes depend on B’s actions and vice versa. This conceptualization is common across widely varied literatures. It can be found implicitly in Thompson’s pioneering work on 2 Coordination failures occur when interacting individuals are unable to anticipate each other’s actions and adjust their own accordingly (Schelling, 1960); in organizations, coordination failures are often manifested as delay, mis-understanding, poor synchronization and ineffective communication. In contrast cooperation failures occur when interdependent individuals are not motivated to achieve the optimal collective outcome because of conflicting incentives. Coordination failures can occur quite independent of cooperation failures – even when incentives are fully aligned (Simon, 1947; March and Simon, 1958; Schelling, 1960; Camerer, 2003; Heath and Staudenmayer, 2000; Grant, 1996; Holmstrom and Roberts, 1998). Cooperation and coordination failures are therefore individually sufficient reasons for the failure of collaboration. 2 interdependence between those executing related tasks (Thompson 1967), in the literature on groups (Kelley and Thibault, 1978) and relationships between organizational units (Van de Ven et al 1976). The same analytical idea also underlies formal models of complementarities in production functions (Milgrom and Roberts 1990), and epistatic interactions in models of organizational adaptation on fitness landscapes (Levinthal 1997), though in many applications in these domains the focus is on interdependent tasks undertaken by a single decision maker. Interdependence is a natural foundational construct for analyzing organization design, which involves two complementary problems- the division of labor and the integration of effort (March and Simon, 1958; Lawrence and Lorsch, 1967; Burton and Obel, 1984). Interdependence is the inevitable consequence of the division of labor as it arises between those who carry out the divided labor. Consequently, much of the literature on formal organization design rests on the premise that specific patterns of the division of labor give rise to specific patterns of interdependence, and that efficient organizational forms “solve” the problems of cooperation and coordination that arise when integrating the efforts of interdependent actors (Simon, 1945; Thompson, 1967; Lawrence and Lorsch, 1967). For this idea to have theoretical and empirical traction, the analysis of organization designs must build on a clear understanding of why certain patterns of interdependence are harder to organize around than others. An influential set of answers to this question are rooted in the notion of bounded rationality (Simon, 1945), with a focus on the costly information processing requirements generated by interdependence. These are the costs of “communication and decision making” (Thompson, 1967: 57) or the “gathering, interpreting and synthesis of information in the 3 context of organizational decision making” (Tushman and Nadler, 1978: 614)3. In this perspective, distinct patterns of interdependence give rise to different information processing requirements (Thompson, 1967), and “variations in organizing modes are actually variations in the capacity of organizations to process information” (Galbraith, 1977: 79). Our approach lies squarely within this perspective. Our key thesis is that different patterns of division of labor generate corresponding patterns of interdependence, which in turn imply a particular distribution of knowledge across actors that is necessary for them to achieve coordinated action. Thus the nature of interdependence shapes epistemic dependence: “who needs to know what about whom”. We claim that taking into account the epistemic consequences of interdependence allows us to establish an improved mapping from types of interdependence to types of organizational structures, which is the fundamental mapping on which most descriptive as well as prescriptive principles of organization design in the information processing perspective are based. We acknowledge an important alternative tradition that focuses on the challenge of aligning incentives to explain why certain patterns of interdependence are harder to organize around. In this perspective, managing interdependence can be complicated when the initial (pre-design) distribution of gains makes it unlikely that optimal actions will be taken - for instance, because of gains from free riding, hold-up, or other forms of rent seeking behavior. This approach to the analysis of interdependence characterizes the extensive literature on social dilemmas (Kollock, 1998), reward interdependence within teams (Kelly and Thibault, 1978), principal-agent models with multiple agents (Holmstrom, 1982), and hold-up problems in transaction cost economics as well as the property rights perspective (Williamson, 1985; Grossman and Hart, 1986). These diverse accounts from social 3 The term “coordination costs” is also sometimes used to describe these costs. However, this is ambiguous as it could refer either to the costs of the information processing necessary to generate coordinated action, or the (opportunity) costs that arise when coordination failures occur. 4 psychology and economics have in common the notion that the central organizational design problem involves using incentives (formal or informal) to change the distribution of gains from managing interdependence in a manner that motivates optimal (if not efficient) actions. While incentive structure does play an important part in this paper, we are largely agnostic to the problem of the distribution of gains from managing interdependence, and focus instead on the epistemic and information processing consequences of interdependence. A fuller integration between the incentive and information perspectives is doubtless desirable, but beyond the scope of this paper. The rest of this paper is organized as follows: The next section discusses Thompson’s highly influential conceptualization of interdependence and implications for information processing and organization design (Thompson, 1967), and notes ambiguities within this framework. We then develop a formal definition of epistemic dependence in section 3, and discuss its application to deriving implications of any pattern of the division of labor, and the consequent role of the organization designer in sections 4 and 5. In the discussion section we show how our approach accommodates a number of traditional organization design topics, and summarize our key contributions in the conclusion. 2. PRIOR CONCEPTUALIZATIONS OF THE LINKS BETWEEN INTERDEPENDENCE & INFORMATION PROCESSING An early conceptualization of the mapping from interdependence to information processing requirements was presented by Thompson (1967). Thompson’s framework is still the dominant one for studying interdependence as a basis for understanding organizational structures. Figure 1 provides some data on the number of papers published by year in five important management journals (Strategic Management Journal, Administrative Science Quarterly, Organization Science, Academy of Management Journal and Academy of 5 Management Review) that a) cite Thompson’s 1967 book b) contain the word “interdependence” and c) the words “organization (al) structure (OR) design” as a fraction of the number of papers that satisfied b) and c) alone. Given its wide usage and utility, we use Thompson’s approach to the analysis of interdependence to briefly illustrate what we see as distinctive about our approach. ---------------------------------Insert Figure 1 about here ---------------------------------- The central idea that arises from Thompson’s work on interdependence is that complex forms of interdependence require high levels of costly information processing activities such as decision making and communication (Galbraith, 1977; Tushman and Nadler, 1978). In Thompson’s classic typology of interdependence complexity (1967), interdependence is “pooled” when tasks depend on each other in a simple additive manner; each task renders a discrete contribution to the whole, and each is supported by the whole. Interdependence is “sequential” when the outputs of one task form the inputs of the other. Finally, interdependence is “reciprocal” when the outputs of each task become the inputs of the other. Influential contributions by Galbraith, Tushman and Nadler elaborated on the mapping from increasing interdependence complexity to increasing need for information processing activities (Galbraith, 1973; Tushman and Nadler, 1978). We note that both Galbraith’s and Tushman and Nadler’s analyses were clearly epistemic in spirit, because their emphasis lay in outlining the conditions under which more or less knowledge was to be made available to interdependent agents through information processing activities. In this respect, our work directly builds on and extends their contributions. However, we also argue that Thompson’s interdependence framework is a problematic basis on which to construct a mapping from the nature of interdependence to the extent of information processing 6 necessary to manage it. Specifically, we see two important ways in which the framework is underspecified, which result in ambiguities in its application. First, Thompson’s framework does not distinguish between the logical sequence of actions and their temporal sequence. Pooled interdependence involves a situation where each action makes an independent contribution to the whole- no task provides an input to another. However, this situation can accommodate variation in the timings of the actions themselves, which may be sequential or simultaneous. Actions may be taken sequentially, even though the results of the actions themselves are pooled to create a final output. As we will show, the timing of actions has critical implications for the epistemic conditions that must hold for actions to be coordinated. Furthermore, if the time it takes for tasks to be executed is greater than zero, then a distinct category for “reciprocal interdependence” is unnecessary, as it can be expressed as a repeated cycle of temporally sequential tasks. Second, Thompson’s framework is silent about the role of incentives, and therefore does not account for the entire class of interdependencies induced by incentive structures (Kelley and Thibault, 1978; Wageman, 1995). It is worthwhile elaborating on this critique in some detail to delimit its scope precisely: Thompson’s work, like much of the literature in classical organization theory (and in contrast to much work in economics) focused on issues of information and coordination rather than issues of incentive conflict. Our point is not to point out that his approach privileges information over incentives; tractability is necessary for theorizing, and indeed the converse critique could apply to the extensive literature that has emphasized the incentive consequences of interdependence over coordination issues (also see Dosi, Levinthal, and Marengo, 2003; Kretschmer and Puranam, 2008). Rather, our point is that incentive structures themselves can critically impact the epistemic implications of various patterns of interdependence. For instance, what is required to be known for coordinated action by two individuals who are rewarded on joint output is not the same as 7 when they are rewarded on individual output. The complexities of managing the different patterns of interdependence Thompson analyses could vary vastly as a function of the incentive structure in place, and it is this missing “moderator” role of incentives that we point to in this critique. As will become clear from the next two sections, without knowing the nature of the incentive structure – individual or joint rewards- and the sequence of actions – sequential or simultaneous- it is difficult to give a precise specification of the underlying coordination problem (also see Weber (2004) for an attempt to model each of Thompson’s interdependence types as a distinct coordination game). Without knowing the nature of the coordination problem associated with a particular pattern of interdependence, or a means to compare these underlying coordination problems, it is in turn difficult to rank them in terms of the need for information processing. Thus, the ambiguities in Thompson’s framework raise doubts about the ordering of his interdependence categories in terms of the need for information processing. The approach we introduce in the next two sections avoids these ambiguities. We are able to specify the underlying coordination problems generated by any pattern of division of labor, and also compare them in terms of the knowledge requirements imposed on the relevant actors, using the concept of epistemic dependence. 3. DEFINING EPISTEMIC DEPENDENCE To define epistemic dependence, consider a simple organization comprising multiple agents, in which each agent i undertakes costly action a i and receives performance based compensation. The agent’s utility is given by u i = π i (a i ,ψ i ) − ci (ai , ξ i ) , where ci is the agent’s cost of action, π i is compensation, and ψ i and ξ i are vectors of exogenous parameters. Total utility for all agents in the system is thus given by U = ∑ u i . The 8 functional form of the agent’s objective function is defined as oi = [π i (a i ,ψ i ) − ci (ai , ξ i ); Ω i ] where Ω i represents any endogenous constraints faced by the agent as part of his objective function. These could include internalized norms or habits of behavior, for instance. Assumption 1: For total utility of the organization to be maximized, for every agent i it is necessary that K i {oi } The “knows that” operator K i {.} from modal (epistemic) logic can be read as “agent i knows {.}”. This assumption makes explicit the notion that each agent is a decision maker (rather than an automaton), who acts on the basis of her own objective function, which may include constraints in the form of habits or norms for behavior in a particular context. Assumption 2: For agent i to accurately predict agent j’s actions, it is necessary that K i {o j } The second assumption is not innocuous. It rules out, for instance, the use of a history of prior observations that may enable an accurate prediction of another’s actions (or at least a history that would be sufficient to generate an accurate prediction). It also rules out the existence of signals that allow infallible predictions about actions. This leaves knowledge of the agents’ objective function as the only means by which an accurate prediction of that agent’s actions can be made. Definition: Given assumptions 1 & 2, if agent i maximizes his own utility only by acting on the basis of an accurate prediction of agent j’s action, then i is epistemically dependent on j. We say there is epistemic dependence between agent i and j under these conditions because it follows automatically that under assumptions 1 & 2 for total utility to be maximized, it is necessary that K i , j {o j } - i.e. agent j’s objective function must be shared knowledge between agent i and j. To see this, consider that for total utility to be maximized, agent j must know 9 her own objective function K j {o j } (Assumption 1). If i maximizes his own utility by acting on the basis of an accurate prediction of j’s action then it is necessary that K i {o j } (Assumption 2). Therefore for total utility to be maximized, it is necessary that K i {o j } ∧ K j {o j } ≡ K i , j {o j } . The notion of epistemic dependence helps to sharpen our discussion of coordination problems in two important ways. First, we may say that whenever one agent’s utility is maximized by acting on an accurate prediction of another agent’s actions, there is a (potential) coordination problem.4 More generally, coordination problems in “real time” settings (e.g. surgical teams, fire-fighters) as well as in “realistic time” settings (new product development, strategic alliances) are fundamentally identical. Any setting in which actions are unobservable- either because they are taking place simultaneously, because of communication/information transmission constraints, or because of timing (it hasn’t happened yet) – but must be predicted, can be modeled as a coordination problem. Communication itself can be seen as a coordination problem, as indeed the modern view of linguistics sees it: when communicating, I need to predict which among several possible meanings you chose to attach to the word you use (eg. Clark, 1996). A coordination failure thus is fundamentally a failure to predict the actions of another in situations where such a prediction is essential for optimal action by oneself. It also follows that whenever there is epistemic dependence, there must be a coordination problem, but the converse is not true. For instance, the need to predict other’s actions in order to take one’s own optimal actions may exist even in situations where Assumption 2 is not valid- where norms, precedents conventions and signals exist to help make this prediction. Since there is no epistemic dependence unless assumption 2 holds, it is clear that an interesting set of 4 Thus any game where at least one agent does not have a dominant strategy features a coordination problem. 10 questions arise about the circumstances under which prior history can serve as a basis for predicting actions, or a system of signals can arise that makes such predictions possible. However, we defer these issues for future work and focus on situations in which assumption 2 does hold. Second, we note that there is bound to be a coordination failure if epistemic dependence exists and the shared knowledge condition (i.e. the objective function of the agent whose action is being predicted is known to itself as well as the predictor) is not met. On the other hand, because the shared knowledge condition is necessary not sufficient, one could also have a coordination failure when (given epistemic dependence) the shared knowledge condition is met, but higher order shared knowledge is not present- common knowledge may not exist (see Appendix A for a formal definition of shared knowledge of any order). For instance, consider the case of epistemic interdependence, which is a small conceptual step from epistemic dependence: If agent i is epistemically dependent on agent j, and agent j is epistemically dependent on agent i, then there is epistemic interdependence between agents i and j. Therefore each must know whatever the other must know in order to act- so that the objective function of each must be shared knowledge between them. However, shared knowledge (of first order- both know it) may not be sufficient: unless both know that both know, and both know that both know that both know (and so on), there may be residual doubt in the mind of each agent about the action of the other. This residual doubt is in principle fully eliminated only with shared knowledge of infinite order- which is common knowledge. If we assume that residual doubt must disappear for coordinated action to take place, then common knowledge is necessary; if some residual doubt is tolerable (because of the relative costs of foregone coordination opportunities vis-à-vis failed attempts at 11 coordination), then only shared knowledge of finite order may be necessary. However, in either case, shared knowledge of first order is always necessary. 4. THE DIVISION OF LABOR AND EPISTEMIC DEPENDENCE Having defined epistemic dependence, we next provide a framework within which to map different ways to divide labor to the resulting epistemic implications. To represent the division of labor, we utilize the notion of a Task Architecture (denoted by A): a configuration of tasks performed by agents who are measured and rewarded in certain ways. Task Architectures capture the nature of the division of labor- which we define as the result of a process of decomposing a single task performed by a single agent, into multiple tasks performed by multiple agents. A task may be thought of as a production technology- it is a transformation of inputs into outputs in a finite (non-zero) time period. The inputs are broadly of two kinds: 1) actions taken by agents (always necessary- if there is no action as an input to a task, only the output of a previous task, then the task is indistinguishable from the previous task), and 2) the outputs of other tasks.5 In the basic formulation of our framework, each agent is assigned a single task, though it is easy to relax this assumption. The fundamental building block of a Task Architecture is the task dyad – a pair of tasks i and j conducted by a pair of agents. More complicated structures are built through scaling (e.g. from dyad to triad) delegation (hierarchical decomposition of tasks) and recursion (e.g. a dyad in which each task is itself a dyad of tasks- see Appendix B). 4.1 Dimensions of Task Architecture 5 One may also consider Resources to be another kind of input. A Resource is an input that cannot itself be viewed as the output of any other task within the Task Architecture being analysed. 12 Task architecture can be described along two basic dimensions, which capture the nature of 1) task sequencing and 2) incentive breadth 1. Task sequencing (S): This refers to the timing of actions by the agents. For instance, in a dyad of tasks, the two agents may act sequentially (q) or simultaneously (u). Note that “simultaneous” need not actually mean synchronized actions by the agents; as long as each did not know the actions/outputs of the other agent when they acted, their actions are effectively simultaneous. Conversely, sequential actions may not be restricted to only those cases where the output of the first task is strictly necessary for the second task, but more generally refers to the case of one agent acting only after the action/output of the other agent has become visible. 2. Incentive Breadth ( π ): In a dyad, narrow incentives correspond to individual incentives, whereas broad incentives correspond to dyad level incentives (e.g. Kretschmer and Puranam, 2008). More generally, the incentives for unit i (where i may be an individual agent, group of agents, group of groups.. etc) are narrow if performance is at the same level of aggregation as i, and broad if performance is measured at the next higher level of aggregation. To illustrate our notation for capturing the nature of a simple Task Architecture comprising a single task dyad, we could write A :{ S (q ) | π 1 , π 2 } to denote sequential actions, with narrow incentives for agent 1 and agent 2. We take the perspective of a Designer who can make choices about one or both dimensions of Task Architecture- sequencing and incentive breadth. As before, we assume that the agent maximizes his utility given by u i = π i (a i ,ψ i ) − ci (ai , ξ i ) , subject to constraint Ω i . The designer maximizes utility from the final output of the Task Architecture 13 net of incentive compensation paid out U d = Π A − ∑ π i . Total utility is thus given by U = U d + ∑ U i . In an analysis focusing on incentive considerations, we would now proceed to find the optimal incentive structure a designer would set (as in standard principal agent models) subject to some participation constraint for the agents; however, our goal is to explore the epistemic requirements necessary to maximize total utility. In other words, we aim to specify the distribution of knowledge across agents that would be necessary to enable maximization of the sum of the agents’ and designer’s utilities, i.e. total utility. To do so, corresponding to every Task Architecture A we show how to construct an Epistemic Structure- E which captures “who needs to know what about whom” for total utility to be maximized. E thus describes the necessary epistemic conditions for total utility to be maximized for a given A. In all there are 4 (22) possible variants on task architecture for a single task dyad. The corresponding epistemic structures are shown in Figure 2. A “1” in a cell indicates that the column actor needs to know row knowledge element for total utility to be maximized. The task architectures in which there is epistemic dependence between agents (i.e. off-diagonals are non zero in the matrices in E) are those, by definition, in which one agent must act on the basis of a prediction of the acts of another. This situation arises under the conjunction of two conditions: a) whenever agent i faces broad incentives – where the returns to agent i’s actions depend at least partly on j’s actions so that j’s actions feature as part of the vectors ψ i , ξ i and b) when i is unaware of j’s action at the time of acting, which could be the case if i and j act simultaneously, or i acts sequentially before j. Consider the case where the two agents have broad incentives and act simultaneously. In effect, both agents act before knowing the actions/outputs of the other, so that each is epistemically dependent on the other. Therefore each must know whatever the other must 14 know in order to act- so that the objective function of each must be shared knowledge between them. In contrast, consider the case of sequential action with broad incentives for (at least) the first agent. There is an asymmetry that is immediately visible; the first mover needs to know whatever the second mover needs to know in order to act but the converse is not true.6 Here, there is only epistemic dependence but not epistemic interdependence. Put simply, epistemic dependence between agents in an interdependent dyad arises only when broad incentives are used; and given broad incentives, epistemic dependence is greater for simultaneous than for sequential actions. In other dyadic task architectures, there is no epistemic dependence between agents. In Appendix B, we show how epistemic dependence can be represented for task architectures beyond the simple dyad. ---------------------------------Insert Figure 2 about here ---------------------------------- 4.2 From Epistemic Structures (E) to Information Structures (It ) While the epistemic structures (E) gives the necessary knowledge conditions to maximize total utility, in any given task architecture, the agents may not possess such knowledge. The Information Structure (It) captures their actual state of knowledge at any point in time. The Information Structure has exactly the same form as the matrix representing E. However, the entries in the cells now capture the probability that the (column) agent knows the (row) knowledge element. 4.3 The Information Processing Implications of Epistemic and Information Structures 6 Readers familiar with game theory will recognize the well known principle of “backward induction” 15 In order to compare Task Architectures in terms of the relative costliness of information processing required to maximize total utility in each, we define a few useful metrics on any E and It: Epistemic Load (EL) of the task architecture, which is simply the number of all positive elements in any E7. To the extent that memory is expensive, or every piece of knowledge must be communicated to agents, EL is one way to capture the magnitude of information processing costs in any task architecture. Epistemic Dependence (ED) is measured as the sum of all off-diagonal elements in E. The diagonal elements in E8 tell us whether the agent needs to know the relevant knowledge item about himself, whereas off-diagonal elements tell us what agent i needs to know about other agents in the system. If knowledge about an agent is typically private, and expensive to communicate or be learnt by another agent, then ED is another measure of information processing costs which focuses on the cost of communication among interdependent agents in order to create sufficient epistemic conditions for coordinated activity. A visual representation of ED also shows us the structure of the network linking agents who need to communicate. It is worth noting here that in our framework we presume that the division of labor always involves some degree of specialization, i.e. that the tasks are qualitatively different in a task architecture. However, this need not be so, as one can also see division of labour with homogenous efforts (arising from individual effort constraints) – for instance a group of workers trying to move furniture. The major difference in the case of division of labor 7 8 n n i =1 j =1 Formally, EL = ∑ ∑ δ eij , where δ takes on a value of 1 if eij is positive and 0 otherwise. Formally, ED=EL-tr(E) 16 without specialization lies in the lower cost of generating shared knowledge to cope with epistemic dependence. The nature of the information processing problem in the first case (with specialization) compared to the latter (without specialization) is fundamentally different – without specialization both agents hold the same general knowledge, in the case with specialization, each agent has task specific knowledge. The cost of creating shared knowledge about each other’s objective functions and constraints, when it is required to be shared, should be higher with specialization than without. Thus we can expect that the information processing costs for a given level of epistemic dependence are higher with specialization than without. Reliability ( ρ ) We can combine information from E and It to calculate the reliability of any task architecture: first create a new square matrix from the element by element multiplication of E and It and then multiply all the non-zero elements of the resulting matrix together to yield a scalar in the range [0, 1]. Reliability captures the joint probability that all agents know everything they need to for total utility to be maximized. Reliability as defined here provides a precise way of assessing the extent of task uncertainty (Galbraith) - “the difference between the information required to perform a task and the information already possessed” (1973: 5). Epistemic Load and Epistemic Dependence offer two precise ways in which we can assess the de novo information processing costs of any task architecture. Epistemic Load includes both the agent’s knowledge of own objective function and constraints as well as Epistemic Dependence with other agents. If we assume that knowing one’s own objective function and constraints is costless, then Epistemic Dependence may be a better measure of information processing costs; else we may use the more comprehensive Epistemic Load measure. In any case, EL= ED + the number of agents in the task architecture. However, both these measures assume that agents have no prior knowledge. In contrast, Reliability, which 17 suggests a performance metric for any task architecture (given the epistemic and information structures) may also be used to assess the extent of information processing necessary given what the agents now. Put simply, the lower the reliability, the higher the information processing costs required to ensure that It =E 5. THE ROLE OF THE DESIGNER By analyzing task architectures in terms of epistemic dependence between agents, it is also possible to conceptualize the role of a designer in epistemic terms. We assume that the goal of the designer is to maximize reliability of the task architecture at minimal information processing cost. Broadly, the designer can do this in two ways: first, by shaping the task architecture in a manner that minimizes the epistemic burden on the agents; second, in creating communication channels between agents who are epistemically dependent so that the shared knowledge conditions are met through communication between agents. These correspond closely to the classical distinction drawn between coordination through programming and feedback in organizations (March and Simon, 1958). Our discussion below shows the compatibility between the classic conceptualizations of coordination modes and the notion of epistemic dependence, but also goes a step further: it points to the epistemic conditions for the designer to be able to adopt either mode of coordination. 5.1. Design as Programming: Shaping task architecture In this mode of coordination, the designer’s role consists of creating plans, schedules and programs to enable integration of effort without the need for communication among agents (March and Simon, 1958). According to March and Simon, “[t]he type of coordination 18 (...) used in the organization is a function of the extent to which the situation is standardized”, and hence, “[t]he more stable and predictable the situation, the greater the reliance on coordination by plan.” (1958: 182). In epistemic terms, we may think of this role of the designer as involving the selection of a task architecture A in a manner that minimizes the epistemic burden on the agents. We know that epistemic dependence is lowest in task architectures characterized by narrow incentives, followed by broad incentives with sequential actions, and highest in broad incentive structures with simultaneous actions (see Figure 2). Therefore, we need to understand under what conditions a designer can impose a narrow incentive structure or at least sequence actions. We argue that knowing what outputs to expect from each task and the logical sequencing of tasks constitute a form of architectural knowledge (Henderson and Clark, 1990; von Hippel, 1990; Baldwin and Clark, 2000). This comprises knowledge about task decomposition and task integration. Specifically, the designer must know the different tasks and the order in which they must be performed for the final system level output to be produced. Having architectural knowledge thus implies knowing what outputs to expect from each task in the entire task architecture. Architectural knowledge is therefore required if the designer wants to implement a narrow incentive structure or sequence actions. Conversely, in the absence of architectural knowledge, the task architecture will admit broad incentives with simultaneous sequencing of action only. This task architecture generates the highest level of epistemic dependence between agents. Limited architectural knowledge thereby results in coarser partitioning of sub-tasksThe key point here is that the architectural knowledge of the designer helps to reduce epistemic dependence between agents. Further, the designer can also influence the epistemic load of the agents in terms of what they need to know about themselves, not only about others (i.e. epistemic dependence). 19 If the designers possess knowledge of the production technology used by the agents- the knowledge required at the task level to convert actions (and inputs) to outputs, then the agents need not know their own production technologies, because the designer can set incentives based on actions rather than outputs. Thus, being paid on output implies a higher epistemic requirement for the agent- who must know his own production technology. The corollary is that if the agent’s incentive pay is based on his actions, then the agent need not know his own production technology. The epistemic requirement is lowered because the agent’s objective function becomes simpler. This is an important epistemic distinction between output vs. behavior based contracts, and also points to the conditions under which a designer can impose a high degree of formalization (i.e. action based contracts for agents) (Galbraith, 1973). Thus, whether programming takes the form of designing task architectures with narrow incentives or action based incentives- which is after all what plans, schedules and programs are- viewed from the epistemic perspective, its function is to exploit the knowledge of the designer to minimize the epistemic burden on the agents. An equally important insight is that designers with different levels of knowledge may well choose different task architectures for the same set of tasks. For instance, consider a task dyad assigned to the pair of agents {1,2} and {3,4}, each with its own designer, D1 and D2. D1 may sequence the tasks simultaneously, while D2 may do so sequentially. We know from Figure 2 that if both designers use narrow incentives, then in fact the two pairs of agents have identical knowledge requirements. However, even if D1 used broad incentives, D2 could still plausibly use narrow incentives as long as he knows what to measure- in other words, has the requisite architectural knowledge. Similarly, if {1,2} knew their production technologies they could work to an output based measurement, while {3,4} who are ignorant about the relevant production technology may be measured on their actions. 20 By explicitly allowing the Designer to modify A (task architecture) we are clearly eschewing the strong form of technological determinism- where there is only one “true” underlying task architecture. Indeed, with a bounded rational designer, it appears to us that at least the strongest forms of technological determinism are impossible to sustain, as the Designer’s choices about A can only reflect his understanding at a point in time. Different Designers with different levels of knowledge may therefore enforce quite different A for the same basic set of tasks. Conversely, with sufficient knowledge (almost) any task architecture is feasible.9 5.2. Design as Feedback: Shaping Organizational Architecture Programming is feasible only with high levels of architectural knowledge. If such knowledge is not available to the designer, an alternative may be to let the agents communicate among themselves in order to create the shared knowledge made necessary by epistemic dependence. Even in this case, the designer must possess some level of architectural knowledge- for instance, he must know who is epistemically dependent with whom- but not enough to be able to appropriately redefine the epistemic structure. Instead, the designer can make the knowledge specified in E available to agents by creating appropriate channels for the agents to communicate directly with each other (enabling feedback). According to March and Simon coordination by feedback becomes necessary “[t]o the extent that contingencies arise, not anticipated in the schedule, (...) [which] requires 9 It is even possible to imagine adopting simultaneous scheduling for tasks that appear to have a logical sequence to them (task 2 depends on the output of task 1 as an input, for instance). As long as both agents in the dyad meet the knowledge requirements imposed by such a high level of epistemic dependence -the downstream agent would literally have to predict the upstream agents output before it was delivered, and the upstream agent would have to predict the ability of the downstream agent to do this etc.- the task architecture is still technically feasible (consider, for instance attempts at concurrent engineering in software development). However, a caveat here is that no physical good is necessary from upstream to downstream for downstream to start working- all that is required is a prediction of what the output will be. 21 communication to give notice of deviations from planned or predicted conditions, or to give instructions for changes in activity to adjust to these deviations.” (1958: 182) The marked distinction between coordination by plans and coordination by feedback is that the former is based on pre-established schedules, whereas the latter “involves transmission of new information” (1958: 182). In prior literature, grouping and linking have been considered as the two basic ways in which communication between agents can be induced (Nadler and Tushman, 1997). Grouping constitutes placing activities that need to be tightly coordinated within common organizational boundaries; linking refers to the creation of channels of communication and influence that link activities in different groupings. Grouping and linking jointly describe the organizational architecture. Specifically, the grouping decision leads naturally to a consideration of organization structure- as grouping results in sub-units within the organization. 10 Coordinating interdependent activities through grouping therefore implies the creation of an organizational architecture that eases communication between epistemically dependent agents. The distinction between programming and feedback thus corresponds to an important but often ignored distinction between task architecture and organizational architectural. Put simply, designer’s shape organizational structures when they cannot shape task architectures; the latter requires the designer to possess a higher level of architectural knowledge. 6. DISCUSSION: IMPLICATIONS FOR ORGANIZATION DESIGN 10 Two assumptions are implicit in the discussion of grouping: a) communication is easier within rather than between groups b) communication efficiency declines with group size. Both assumptions are necessary and jointly sufficient to explain the presence of organizational sub-units. 22 We now show how the approach we have developed can be used to state more precisely some well known ideas in the literature on interdependence, information processing and organization design. Our coverage of topics cannot be exhaustive, but we will select some examples with the idea of demonstrating the generative powers of the language we have constructed. We limit ourselves to a discussion of the underlying theoretical mechanisms in terms of the simple interdependent dyad – interested readers will find details of how to scale up to arbitrarily large sized organizations with both vertical and horizontal differentiation in Appendix B. 6.1. The Mapping Between Interdependence and Information Processing The central purpose of this paper was to develop an improved mapping between interdependence and the resulting need for information processing. In the task dyad, our analysis makes clear that four task architectures are feasible (Figure 2). Further, the lowest epistemic load (EL) and epistemic dependence (ED) occur with narrow incentives, while maximum EL and ED occurs with broad incentives and simultaneous moves. The task architecture with broad incentives and sequential moves is intermediate between these two extreme cases. Thus in the absence of prior knowledge (i.e. It=0), this ordering also captures the extent of information processing required. However, as our analysis makes clear, if the “blank slate” condition does not hold and It ∫0, then this ordering need not hold. To see, this consider Figure 3 below, which shows epistemic structures E1-E3 as well as associated information structures I1-I3 for three dyadic task architectures. ---------------------------------Insert Figure 3 about here ---------------------------------- It is obvious in Figure 3 that while EL(E1)>EL(E2)>EL(E3), it is also true that ρ 1> ρ 2> ρ 3. If the agents in these task architectures all began from blank slate conditions 23 with no prior knowledge (I1=I2=I3=0), we would have no hesitation in saying that information processing costs would be highest in the first case and lowest in the last. However, if the information structures are as shown in the figure, then in fact information processing costs (measured as a decreasing function of reliability) would be highest for the last case and lowest in the first. Put simply the history of a task architecture, which in turn may influence the current knowledge states of the agents (i.e. their information structures), cannot be ignored in assessing its information processing implications. How do our conclusions compare with received wisdom? We have already discussed Thompson’s (1967) conceptualization of interdependence into pooled, sequential and reciprocal categories extensively. In terms of our framework, Thompson’s rank ordering of sequential interdependence as higher than pooled interdependence in terms of information processing requirements can be justified only in special cases in our framework. If we assume that pooled interdependence (“when tasks depend on each other in a simple additive manner; each task renders a discrete contribution to the whole, and each is supported by the whole” pp. 54, 64) is associated with a narrow incentive structure and simultaneous task scheduling, whereas sequential interdependence implies broad incentives with sequential actions, then Thompson’s ranking holds Unless measured in this way, we would expect to see little if any empirical support for Thompson’s ordering in terms of information processing needs. As we have noted, Thompson’s notion of reciprocal interdependence (“when the outputs of each task become the inputs of the other” –p.55) is not logically definable at a point in time (because time to perform a task cannot be zero). However, it might well capture repeated instances of sequential task scheduling (i.e. task A feeds into task B, which then feeds into task A, which then feeds into task B etc.). Yet if the nature of the interaction does not change by period, then it is still not clear why this should necessarily generate higher information processing needs than pooled or sequential interdependence. There is also a 24 second possibility- perhaps reciprocal interdependence points to the need for repeated interaction because the agents lack knowledge of E or its elements. By interacting repeatedly, they may build such knowledge through communication. It is thus useful to look closely at cases of reciprocal interdependence to distinguish a) a pair of sequential architecture tasks that reverse direction after every period, but in which the relevant Epistemic Structures are fully known, from b) repeated interactions in a trial and error learning process meant to overcome uncertainty or ignorance about E. If the latter case applies, then this is better viewed as an organizational architecture meant to create feedback (i.e. communication between agents) rather than as a form of task architecture. But in either case, we would not expect strong evidence to support the idea that reciprocal interdependence must generate higher information processing needs than the other two categories of interdependence. More generally, in the absence of the blank slate condition (i.e. It=0), it is hard to see why we should expect any evidence for the increasing information processing costs of pooled, sequential, and reciprocal interdependence. Van de Ven et al. (1976) extended Thompson’s framework by adding an additional category of interdependence that they called “team”. In their own words: “Team work flow refers to situations where the work is undertaken jointly by unit personnel who diagnose, problem-solve and collaborate in order to complete the work. In team work flow, there is no measurable temporal lapse in the flow of work between unit members, as there is in the sequential and reciprocal cases; the work is acted upon jointly and simultaneously by unit personnel at the same point in time.” (p. 325) The authors conceptualize this fourth category to stand at the top of Thompson’s Guttman scale of task interdependence, requiring the most complex (group) coordination mechanisms. Within our framework, it is easy to see that team interdependence can be modeled as simultaneous scheduling with broad incentives. It would then generate the highest values of epistemic load and dependence and would indeed describe 25 a situation with the highest possible information processing costs, provided the blank slate assumption is met. Adler (1995) explicitly considered differences in situations where the actors in a task architecture begin from a relatively more or less knowledgeable state. Adler applied Thompson’s concept of task interdependence without modification; however, he distinguished between two types of uncertainty, namely the degree of fit novelty and the degree of fit analyzability. “The fit novelty of a project can be defined as the number of exceptions with respect to the organization’s experience of product/process fit problems.” (1995: 157) On the other hand, “[f]it analyzability can be defined (following Perrow (1967)) as the difficulty of the search for an acceptable solution to a given fit problem..” (1995: 158). Within our framework, these situations correspond to the absence of architectural knowledge on the part of the designer and/or low reliability structures (fit novelty), and difficulties in building the relevant knowledge (fit analyzability). Adler’s approach thus strongly aligns with our own conclusions that information processing requirements must depend not only on the nature of the task architecture (and resulting epistemic structure) but also existing knowledge endowments of the actors involved. Empirical measures of interdependence are likely to be useful only to the extent both these aspects are accounted for. 6.2 Span of Control, Delegation and the Depth of the Hierarchy The epistemic perspective offers us a clear distinction between the role of design vs. management: while the former involves manipulations to task architecture (programming) or organizational architecture (feedback) in an effort to enhance reliability at minimal information cost, the latter requires managing the exceptions that arise because of inadequacies of design (Galbraith, 1973). This calls for a view of hierarchical supervision as exception management rather than as simply telling the agents what to do. This is because if 26 the designer was sufficiently endowed with architectural knowledge to be able to simply specify action based contracts (i.e. telling agents what to do) – then this is by definition simply a form of coordination by programming. Whether the instructions on what to do are delivered in written plans or periodic oral communication does not appear to us to be a critical distinction. On the other hand, exception management by definition cannot be conducted by plans or schedules (else they would not qualify as exceptions)- indicating this is a distinct task for managers in addition to the task of designing. Of course, managers in an organization may possess both design and exception management responsibilities for their immediate subordinates or only the latter (also see Appendix B for how to model centralization vs. delegation of design). This role of managers in exception management provides a natural basis for limits to the span of control (Galbraith, 1973). Imperfect architectural knowledge should imply the need for exception management, and the greater the number of subordinates, the greater the efforts needed for exception management. If there is a finite limit to the exception management capacity of a manager, then there is a finite span of control. This in turn influences the need for delegation, as well as the number of hierarchical layers for any given total size of the organization. Thus, we should expect that limited architectural knowledge implies a limited span of control, resulting in greater delegation and deeper hierarchies. Conversely, flatter organizations require a high level of architectural knowledge on the part of designer. Thus, we should expect that combinations of limited architectural knowledge and imperfect information structures should imply more need for managing exceptions and a sharply limited span of control, resulting in greater delegation and deeper hierarchies. Conversely, flatter organizations require superior architectural knowledge on the part of designer or superior information structures for the agents. 27 6.3 Environmental Attributes and Information Processing The attributes of the organization’s task environment- comprising the organization’s consumers and users, suppliers, competitors, and regulatory groups- have also been argued to have information processing implications for the organization (Galbraith, 1973). A fair number of environmental characteristics relevant to a firm’s decision-making process have been proposed and analyzed. Environmental uncertainty, in much of the early information-processing literature (Burns & Stalker, 1961; Lawrence & Lorsch, 1967; Galbraith, 1973, 1977) is argued to be a primary driver of task uncertainty and consequently information processing. It is “caused by high levels of technological and market change requiring the organization to innovate in order to remain effective and competitive” (Donaldson, 2001: 22). Eisenhardt and Bourgeois (1988) argue that in high-velocity environments, “changes in demand, competition and technology are so rapid and discontinuous that information is often inaccurate, unavailable, or obsolete.”(1988: 818). Equivocality, has been defined by Daft & Lengel as a “measure of the organization’s ignorance of whether a variable exists in the space.” (1986: 567) and hostile environments are characterized by Burton and Obel by “precarious settings, intense competition, harsh, overwhelming business climates, and the relative lack of exploitable opportunities.” (1998: 171). Interestingly, the consequences of these various forms of environmental change can ultimately be described in epistemic terms. As Burton and Obel clarify (1998: pp. 167-171), the impact of environmental attributes can be traced through their impact on interdependence within the organization. Thus, we can represent the impact of environmental attributes in terms of an underlying epistemic and/or information structure. In other words, changes in the environment may a) exogenously alter the epistemic structure and/or b) the agents’ information structure leading in either case to a need for more information processing. 28 An example will highlight this point: Given a task structure with high epistemic interdependence (such as the one depicted in Figure 4a – with sequential tasks and broad incentives).Reliability will be highest with the information structure depicted in Figure 4b. However, the effect of environmental uncertainty, velocity, equivocality, and hostility is effectively a reduction in the agents’ knowledge as it becomes outdated – and hence a reduction in the structure’s reliability given that I (as depicted in Figure 4c) does not map as closely onto E anymore. We can thus explain why in the presence of environmental uncertainty leading to lowered probabilities of knowing what they need to know, task architectures which minimize epistemic load will always have higher reliability. However changes in environment may also alter the underlying epistemic structure. In particular, if the change is in the direction of greater epistemic dependence, then the need for information processing should also increase. Thus if the new epistemic structure is as shown in Figure 4d (simultaneous tasks with broad incentives), then we should expect to see a dramatic decrease in reliability as well as the need for substantial information processing. ---------------------------------Insert Figure 4 about here ---------------------------------- Thus, the epistemic perspective helps us to think of environmental changes as either affecting epistemic structure in the direction of greater epistemic dependence (for instance, by destroying the architectural knowledge of the designer), or weakening the information structure while preserving the epistemic structure. Each kind of change would result in a greater information processing requirements. It is therefore critical to parcel out these effects in empirical research- otherwise conclusions can be severely misleading. 29 7. CONCLUSION We have introduced an analytical framework that can help to refine our understanding of the links between interdependence, information processing and organization design. Specifically, building on the central concept of epistemic dependence, we show how it is possible to describe differences in the underlying coordination problems associated with different divisions of labor, and to compare the resulting implications for information processing. Through a few applications, we hope to have also illustrated that our approach is highly compatible with prior wisdom- but at the same time offers opportunities for refinement and novel theorizing. We have touched upon several topics central to the study of organization design such as coordination by plan vs. feedback, formalization, delegation, span of control, the depth of the hierarchy, delegation, management of exceptions and environmental uncertainty. However, we recognize that neither this list nor our coverage of these topics is exhaustive. Rather the goal in this paper has been to demonstrate the promise of a new approach to tackling these classic problems, through a few initial steps. We hope to take many more. 30 FIGURE 1 Papers on Interdependence & Design citing Thompson (1967) 120 Percentage 100 80 Percentage 60 40 20 19 67 19 70 19 74 19 77 19 80 19 83 19 86 19 89 19 92 19 95 19 98 20 01 20 04 20 07 0 Year FIGURE 2 The Epistemic Structures for Four basic Task Architectures for a Task Dyad incentive structure simultaneous K{oA1} K{oA2} incentive structure sequential K{oA1} K{oA2} π1, 2 π1, π2 A1 1 0 A2 0 1 A1 1 1 π1, 2 π1, π2 A1 1 0 A2 1 1 A2 0 1 A1 1 1 A2 0 1 31 FIGURE 3 Comparing Information Processing Costs across Epistemic and Information Structures E1 E2 E3 A1 A2 A1 A2 A1 A2 1 1 1 1 1 1 0 1 1 0 0 1 A1 A2 A1 A2 A1 A2 1 0.9 0.9 1 1 0.7 0 1 0.5 0 0 0.5 I1 I2 I3 FIGURE 4 Example of the effects of the Environmental Uncertainty on E and on I incentive structure E K:{oA1} K:{oA2} Figure 4a π1, 2 A1 A2 1 1 1 1 I A1 1 1 I A2 1 1 Figure 4b A2 0 0 E: π1,2 A1 A2 1 1 1 1 Figure 4c Figure 4d A1 1 0 32 APPENDIX A: SHARED KNOWLEDGE We present a formal definition of shared knowledge using standard notation from epistemic (modal) logic. K1j denotes that Agent 1 knows j Shared knowledge of first order j is shared knowledge (of first order- which is implied if not otherwise specified) among the j agents of a group G if K1j ⁄ K2j ⁄ K3j ….⁄ Kjj This is written as E1Gj and read “Everyone in group G knows j” Shared knowledge of any order By abbreviating EGEGn-1j = EGnj, and defining EG0j=j, we can define shared knowledge of j to any order n with the axiom: ∧ in=1 E nϕ Common knowledge is shared knowledge of order n =∝ Cϕ = ∧ in=1 E nϕ where n =∝ . APPENDIX B: FROM TASK DYADS TO COMPLEX ORGANIZATIONS Here we show how epistemic dependence can be represented for task architectures beyond the simple dyad. More complex structures are built through scaling (e.g. from dyad to n-task structures), delegation (managers at each level design task architectures for their subordinates) and recursion (e.g. a dyad in which each task is itself a dyad of tasks). 1. Simple scaling from task dyads to n-task structures The general n-task versions of the four basic epistemic structures in any task dyad are shown in Figure B1. Recall that a “1” in a cell indicates that the column actor needs to know row knowledge element for total utility to be maximized. When an agent i must act on the basis of predictions about the acts of any other agent j, there is epistemic dependence between agents. These cases appear as entries with non zero off-diagonals. 2. Building nested, hierarchical n-task structures The epistemic structures shown in Figure B1 are quite simple in that they describe a flat structure with a designer and a single layer of agents below him – there are no organizational subdivisions so that all agents are housed in a single department. The next step is to consider ways in which multiple hierarchical levels and horizontal subdivisions can be captured in our formalism. It turns out that we can proceed to build up such complex structures in two very different ways, either by delegation or centralization of organization design. Note that in 33 either case, the task of managing exceptions as they occur may still be delegated to subordinates. FIGURE B1 n-Task Versions of the Basic Epistemic Structures incentive structure simultaneous K{oA1} K{oA2} ... K{oAn} incentive structure sequential K{oA1} K{oA2} ... K{oAn} A1 1 0 ... 0 π1,π2 A2 ... 0 ... 1 ... ... ... 0 ... An 0 0 ... 1 A1 1 0 ... 0 π1,π2 A2 ... 0 ... 1 ... ... ... 0 ... An 0 0 ... 1 A1 1 1 ... 1 π1,2 A2 ... 1 ... 1 ... ... ... 1 ... An 1 1 ... 1 A1 1 1 ... 1 π1,2 A2 ... 0 ... 1 ... ... ... 1 ... An 0 0 ... 1 2.1 Delegation of design. In this approach, we assume “design is always one level deep”. The agents below the designer could themselves be middle managers who design for their subordinates and so on, giving rise to a hierarchy. Assume there are m middle managers, each responsible for a subgroup including one or more subordinates. Each middle manager is responsible for designing the task architecture for her own group, in addition to managing exceptions that arise because of inadequate design for their respective subunits. This has several implications: a) Only the responsible middle manager has detailed architectural knowledge and awareness of the epistemic conditions within her own unit. b) There is implicitly a design cycle and a production cycle. For instance in a three level organization, the organizational designer (typically the CEO), designs the incentive structure for middle managers. The middle manager’s action is based on these contracts. The middle manager in turn designs the incentive structure for the subordinates. The subordinates then complete their production cycle after which the middle managers complete theirs. Thus, we assume alternating cycles of design and production that have to move in serial from the bottom up, one layer in each time interval. c) Epistemic dependence occurs only within sub-groups, not between members in different sub-groups. The sub-groups are essentially modular units with high epistemic dependence within but not between them (Simon, 1962). Every sub-unit must have a middle manager-who aggregates the output of his subordinates- and these 34 managers may indeed be epistemically interdependent- but in general, individuals in one group are not interdependent with individuals in other groups. We show a simple example in Figure B2. In the structure depicted in Figure B2, the designer D delegates the design task for the lowest level of the hierarchy to middle managers M1 and M2. Thus, the designer only needs to know architectural knowledge necessary to define the incentives for M1 and M2 and the middle managers, in turn, need to possess architectural knowledge relevant to the sub-ordinates they directly oversee; in addition, given that M1 moves sequentially before M2, she also needs to know M2’s output function. A possible resulting Epistemic structure is shown in Figure B3. FIGURE B2 A Simple Hierarchy with one Layer of Middle Managers FIGURE B3 The Corresponding Epistemic Structure E to A in Figure B2 K{oD} K{oM1} K{oM2} K{oS1} K{oS2} K{oS3} K{oS4} K{oS5} K{oS6} K{oS7} K{oS8} M1 0 1 1 1 1 1 1 0 0 0 0 M2 0 0 1 0 0 0 0 1 1 1 1 S1 0 0 0 1 1 1 1 0 0 0 0 S2 0 0 0 1 1 1 1 0 0 0 0 S3 0 0 0 1 1 1 1 0 0 0 0 S4 0 0 0 1 1 1 1 0 0 0 0 S5 0 0 0 0 0 0 0 1 0 0 0 S6 0 0 0 0 0 0 0 0 1 0 0 S7 0 0 0 0 0 0 0 0 0 1 1 S8 0 0 0 0 0 0 0 0 0 1 1 2.2 Centralization of design. Centralization of design is the natural contrast to delegation of design. This approach assumes that a central designer (e.g. CEO) designs the entire organization so that each subunit is selfmanaged. In this case middle managers exist primarily to manage exceptions that arise within their sub-units. The only reason for layers of organization to exist in this case is task 35 sequencing. There are layers in the organization only to the extent that certain tasks must occur before others- there is therefore a “hierarchy of timing”. The formal characterization of the central design model is somewhat more involved than the delegation model because interdependencies across units influence every subordinate. Our aim is to formally specify the epistemic structure required to achieve coordination in any activity system comprising m groups of n members each. Epistemic structures are represented as before, and we use the general n-member versions of the basic epistemic structures that are shown in the above Figure B1. Building blocks It is useful to express the general n-member versions of the basic epistemic structures in matrix form. The three distinct epistemic structures shown in Figure B1 are shown in matrix form below, i.e. matrices I, L and J. The 0 and H matrices are helpers that allow us to use recursive expansion and thereby model organizations that vary in horizontal span and number of hierarchical layers. 1 0 L 0 0 1L 0 In = M M O M 0 0 L 1 1 0 L 0 1 1L 0 Ln = M M O M 1 1L 1 0 0 L 0 0 0 L 0 0n = M MO M 0 0 L 0 1 1 L 1 1 1L 1 Jn = M M O M 1 1L 1 0 1L 1 1 0 L 1 Hn = M M O M 1 1L 0 The general case defines an epistemic structure E composed of block matrices that each represents a distinct epistemic category. Coordination between m groups is regulated by matrices Bm, where Bm stand for any m × m matrix that can be made from the building blocks presented above. Coordination within subunits with n members is regulated by two matrices Wn, where Wn stand for any n × n matrix, including those that can be made from the building blocks presented above. Coordination within groups with arbitrary structures represented by Bm and Wn requires the following epistemic condition: ([ Bm ][ H m ]) ⊗ J n + I m ⊗ Wn E = ([ Bm ][ H m ]) ⊗ J n + I m ⊗ Wn This formalism extracts the overall epistemic matrix E for m groups that each employ n agents. The method used here is matrix expansion by tensor product. The advantage of this method is that there are virtually no restrictions on the specification of the two required input matrices: 1. A matrix Bm defining the epistemic structure among m groups. 36 2. A matrix Wn defining the epistemic structure among the n agents that are employed in each of the m groups. The number of agents n can vary across groups and there is no restriction on the size of m and n. In addition to the two required inputs, there are three “helpers”, the matrices Im (m × m identity matrix), Hm (m × m identity matrix with zeros on the diagonal), and Jn (n × n unit matrix). The helper matrices allow us to decompose the formal representation into two meaningful additive components. Each component is a matrix expansion by tensor product. The first tensor product represents coordination between groups and the second represents coordination within. As previously mentioned, square brackets denote element-by-element multiplication. This operation eliminates self-referential interdependencies in Bm (betweengroup interdependence). A simple example illustrates the formalism. Suppose there are two subunits and the epistemic structure between the two units, B2, is defined in terms of the basic task architecture I2 (narrow incentives). Further suppose that there are three members in each group and that the epistemic structure within the unit, W3, is defined in terms of the basic task architecture J3 (simultaneous actions and broad incentives). We then get: E = ([ B 2 ][ H 2 ]) ⊗ J 3 + I 2 ⊗ W3 = ([ I 2 ][ H 2 ]) ⊗ J 3 + I 2 ⊗ J 3 The resulting Epistemic structure is shown in Figure B4. FIGURE B4 Example of a Nested Task Structure: Two groups with three agents each A1 K{oA1} K{oA2} K{oA3} K{oA4} K{oA5} K{oA6} A2 1 1 1 0 0 0 A3 1 1 1 0 0 0 A4 1 1 1 0 0 0 A5 0 0 0 1 1 1 A6 0 0 0 1 1 1 0 0 0 1 1 1 Recursive expansion Multilevel systems can be generated by recursively expanding Wn while defining Bm at each recursive pass. This scheme is quite flexible because the number of groups, m, can be varied for each pass in the recursion (the number of agents, n, is defined from the expanding matrix Wn). A new epistemic structure Wn can even be applied in each recursive pass. Our recursive algorithm allows us to generate any epistemic structure E, given an initial definition of m, n, Bm and Wn as well as proper definitions of m and Bm for each subsequent pass in the recursion. 37 REFERENCES Adler, P. S. 1995. Interdepartmental Interdependence and Coordination: The Case of the Design/Manufacturing Interface. Organization Science, 6(2): 147 - 167. Baldwin, C. Y., & Clark, K. B. 2000. Design Rules. Cambridge, MA: The MIT Press. Burns, T., & Stalker, G. M. 1961. The Management of Innovation. London: Tavistock. Burton, R. M., & Obel, B. 1998. with contributions by S.D. Hunter III., M. Sondergaard, and D. Dojbak. Strategic Organizational Diagnosis and Design: Developing Theory for Application (2nd ed.). Boston/Dordrecht/London: Kluwer Academic Publishers. Burton, R. M., & Obel, B. 1984. Designing Efficient Orgnaizations: Modelling and Experimentation: North-Holland. Camerer, C. 2003. Behavioural Game Theory: Experiments in Strategic Interaction. Press: Princeton University Press. Clark, H. 1996. Using Language. Cambridge, UK: Cambridge University Press. Daft, R. L., & Lengel, R. H. 1986. Organizational Information Requirements, Media Richness and Structural Design. Management Science, 32(5): 554-572. Donaldson, L. 2001. The Contingency Theory of Organizations. Thousand Oaks, CA: Sage Dosi, G., Levinthal, D. A., & Marengo, L. 2003. Bridging contested terrain: linking incentive-based and learning perspectives on organizational evolution. Industrial and Corporate Change, 12(2): 413-436. Eisenhardt, K. M., & Bourgeois III, L. J. 1988. Politics of Strategic Decision Making in High-Velocity Environments: Toward a Midrange Theory. Academy of Management Galbraith, J. R. 1973. Designing Complex Organizations. Reading, MA: Addison-Wesley. Galbraith, J. R. 1977. Organizational Design. Reading, Mass.: Addison-Wesley. Grant, R. M. 1996. Toward a knowledge based theory of the firm. Strategic Management Journal,, 17: 109-122. Heath, C., & Staudenmayer, N. 2000. Coordination Neglect: How Lay theories of organizing complicate coordination in organizations. Research in Organizational Behavior,, 22(53-191.). Henderson, R. M., & Clark, K. B. 1990. Architectural innovation: The reconfiguration of existing product technologies and the failure of established firms. Administrative Science Quarterly,, 35(9-30.). Holmstrom, B. 1982. Moral hazard in teams. Bell Journal of Economics, 13(2): 324-340. Holmstrom, B., & Roberts, J. 1998. The Boundaries of the Firm Revisited. Journal of Economic Perspectives, 12(4): 73. Kelley, H. H., & Thibaut, J. W. 1978. Interpersonal Relations: A Theory of Interdependence: John Wiley & Son. Kollock, P. 1998. Social Dilemmas: The Anatomy of Cooperation. Annual Review of Sociology, 24(1): 183. Kretschmer, T., & Puranam, P. 2008. Integration through Incentives in Differentiated Organizations. Organization Science., 19(6): 860-875. Levinthal, D. A. (1997). "Adaptation on rugged landscapes." Management Science 43(7): 934. 38 Lawrence, P. R., & Lorsch, J. W. 1967. Organization and Environment: Managing Differentiation and Integration. Boston: Harvard Graduate School of Business Administration. March, J. G., & Simon, H. A. 1958. Organizations: Wiley. Milgrom, P. and J. Roberts 1990. The economics of modern manufacturing: Technology, strategy, and organization." American Economic Review 80(3): 511-28. Nadler, D. A., & Tushman, M. L. 1997. Competing by Design: The Power of Organizational Structure: Oxford University Press. Schelling, T. 1960. The Strategy of Conflict. Cambridge, MA: Harvard University Press. Simon, H. A. 1945. Administrative behaviour. NY: Macmillan Press. Simon, H. A. 1957. Administrative Behavior (2nd ed.): Macmillan. Thompson, J. D. 1967. Organizations in Action. New York: McGraw-Hill. Tushman, M. L., & Nadler, D. A. 1978. Information Processing as an Integrating Concept in Organizational Design. Academy of Management Review, 3(3): 613-624. Van de Ven, A., Delbeq, A., & Jr., K. R. 1976. Determinants of Coordination Modes within Organizations. American Sociological Review, 41(2): 322-338. Von Hippel, E. 1990. Task Partitioning: An Innovation Process Variable. Research Policy 19: 407-418. Wageman, R. 1995. Interdependence and Group Effectiveness. Administrative Science Quarterly, 40(1): 145 - 180. Weber, R. 2004. Organizational Coordination: A Game-Theoretic View. Carnegie Mellon University, Working Paper. Williamson, O. E. 1975. Markets and hierarchies: Analysis and antitrust implications. New York: Free Press. 39