Computing and Information Systems © University of Paisley 2005 Beyond Semantics: Verifying Information Content Containment of Conceptual Data Schemata by Using Channel Theory Yang Wang and Junkang Feng There is an ever-growing need to consider semantics in various research fields. Although numerous methods and solutions have been proposed and used along this line, some fundamental issues that need to be addressed still remain. The one that we are particularly interested in is how something that has semantics could have impacts on its receiver. It would seem that the impacts results from the capability of yielding knowledge. And by Dretske’s notion of ‘to know’, an essential component of this capability is information provision. Semantics has a role to play only if it contributes to the provision of information. Our approach to addressing this issue is therefore to use a number of theories that address the problem of information flow. In this paper we focus on the problem of ‘information content’ of some informationbearing objects or events, such as a piece of data that are constructed in some particular way, which we call ‘data construct’. To verify whether the information content of a given data construct contains a given piece of information systematically is not easy, particularly so when the information is beyond the literal meaning of the data. In this paper, we describe how a sophisticated theory about information flow, namely the Channel Theory, can be used for such a task. The main point we make in the paper is that this task can be accomplished through identifying nested information channels that cover both some intra-properties and inter-properties of a schema. Our work so far seems to show that this is a promising avenue. Such an approach introduces mathematical rigor into the study, and results in a sound means for the verification. This in turn provides the practice of conceptual modeling and validation with a desirable and reliable guidance. 1. INTRODUCTION Semantics has become an increasingly significant factor in various information system (IS) research fields such as system integration, knowledge representation and management, and semantic web services. In order to develop sound semantics-based approaches and methods, the need for solid theoretical bases cannot be overlooked. Our interest resides in 1 how semantics of something could have impact on its receiver. It would seem that the impact results from the capability of yielding knowledge. And by Dretske’s notion of ‘to know’ (Dretske 1981, p.92), an essential component of this capability is information provision. Therefore we have been exploring the relevance of semantic information and information flow theories, such as Dretske’s work (1981), Devlin’s work (1991) and the Information Flow theory (also called Information Channel Theory, CT for short) put forward by Barwise and Seligman (1997), to information systems. In this paper, we present our thinking on what we call ‘information content containment’, i.e., how an information-bearing object, such as a data construct, conveys information that constrains a given piece of information. Such a concept is put forward as an essential part of the notion of ‘Information Bearing Capability’ (IBC) by Feng and his colleagues in (Feng 1999, Xu and Feng, 2002, and Hu and Feng, 2002). This notion and many associated ideas are a result of observations through years of experience. There are four conditions identified that enable a data construct to represent a particular piece of information (which is called ‘the IBC Principle’ for convenience and will be given shortly), one of which is called the ‘information content containment’ condition. To verify whether these conditions hold of a particular system is not straightforward, as it is not simply a matter of checking the literal meaning of the data. To this end, we have been making considerable effort and achieved some preliminary yet encouraging results. Among what have been achieved, it was found that employing a sophisticated theory of information flow within a notional distributed system, namely the Channel Theory (CT) introduces much needed rigor into this work and can serve as a sound means for verifying whether the ‘Information Content Containment’ condition is satisfied over any given pair of two elements, one of which is a data construct and the other a piece of information. That is, whether the ‘information content' of the data contains the information. In this paper we describe our approach by using examples in conceptual data modeling, and avoid any lengthy description of the CT per se. The rest of the paper is organized as follows. In the next section, the IBC Principle is introduced. After that, the reasons why we choose the Channel Theory as our intellectual tool for the job in hand are described. In Section 5, we present details on how we verify the ‘Information Content Containment’ condition by using the method of CT summarised in section 4. In Section 6, some related works that are concerned with ‘information content preserving’ in schema transformation and with contributions of CT to semantic interoperability and ontology mapping are briefly summarised. Finally, we give conclusions and indicate directions for further work. 2. THE ‘IBC’ PINCIPLE AND ITS RELEVANCE In practice, due to the lack of understanding of the difference and link between information and data, problems can occur. For example, it is difficult to identify redundant or conflict information requirements; the transformation from human level models to machine level design and implementation may not be information content preserving; a query to a system may receive unsound and/or incomplete answers. An underlying reason for this could be that the system’s capability of bearing information is overestimated or mis-interpreted. Some of such problems were recognized as ‘connection traps’ by Codd (1970) and Howe (1989), and discussed in detail in Feng and Crowe’s work (1999). We envisage that the ideas of ‘Information Bearing Capability’ should help. The notion of IBC and the work around it were developed over a number of years as shown in series of works (Feng 1999; Xu and Feng 2002 and Hu and Feng, 2002). The version of it that we present here is that which takes into account the three major parties that are involved in information flow, namely information source (S), information bearer (B), and information receiver (R). We call such a thought the framework of SBR for simplicity and convenience purposes. The IBC Principle (Feng, 2005) is now specified as follows: This Principle is concerned with token level (in comparison with the ‘type level’) data (or media) constructs’ representing individual real world objects and individual relationships between some real world objects. For a token level data construct (or a ‘media construct’ in general), say t, to be capable of representing an individual real world object or an individual relationship between some real world objects (or a ‘referent construct’ in general), say s, which is neither necessarily true nor necessarily false1, 1 See Floridi (2002). 2 The information content of t when it is considered in isolation must include s, the simplest case of which is that the literal or conventional meaning of t is part of its information content, and the literal or conventional meaning of t is s; And t must be distinguishable (identifiable) from the rest of the data constructs in a system, say Y, that manages data including t (or from the rest of the media constructs in a system Y that manages media constructs including t, in general). The above two conditions were formulated from the viewpoint of the relationship between S and B under the SBR framework. For a data construct (or a ‘media construct’ in general) t that is capable of representing an individual real world object or an individual relationship between some real world objects (or a ‘referent construct’ in general) s to actually provide information about s, t must be accessible by the only means available to system Y; In the case that t has neither literal nor conventional meaning and in the case that neither the literal nor the conventional meaning of t is s, the information receiver, i.e., the R in the SBR framework, must be provided with means by Y to infer s from t. The above two conditions were formulated from the viewpoint of R’s obtaining information about S via B under the SBR framework. This Principle is scalable to more (i.e., t could be an instance of an entire system among many related systems, for example) or less (e.g., an instance of a simple attribute of an entity in an ER schema) complex cases, and hence flexible in terms of applicability. We call the first condition of the above four the ‘Information Content Containment’ condition. And the concept of ‘information content’ of a sign/message can be defined as follows (Dretske 1981, p.45): ‘A state of affairs contains information about X to just that extent to which a suitably placed observer could learn something about X by consulting it.’ In this paper, we focus on how we could verify whether a state of affairs contains information about X. And our approach is to make use of theories on information flow. 3. SOME THEORIES ON INFORMATION FLOW 3.1 Dretske’s Semantic Theory of Information Dretske put forward a theory (Dretske, 1981), which not only captures, following Shannon (Shannon and Warren, 1949), the quantitative aspect of communication of information, but also addresses the information content of an individual message. These form an account of how information can flow from its source to a cognitive agent, i.e., the receiver. Although his theory has been widely cited, we also note its objections. Firstly, as Dretske’s theory is based upon probability, certain conditions must be maintained for a probability distribution to occur. In the ‘real world’, this might be a stringent requirement. Secondly, Dretske includes the ‘internal’ contribution of the prior knowledge k and the ‘external’ contribution of objective probabilities to the conceptualisation of information flow. However, Dretske has to accept that different ways of determining relevant possibilities and precisions give different probability measures and therefore different information flow and knowledge (Barwise and Seligman, 1997). Finally, Dretske goes beyond Shannon’s work (1949) that concerns solely the quantitative aspect of communication with many messages (Devlin, 2001), and tackles explicitly the content of information that an individual message bears. And yet it is difficult to identify the information content of an individual message. 3.2 Devlin’s ‘Infon’ and Situation Theory To model information flow, the mechanism used by Devlin (1991) is made up of situation types and constraints, which connect situation types. This theory seems to emphasise the ‘soft’ aspect of information flow, i.e., what is going on in people’s mind. Moreover, it does not seem to address the issue of how a constraint gets established, which would not rule out the possibility of them being arrived at subjectively. 3.3 The Information Channel Theory Taking into consideration the shortcomings of the above two theories on modelling information flow, we choose the Channel Theory (Barwise and Seligman, 1997) as a tool to verify whether a given data construct satisfy the ‘information content containment condition’ of the aforementioned IBC Principle. The reasons why we think that this is appropriate are summarised below. In general, saying that ‘B bears information about S’ is the same as saying ‘there is information flow from S to B’. And the latter is in the language of CT, which is a systematic approach to mathematically modelling and analysing information flow. 3 With CT, information flow is possible only because what is involved in information flow can be seen as components of a distributed system. That is to say, the notion of ‘distributed system’ provides us with a way to model and formulate, if possible, an ‘information channel’ within which information flows. Information flow requires the existence of certain connections (which may be abstract or concrete) among different parts that are involved in information flow. Following Dretske (1981, p.65), such a ‘connection’ is primarily made possible by the notion of ‘conditional probability’ where the condition is the Bearer. However, this is perhaps only a particular way for ‘connections’ to be established. It would be desirable to see in general how a connection becomes possible. The notion of ‘information channel’ in CT helps here. In CT terms, information flow is captured by ‘local logics’ (which are roughly conditionals between ‘types’) within a distributed system. Most frequently, information connections lie with ‘partial alignments’ (Kalfoglou and Schorlemmer, 2003) between system components. CT is capable of formulating such alignments by using concepts of ‘infomorphisms’ ‘state space’, and ‘event classification’. The property of ‘state space’ enables every instance and state relevant to the problem to be captured. Consequently, informational relationship between components connecting through the whole system can be modelled as the inverse image of projections between their corresponding state spaces. Such inverse images are also known as ‘infomorphisms’ between event classifications. This is exactly how the ‘infomorphisms’ for the ‘core’ (which is a classification for the whole distributed system) of an information channel can be found. We ask our readers, who are not familiar with CT, to bear with us here as we will use example to show the basics of CT shortly. 4. A METHOD OF USING CT TO VERIFY INFORMATION CONTAINMENT We have developed a method for the task at hand, which consists of a series of steps below. The reader is referred to the appendix for those numbered terms: Identify component classifications[1] relevant to the task at hand and then use them to construct a distributed system; Validate the existence of infomorphisms[2] between the classifications and construct an information channel[3] relevant to the task at hand; Construct the core of the channel by identifying those parts of the component classifications (including normal tokens[4]) that contribute to the desired information flow; Find the local logic[5] on the core, i.e., the entailment relationships between types of the classification for the ‘core’ that is directly relevant to the information flow, i.e., the information content containment at hand; Arrive at desired system level’s theory applying the f-Elim rule core. [7] [6] by on the local logic on the We describe this method in detail by means of examples in the next Section. 5. VERIFYING CONTAINMENT INFORMATION We believe that to verify whether the information content of the schema contains something about a specific ‘real world’ application domain takes two steps, which can be roughly seen as corresponding to the syntactic level and the semantic level of (Organisational) Semiotics (Stamper, 1997; Liu, 2000 and Anderson 1990). We use Figure 1 to show our points. has has Marks (0,n) (1,1) finished student (1,1) (1,n) Qualifications added Figure 1 Topological connection t 5.1 Semantic Level On this level, the problem being addressed is ‘meaning, propositions, validity, signification, denotations,…’ (Stamper, 1997) Accordingly, what should be looked at is how a topological connection, which is a connection between entities made possible by an Entity-Relationship (ER) schema, is able to represent a real world relationship. That is to say, this level is concerned with the relationship between a conceptual data schema and the ‘real world’ that the schema models. Specifically, we want to see whether and how information flows from the ‘real world’ to the schema. In other words, the information content of the schema includes something about a specific ‘real world’ application domain. In the world of CT, information flow is ever possible only within a distributed system, i.e., a system made up of distinct components, the behaviour of which is government by certain regularities. Within such a system, ‘what’ flows is captured by the notion of 4 ‘local logics [5]’ and ‘why’ information can flow is explained by the concept of ‘information channel [3]’. Therefore, it is essential to construct relevant and appropriate ‘distributed systems’. Considering the condition of ‘information content containment’ of the IBC Principle under the SBR framework, the information source is the ‘individual real world object or individual relationship between some real world objects (or a ‘referent construct’ in general) s’, and the information bearer is the ‘topological relation t’. On this level then, to make sure that a conceptual data schema is capable of representing some particular real world individuals involves constructing a distributed system justifiably that supports information flow between these two different types of things. The ‘connections’ between them, which are instances of the distributed system, are crucial. We view the process of constructing a conceptual schema for representing some particular real world as a ‘notional system’.2 It is interesting to note that the notion of ‘distributed system’ in CT is similar to the concept of ‘notional system’ in Soft Systems Methodology (SSM) (Checkland, 1981). In addition, sometimes we use the term ‘semantic relations’ to refer to something in the ‘real world’ in contrast to ‘topological connections’, which are elements within a data schema and can be seen as something on the syntactic level. Semantic relations are not unlike the notions of ‘objects in the “reality”’ and ‘social conventions’ in semiotics (liu, 2000). In such a distributed system (It is called ‘representation system [8]’in particular), the semantic relations and the conceptual schema can be seen as components. Using the notions of CT, the conceptual model is our source classification (not to be confused with the ‘information source’ in aforementioned SBR), and the semantic relation is target classification. The tokens of the schema consist of all the instances of data constructs, i.e., the topological connections between data values. The types of this classification are different data construct types. An entity, a relationship or a path in a conceptual model can all be different ‘types’. As the process of ‘constructing a schema’ is modelled as a distributed system, the classification that serves as the core of our information channel is made up of connections that are causal links between our conceptual model (i.e., the schema) and what it models for and types that are ways of classifying these links. The logic on this classification captures the reasoning of the schema construction. That is, the 2 See Checkland (1981) constraints made up the logic on the ‘core’ of the system represent the rules employed by the database designer about how to modeling real world objects to the conceptual model. The normal tokens of the logic are the links that must satisfy the constraints, among all possible tokens of the system. Now let us use the example in Figure 1 to illustrate our ideas. Assume that there is a semantic relation s, say ‘After having successfully passed all required modules, a student receives a new qualification (a degree or diploma)’. We show how the topological connection t (as shown in Figure 1) comes to represent the semantic relation s. The information channel and infomorphisms for justifying this are shown in Figure 2. f1 (topological connection between ‘student’, ‘marks’ and ‘qualifications’) ├ ζ C f2 (semantic relation between ‘student’, ‘marks’ and ‘qualifications’) In the above entailment relation, f1 (topological connection between ‘student’, ‘marks’ and ‘qualifications’) is the result on the core of applying function f1 to a type of the classification CM, namely topological connection between ‘student’, ‘marks’ and ‘qualifications’. f2 (semantic relation between ‘student’, ‘marks’ and ‘qualifications’) can be interpreted the same way. The f-Elim Rule allows us to move from this constraint in the logic on the core of the channel to the component level, and we have: topological connection between ‘student’, ‘marks’ and ‘qualifications’ ├CM+SR semantic relation between ‘student’, ‘marks’ and ‘qualifications’. Figure 2 Information channel diagram 1 In this diagram, C is the ‘core’ of the information channel that is the process of constructing a schema like the one in Figure 1; CM and SR are classifications of the conceptual model (e.g., a data schema) and a real world application domain respectively. Infomorphisms f1: CM C and f2: SR C enable the causal relations between these two component classifications and the core, which is also a classification. Combining these two infomorphisms, we obtain an infomorphism f = f1 + f2 from CM+SR to C. All of them formulate, following BS97 (Barwaise and Seligman, 1997, p.235), a representation system R. In the example, business constraint that ‘if a student has passed all required modules, he/she is awarded a new qualification’ can be captured as part of the local logic ζ on the core, namely: Note that this move is only valid for those tokens that are actually connected by the channel. This is determined by the properties of the f-Elim Rule, namely it preserves non validity (i.e., ‘completeness’) but not validity (i.e., not ‘soundness’). The local logic ζ that holds on the core, i.e., the distributed system involving topological connections within the schema and semantic relations within the real world application domain takes, as a prerequisite, the tokens of the classification CM (Conceptual Model, i.e., the schema) that are required for linking to those of the classification SR (Semantic Relation, i.e., the ‘real world’) in order to form the instances of the core do exist. This would rely upon the inner relationship within the model, which may be captured as information flow on a lower level – what we call syntactic level. That is to say, syntactic level information flow supports that of semantic level. 5.2 Syntactic Level topological connection between ‘student’, ‘marks’, and ‘qualifications’ R semantic relation between ‘student’, ‘marks’ and ‘qualifications’ To formulate it in CT terms results in an entailment relation between types of the ‘core’ classification: 5 On this level, we look at the inside of a schema (it is the source system as mentioned above) and find out whether and how a part of the schema may bear information about another, namely if information flows between parts of the schema. This is a type of constrains that the schema must adhere to, and eventually the database that is constructed according to the schema must also. Therefore information flow within a schema and a database plays a role in determining what a conceptual schema is capable of representing. For the example shown in Figure 1, the problem to be addressed here is what information flow must exist within the path as a necessary condition in order to enable the aforementioned information flow on the semantic level. In CT terms, this is a matter of making sure that the tokens of the classification CM (Conceptual Model, i.e., the schema) that are required for linking to those of the classification SR (Semantic Relation, i.e., the ‘real world’) in order to form the ‘connections’ (i.e., the instances of the core) do exist. As above, the task is to find out whether it is possible to construct an appropriate ‘distributed system’ and an associated information channel that would support the desired logics. We will use the same example in Figure 1. But this time, we will look at ‘informational relationships’ between parts of the schema rather than that between the real world and the schema as a whole. Specifically, we examine whether and how the entity ‘mark’ actually provides information about entity ‘qualifications’ by both connecting to the entity of ‘student’. On this level, the idea of ‘information source’ and ‘information bearer’ still applies. The relationship ‘marks-student’ is now the information bearer, t, while the relationship ‘student-qualification’ is the information source, s. The notional relevant distributed system will be constructed by using these two relationships accordingly. ‘There are many ways to analyze a particular system as an information channel’, and ‘if one changes the channel, one typically gets different constraints and so different information flow’ (Barwise and Seligman, 1997, p.43). We will use entity ‘student’ to illustrate how to construct a distributed system and information channel for a particular purpose. To help this, the schema is now extended as shown in Figure 3. The tokens a, a’….of A are individual ‘markstudent’ relationships at various times. There are many ways to classify the tokens. The way adopted here is ‘a student receives his/her final mark for a module’. For example, there are marks for ‘SPM’ (‘Software Project Management’ module), ‘ISTP’ (‘Information Systems Theory and Practice’), etc. These then become the types of the classification. Classification B: for relationship ‘studentqualification’. Like classification A, the tokens b, b’ …of B consist of individual ‘student-qualification’ relationships at various times. To classify these tokens we use types like ‘a student is awarded a particular qualification’. Examples of the qualifications are ‘BSc IT’, ‘BA BA’, etc. State space [9] SA: for classification A. The tokens x, x’… of SA consist of individual ‘mark-student’ relationships at various times. The states consist of 0 and 1 for each model. The state of si (i is the name of a model.) is 1, if ‘a student receives his/her final mark that is no less than 50 for a module’. Otherwise it is 0. For example, if MSPM ≥ 50, then sSPM = 1SPM, otherwise sSPM = 0SPM. Consequently, the set of state rA for SA should consist of each state of every included module, namely rA = {{0SPM,, 0ISTP …, 0OAD}… {1SPM , 1ISTP…,1OAD}}. If there are altogether 8 modules, the cardinality of rA is 28 = 64. Event classification [10] Evt(SA ). The event Classification A associated with SA, namely Evt(SA) has the same tokens as SA, but the types of it will be all the subsets of rA, for instances, {Φ}, {0SPM , 0ISTP,…, 0OAD}, …, {{0SPM , 0ISTP…, 0OAD}, {0SPM , 0ISP …, 1OAD}},…, {{0SPM , 0ISTP…, 0OAD}, {0SPM , 0ISTP…, 1OAD}, …, {1SPM , 1ISTP…, 1OAD}}. That is, Evt(SA) is the power set of SA. The total number of types of Evt(SA) is therefore 264. The relationships between the types and the tokens of A and Evt(SA) follow the tokenidentical infomorphism gA: A Evt(SA ) (Barwise and Seligman, 1997, p.55) State space SB: for classification B. Figure 3 Extanded Schema Now, we can define the ‘classifications’ involved. Classification A: for relationship ‘markstudent’. 6 The tokens y, y’… of SB consist of individual ‘student-qualification’ relationships at various times. Like SA, the states are also 0 and 1 for each qualification. The state of sj (j is the name of a qualification.) is 1, if the ‘degree’ attribute of ‘student-qualification’ is set to an appropriate degree name. The state of sj is 0 if there is no new degree is added. For example, if degree, BSc BIT, is added, the state SBSc BIT = 1, otherwise, SBSc BIT = 0. Therefore, the set of state rB for SB should consist of each state of every added degree, such as rB = {1BSc BIT, 0BA BA …}. If there are altogether two possible added degrees, the cardinality of rB is 22 = 4. Event Classification Evt (SB ). The same tokens are present in the event classification Evt(SB). Similarly, the types of it are also all the possible subsets of rB. For example, {Φ}, {0BSc BIT }, {0BA BA }, …, {0BSc BIT , 1BA BA },…, {1BSc BIT , 0BA BA},…, {0BSc BIT , 0BA BA , 1BSc BIT , 1BA BA}. If there are totally two degrees, the number of types for Evt(SB) will be 24. Also, there is a natural infomorphism gB: B Evt(SB ). We define a classification, say W, on which a desired local logic lives, which is the sum of Classification A and Classification B[11], namely W = A + B. Any token of this classification consists of two parts, <a, b>, where a and b are instances of the ‘mark-student’ relationship, and the ‘student-qualification’ relationship respectively. The types of W are the disjoint union of the types of classification A and B. It is important to notice that there are no one-to-one relationships between tokens of classifications A and B. Although W connects A and B, not all tokens of W represents meaningful and useful connections between their tokens. For example, a student might have marks, which will not relate to a certain degree if he changes his stream of a course in the middle of a semester. That is to say, only a subset of the tokens of W actually participates in the actual information flow. Therefore, classification W is not the information channel that we are after. Our aim is to find the partial alignment (Kalfoglou and Schorlemmer, 2003) for the core of channel. To this end, we define state space S for classification W. The tokens c, c’…. of S are arbitrary instances of the classification W at various times. The set of states of S is {rA, rB}. An instance, say w, of classification W is in state <rA1, rB1>, if the state of w’s ‘mark for a module’ part is rA1, and the state of w’s ‘qualification awarding’ part is rB1. There are natural projections [12] associated with state space S, i.e., pA: S SA, pB: S SB. In order to find the real informational relationship, i.e., that a student has passed all required modules for a course bears the information that the student is awarded a certain degree, we need to restrict the state space by eliminating those invalid states, such as 7 <{0SPM, 1ISTP, …1OAD }, {1BSc BIT }>. As a result, we have a subspace S*of S. This subspace inherits all the properties of space S including the natural projections. As we did above regarding Evt(SA ) and Evt(SB), we can find the event classification Evt(S*) for S*. Such an event classification enables the existence of partial relationships between tokens of Evt(SA ) and Evt(SB ). This is what we want to model as the core of the information channel. According to a proven proposition, i.e., Proposition 8.17[13] (Barwise and Seligman, 1997, p.109), there are infomorphisms between event classifications Evt(SA), Evt(SB) and Evt(S*) on inverse directions of the natural projections between them. The infomorphisms from the component classifications to the core of the information channel C are defined as follows. The infomorphism fA: A C is the composition of the infomorphisms gA: A C. Evt(SA) and Evt(pA ) : Evt(SA ) The infomorphism fB: B C is the composition of the infomorphisms gB: B C. Evt(SB) and Evt(pB) : Evt(SB) The relationship between information channel C, component classifications A and B can now be seen in Figure 4. Figure 4 Information channel diagram 2 The core Classification C supports a local logic LC, which is concerned with entailment relations between sets of states of state space S. Here, as shown in the conceptual schema diagram, following a regulation in this particular organisation, there is a system level constraint, namely ‘if having passed all required modules like SPM, ISTP …(5), OAD , the student obtains a new degree, BSc BIT’. Resulted from this, there should be a constraint supported by the LC: fA (a student gets final marks that are no less than 50 for SPM, ISTP …(5),OAD) ├Lc fB (a student is awarded BSc BIT degree) If we apply the f-Elim Rule to infomorphisms, fA and fB, this logic can be moved (translated) from the core to the level of the component classifications A + B as a regular theory: a student gets final marks that are no less than 50 for SPM, ISTP …(5), OAD├ Ls a student is awarded BSc BIT degree This constraint is valid only under the condition that relationships ‘marks-student’ and ‘studentqualification’ are satisfied through the channel C. We would like to point out before we leave this section that the information channels on the ‘syntactic’ level are conceptualisation of some particular ‘intra properties’ of a schema, which are actually nested inside the channels on the ‘semantic level’ that involve the schema and the real world that it models. And the latter captures what might be called ‘inter properties’ of some kind of the schema. 6. RELATED WORKS Sophisticated concepts stemming from CT were formulated for explorations on semantic information and knowledge mapping, exchanging, and sharing among separate systems. Kent (2002b; 2002a) exploits semantic integration of ontologies by extending firstorder-logic-based approach (Kent, 2000) based on CT. An information flow framework (IFF) has been advocated as a meta-level framework for organising the information that appears in digital libraries, distributed databases and ontologies (Kent, 2001). Based on Kent’s work, Kalfoglou and Schorlemmer developed an automated ontology mapping method (Kalfoglou and Schorlemmer, 2003a) by using concepts of CT in the field of knowledge sharing. Furthermore, they extended their thoughts into developing a mechanism for enabling semantic interoperability (Kalfoglou and Schorlemmer, 2003), which seems to be a mathematically sound application of CT. All these research results show that CT is a powerful intellectual tool for finding relationships between items with semantics. In this paper, we have shown how CT might be used in conceptual modeling in terms of determining whether a conceptual model is capable of ‘containing’ certain required information. 7. CONLUSIONS AND FUTURE WORK The work we present here draws on a number of works and is a result of addressing something that they seem to have missed. Conceptual modelling is considered as a valuable tool in IS development. However, it would seem that fundamental questions, such as ‘what enables conceptual models to be what they are, and why they are capable of providing required information’, have not been answered convincingly. Hull (1986), Miller et al (1993; 1994) and Kwan and Fong (1999) advocate a theory concerning the ‘information capacity’ (IC) of a data schema, which is close to our ideas. Their works focus on schema transformation and integration by using the mathematical notion of mapping and developing correctness criteria (1993). Measures of equivalence or dominance of schemas based on the preservation of the information content of schemas are offered (Batini et al., 1992; Miller, 1994). In addition to IC, other relevant works are in the area of analysing relationships between entities during conceptual modelling of real-world applications (Dey et al., 1999) and meanings of relationships on ontological aspects of investigation (Wand et al., 1999). One thing that seems in common among these theories and approaches is that they concentrate on the syntactic aspect of the problem. The notion of ‘information content’ seems to be defined intuitively and concerned only with the data instance level. We believe that it would be worth extending our attention beyond data, and looking at sufficient conditions that might exist that enable data constructs to represent information. 8 Why does a conceptual model that uses meaningful building blocks have potential impacts on its user? We observe that the potential impacts come from the information provision (bearing) capability of the conceptual model. The semantics of the building blocks do have a role to play but it is so only when it contributes to the provision of information. We have not gone into details to argue for this observation in this paper, rather we have provided a way for verifying whether the information content of a data construct contains a given piece of information about something. The main finding presented in this paper is that by identifying and articulating nested information channels that cover both some intra-properties and inter-properties of a schema this task can be accomplished systematically. Due to space constraints, we are unable to present our work on other conditions under the umbrella of the IBC Principle. But our aspirations in pursuing it and our approach have been described in this paper. The ideas embodied by the IBC Principle do seem to make sense, helpful, and moreover justifiable not only by using Dretske’s theory (1981) and Devlin’s theory (1991), on which the Principle is based, but also by using Channel Theory (Barwise and Seligman, 1997). This work also extends the application scope of Channel Theory into conceptual modeling for information systems. This is encouraging. We will continue our investigation along this line to further develop our theoretical thinking and to consolidate and further explore the practical relevance of our theoretical thinking, for example, in the areas of schema optimization and query answering capability of a schema. References Anderson, P. B. (1990). “A Theory of Computer Semiotics: Semiotic Approaches to Construction and Assessment of Computer Systems”, Cambridge, University Press, Cambridge. Barwise, J. and Seligman, J. (1997). “Information Flow: the Logic of Distributed Systems”, Cambridge University Press, Cambridge. Batini, C., Ceri, S. and Navathe, S. B. (1992). “Conceptual Database Design: An EntityRelationship Approach”, The Benjamin/Cummings Publishing Company, Inc. Redwood City, California. Checkland, P. (1981). “Systems Thinking, Systems Practice”, John Wiley & Sons, Chichester. Codd, E. F. (1970). “A Relational Model of Data for Large Shared Data Banks”. Com. Of ACM, 13.6, pp. 377-387. Devlin, K. (1991). “Logic and Information”, Cambridge. Devlin, K. (2001.) “Introduction to Channel Theory”, ESSLLI 2001, Helsinki, Finland. Dey, D., Storey, V. C. and Barron, T. M. (1999). “Improving Database Design through the Analysis of Relationships”. ACM Transactions on Database Systems, 24(4):453-486. Dretske, F.I. (1981). “Knowledge and the Flow of Information”, Basil Blackwell, Oxford. Feng, J. (1999). “An Information and Meaning Oriented Approach to the Construction of a Conceptual Data Schema”, PhD Thesis, University of Paisley, UK. Feng, J. (2005). “Conditions for Information Bearing Capability”, Computing and Information Systems Technical Reports No 28, University of Paisley, ISSN 1461-6122. Feng, J. and Crowe, M. (1999). “The Notion of ‘Classes of a Path’ in ER Schemas”. In Proceedings of Third East European Conference on Advances in Databases and Information Systems, ADBIS’99, Springer, Berlin, ISBN 3-540-66485-8. Floridi, L. (2002).” What is the Philosophy of Information? Mataphilosophy”, 33(1-2), 123-45. Howe, D. R. (1989). “Data Analysis for Database Design”, 2nd edn, Edward Arnold, London. Hu, W. and Feng, J. (2002). “Some considerations for a semantic analysis of conceptual data schemata”, in Systems Theory and Practice in the Knowledge Age. (G. Ragsdell, D. West. J. Wilby, eds.), Kluwer Academic/Plenum Publishers, New York. 9 Hull, R. (1986). “Relative Information Capacity of Simple Relational Database Schemata”, SIAM Journal of Computing, 15(3):856-886. Kalfoglou, Y. and Schorlemmer, M. (2003). “Using Information Flow Theory to Enable Semantic Interoperability”, In Proceedings of the 6th Catalan Conference on Artificial Intelligence (CCIA '03), Palma de Mallorca, Spain, October 2003 Kalfoglou, Y. and Schorlemmer, M (2003a). “IFMap: an ontology mapping method based on information flow theory”. Journal on Data Semantics, 1(1):98– 127. Kent, R. E (2000). “The Information Flow Foundation for Conceptual Knowledge Organization”. In: Dynamism and Stability in Knowledge Organization. Proceedings of the Sixth International ISKO Conference. Advances in Knowledge Organization 7 (2000) 111–117. Ergon Verlag, Würzburg. Kent, R. E. (2001). “The information Flow Framework”. Starter document for IEEE P1600.1, the IEEE Standard Upper Ontology working Group, http://suo.ieee.org/IFF/. Kent, R. E. (2002b). “The IFF Approach to Semantic Integration”. Presentation at the Boeing MiniWorkshop on Semantic Integration, 7 November 2002. Kent, R. E. (2002a). “Distributed Conceptual Structures”. In: Proceedings of the 6th International Workshop on Relational Methods in Computer Science (RelMiCS 6). Lecture Notes in Computer Science 2561. Springer, Berlin. Kwan, I. and Fong, J. (1999), “Schema Integration Methodology and Its Verification by Use of Information Capacity”, Information Systems, 24(5):355-376. Liu, K. C. (2000). “Semiotics in Information Systems Engineering”, Cambridge Press. Miller, R. J., Ioannidis, Y. E. and Ramakrishnan, R. (1993). “The Use of Information Capacity in Schema Integration and Translation”, in Proceedings of the 19th International Conference on Very Large Data Base, Morgan Kaufmann, San Francisco. Miller, R. J., Ioannidis, Y. E. and Ramakrishnan, R. (1994). “Schema Equivalence in Heterogeneous Systems: Bridging Theory and Practice”. Information Systems, 19(1):3-31. Shannon, C. E. and Warren, W. (1949). “The mathematical theory of communication”. University of Illinois Press. Stamper, R. (1997). “Organizational Semiotics. In Information Systems: An Emerging Discipline?”, (J. Mingers and F.A. Stowell), McGraw-Hill, London. Wand, Y., Storey, V. C. and Weber, R. (1999). “An Ontological Analysis of the Relationship Construct in Conceptual Modeling”. ACM Transactions on Database Systems, 24(4):494-528. Xu, H. and Feng, J. (2002). “The ‘How’ Aspect of Information Bearing Capacity of a Conceptual Schema at the Path Level”, . The 7th Annual Conference of the UK Academy for Information Systems, UKAIS'2002. Leeds . ISBN 1-898883-149, pp.209-215 Appendix 1. A classification is a structure A = <U, ∑A, ╞A > where U is the set of objects to be classified (the tokens of A), ∑A the set of objects used to classify the tokens (the types of A), and ╞A is a binary relation between U and ∑A determining which tokens are of which type. 2. Let A and C to be two classifications. An infomorphism between them is a pair f = <f, f > of functions. For all tokens c of C and all types α of A, it is true that f (c) ╞A α iff c ╞C f (α) This is also to say that two function f and f are in opposite directions: f: A C. In it, f-up maps tokens form A to C. f-down maps types from C to A. 3. Information channel consists of an indexed family C= {fi: Ai C} i I of infomorphisms with a common co-domain C, the core of the channel. 4. A subset of tokens that really participate in the information flow are called normal tokens. 5. A local logic L = <A,├L , NL> consists of a classification A, a set of sequent ├L involving the types of A, the constraints of L, and a subset NL U, the normal tokens of L, which satisfy ├L. 6. Theory T = <typ(T), ├ > consist of a set typ(T) of types, and a binary relation ├ between subsets of typ(T). Pairs <Γ, Δ> of subsets of typ(T) are called sequents. If Γ├ Δ, for Γ, Δ type(T), then the sequent Γ├ Δ is a constraint. T is regular if for all α∈ typ(T) and all sets Γ, Γ’, Δ, Δ’, ∑’, ∑0 , ∑1 of type: 1. Identity: α├ α, 2. Weakening: if Γ├ Δ, then Γ, Γ’├ Δ, Δ’, 3. Global Cut: if Γ, ∑0 ├ Λ, ∑1 for each partition <∑0 , ∑1> ( ∑0 ∪ ∑1 = typ(T) and ∑0 ∩ ∑1 = Φ), then Γ├ Λ. 7. 10 These rules consider mappings of types. f-Intro preserves validity but not non-validity. f-Elim does not preserve validity but preserves non-validity. 8. Representation system R = <C, ζ> consists of a binary channel C = {f: A C, g: B C}, with one of the classifications designated as source (say A) and the other as target, together with a local logic ζ on the core C of this channel. The representations of R are the tokens A. if a∈ tok(A) and b∈ tok(B), a is a representation of b, written a R b, if a and b are connected by some c ∈ C. The token a is an accurate representation of b if a and b are connected by some normal token, that is, some c ∈ Nζ. A set of types Γ of the source classification indicates a type β of the target classification, written Γ R β, if the translations of the types into the core gives us a constraint of the logic ζ, that is, if the translations of the types into the core gives us a constraint of the logic ζ, that is, if f[Γ]╞ζ g(β). The content of a token a is the set of all types indicated by its type set. The representation a represents b as being of type β if a represents b and β if a represents b and β is in the content of a. 9. State Space is a classification S for which each token is of exactly one type. The types of a state space are called states, and we say that a is in state δ if a╞ S δ. The state space S is complete if every state is the state of some token. 10.Event Classification Evt(S) associated with a state space S has as tokens the tokens of S. its types are arbitrary sets of states of S. The classification relation is given by a╞Evt(S) α, if and only it stateS (a) ∈ α. 11. Sum A+B of classification has as set of tokens the Cartesian product of tok(A) and tok(B) and as set of types the disjoint union of type(A) and typ(B), such that for α ∈ typ(A) and β ∈ typ(B), <a, b>╞A+B α iff a╞A α, and <a, b>╞A+B β iff b╞B β. 12. A (state space) Projection f: S1 S2 from state space S1 to state space S2 is given by a covariant pair of functions such that for each token a ∈ tok(S1), f(stateS1 (a)) = statesS2 (f(a)). 13. Proposition 8.17. Given state space S1 and S2, the following are equivalent: (1). f: S1 S2 is a projection; (2). Evt(f) : Evt(S2) infomorphism. Evt(S1) is an Wang.Y is a Researcher and Dr. J. K. Feng a Senior Lecturer at the University of Paisley 11