This content was uploaded by our users and we assume good faith they have the permission to share this book. If you own the copyright to this book and it is wrongfully on our website, we offer a simple DMCA procedure to remove your content from our site. Start by pressing the button below!
, where P is a term for a set of persons. ii) the given sets: d(P), the set of ALPHABETS, and the set of natural numbers N. iii) the attributes: d(name): d(P) o ALPHABETS, d(income): d(P) o N. We can denote a generic attribute of P as {f}P, meaning that persons belonging to d(P) have an attribute f. A graph of Conceptual Classes with two direct descendant of the root has been generalized and represented as in Figure 3,
164
E. Locuratolo and J. Palomaki / Extensional and Intensional Aspects of Conceptual Design
where the directed arrow is a “is-a” relation between classes. Now, the attribute inheritances are described as follows: {n,i}P, {n~s,i~s,m}S, {n~u,i~u,m~u,a}U and E P, and either S E = or S E z . In the case S {n~e,i~e,s}E. Also, U S P, E z , we get {n~s,i~s,s~es,m}S and {n~e,i~e,m~se,s}E. And, if U E z , we get {n~u,i~u,m~u,a,s~u}U. The next section describes an intensional containment relation between concepts. 4. An Intensional Containment Relation In conceptual level we have a relation between concepts, and as such we shall take as a primitive the intensional containment relation between concepts, [3,8]. We can say that a concept a contains intensionally a concept b, and denote it as a b. If the concept a contains intensionally the concept b, then the extension of the concept a is a subset of the concept b. For example, the concept of ‘dog’ contains intensionally the concept of ‘quadruped’, and the set of dogs is a subset of the set of quadrupeds. There are several nonidentical concepts, which are co-extensional, and so we can infer from concepts to its extension, but not vice versa. This is an example of the law of reciprocity, i.e. the more intension, the less extension, and vice versa. By means of the intensional containment relation it is possible to define some operations on concepts, for a more formal presentation see [3,8]: Two concepts are compatible, a A b, if there is a concept x, to which them both are intensionally contained. Def. 2. Two concepts are incompatible, a T b, if they are not compatible. Def. 3. Two concepts are comparable, a H b, if there is a concept x, which is intensionally contained to them both. Def. 4. Two concepts are incomparable, a I b, if they are not comparable.
Def.1.
An intensional negation is defined by means of incompatibility as follows: Def. 5.
An intensional negation of a concept a is a concept b, which is intensionally contained to the every concept x, which is incompatible with it.
That is, the intensional negation of a concept is the greatest lower bound of all those concepts which are incompatible with it. An intensional negation of a concept a is denoted below by a. When two concepts a and b are compatible, the least upper bound exists, and it is denoted by a b. On the other hand, when two concepts a and b are comparable, their greatest lower bound exists, and it is denoted by a
b. Without confusion, if we denote the extension of a concept a by set(a) as well, we can see from the law of reciprocity between intension and extension of a concept that the extension of a concept a b is the intersection of extensions of the concepts a and b, i.e. set(a) set(b), and the extension of concept a
b is the union of extensions of the concepts a and b, i.e. set(a) set(b).
E. Locuratolo and J. Palomaki / Extensional and Intensional Aspects of Conceptual Design
165
In general, intensional negation is very problematic, but for ASSO where the most general super-class determines the universe of discourse, we may modify the Def. 5. so, that the quantifier is restricted to that specific universe of discourse. That is, for example, restricting the universe of discourse by the extension of a concept p, i.e. the set(p) is the most general super-class in question. So, the modified definition of a restricted intensional negation is the following, Def. 5*. A restricted intensional negation of a concept a is a concept b, which is intensionally contained to the every concept x, which is incompatible with it, and moreover, there is such a concept p, which is intensionally contained to the concepts a and b. A restricted intensional negation of a concept a is denoted below by ra.
5. A Partition Method The “is-a” hierarchy defines conceptual classes, if the following properties hold: 1. Classification: each node of the hierarchy is a class. 2. Attribute inheritance: the sub-class inherits all the attributes from the superclass, and may have the same additional attributes. 3. Object inclusion: the sub-class object is a subset of the super-class objects. Now a partition method is the step sequence of partitioning decompositions applied to the conceptual classes represented in Figure 3. The aim of the decomposition is to get all possible conceptual classes implicitly included to the “is-a” hierarchy. Thus, e.g. the subclass <E,{s}> of the super-class
is <E,{n~e,i~e,s}>. A partition of a non-empty set A is a collection of non-empty subsets of A such that: i) For all S and T , either S = T or S T = , and ii) A = S S. Accordingly, a partition of a set A is a collection of non-empty and pairwise disjoint subsets of A, which exhaust the set A. An element of a partition is called a block. Because we are working with concepts, which determine the sets, we are not interested on actual members of those sets. For example, we may understand a partition of a set A as a collection of boxes, where boxes are the subsets of set A. Now even the empty boxes can have labels. Moreover, every element can be only in the one box. So, the requirement of non-emptyness in the definition of in this context is possible to drop out. Drawing the three intersecting sets inside the set P as shown in Figure 4, body:0/a:0 body:0/div:1 body:0/table:2/tr:0/td:0/div:0/div:0 body:0/table:2/tr:0/td:0/div:0/div:1 body:0/table:2/tr:0/td:0/div:0/div:2 body:0/table:2/tr:0/td:0/div:0/div:3 body:0/table:2/tr:0/td:0/div:1 body:0/div:3 body:0/img:4 body:0/img:5
166
E. Locuratolo and J. Palomaki / Extensional and Intensional Aspects of Conceptual Design
we get the following six different blocks, (note, by “set(A) \ set(B)” we mean a set theoretical difference, i.e. the intersection of the set(A) with the complement of the set(B)): 1. 2. 3. 4. 5. 6.
set(P) \ (set(S) set(E)), (set(P) set(S)) \ (set(E) set(U)), (set(P) set(S) set(U)) \ set(E), (set(P) set(E)) \ set(S), (set(P) set(E) set(S)) \ set(U), set(P) set(E) set(S) set(U).
These six blocks 1.–6. are in an extensional level. In an intensional level we get the following correspondent formulas for concepts: 1’. 2’. 3’. 4’. 5’. 6’.
concept(P) r(concept(S)
concept(E)), (concept(P) concept(S)) r(concept(E)
concept(U)), (concept(P) concept(S) concept(U)) rconcept(E) , (concept(P) concept(E)) rconcept(S), (concept(P) concept(E) concept(S)) rconcept(U), concept(P) concept(E) concept(S) concept(U).
In this example the restricted intensional negation, “ r”, is restricted to a concept(P). Accordingly, from a modelling point of view we have the possibility to model the boxes containing the objects both in the intensional as well as in the extensional way. Intensionally the modelling is done by concepts, whereas extensionally the modelling is done by Partitioning. However, since the relation between a concept and its extension is many-one, a given box can be an extension for many non-identical concepts, but not vice versa. 6. Some Further Developments of Partitioning The Partitioning maps the conceptual classes into the object classes. In [5], the graph nodes representing the conceptual classes have been labelled by numbers, whereas the class attributes have been labelled by double indexed functions and the inherited attributes by double indexed primed functions. As the general transformation from conceptual classes to object classes is a complex task, the solution has first been determined for the elementary
E. Locuratolo and J. Palomaki / Extensional and Intensional Aspects of Conceptual Design
167
single-path tree, that is to say a tree formed by nodes with only one direct descendent, and then it has been determined for more general basic case. For each of them, the relationships between the determined object classes and those of the previous cases have been established. This addressed the process generalization first to a generic tree then and to a generic graph. Each graph node is characterized by both its set of objects and its set of attributes. The set of objects is represented by a subset of a given set, whereas each attribute is represented as a function defined in the subset of the specified class and assuming values in a given set. Operations on graphs permit to perform recursive decompositions of conceptual classes until object classes are obtained. The method is correct, that is the objects of the obtained classes are all and only the objects of the conceptual classes and these have all and only the attributes defined in the conceptual classes. The Partitioning is a difficult method resulting into disjoint classes which are not recomposed into a class hierarchy satisfying to the property that each object instance belongs to one and only one class. The Revised Partitioning is composed by two phases, called representation and decomposition, respectively. The former permits describing the conceptual classes, whereas the latter permits decomposing them. As to the representation phase, a label connoting the class name and denoting the class objects has been associated with each graph node, whereas a list of attribute names has been associated with each label (See Figure 1.a)). This model implicitly specifies that the objects of each class are represented by a subset of a given set, whereas each attribute has been formalized as a function defined in the specified set of objects and assuming values in an implicitly specified set. As to the decomposition phase, this is a stepwise approach satisfying to the following properties: y Root partitioning: the root objects of the conceptual classes are partitioned into the root objects of the conceptual classes resulting from each decomposition step. The root labels represent the partitioning. y Root labeling. The root labels of the conceptual classes resulting from each decomposition are defined by combining the root label before the decomposition with the labels of the root direct descendants. y Root structuring: the root labels can be decomposed into two parts separated by the “-“sign: the former on the left of this sign consists of label intersections, whereas the latter on the right consists of label unions. One of the two parts can be empty. y Consistency: the following implicit information is specified through the root labels: only the attributes of class X are associated with a node labeled by X Y, whereas the attributes of all the classes X…Y are associated with a node labeled X…Y. From root partitioning, it follows that each object of the original conceptual classes belongs step by step to one and only one of the obtained conceptual classes. From root labeling, it follows that all the information required for schema decomposition is enclosed in the root labels and its direct descendants. From root structuring and consistency, it follows that the attributes can be associated with the classes exploiting information implicitly specified in the root labels. Furthermore, each object has all and only the original attributes. The root structuring is also exploited in order to construct the class hierarchies satisfying to the property that each object instance belongs to one and only one class.
168
E. Locuratolo and J. Palomaki / Extensional and Intensional Aspects of Conceptual Design
In Figure 5, the first decomposition step of the conceptual classes shown in Figure. 1.a has been represented: the root objects of the original conceptual classes have been partitioned into the root objects of the conceptual classes after the decomposition; the following implicit information is specified through the root labels: only the attributes of the class person are associated with the class Person-Employee, whereas both the attributes of the class person and of the class employee are associated with the class PersonxEmployee.
With a further step of decomposition, all disjoint classes are obtained; the method recomposes the disjoint classes to define the object classes as in Figure 6. In the object classes, all the classes implicitly specified in the original conceptual classes are determined and thus the object classes represented in Figure 3 enclose one class more than the original conceptual classes. The concepts of multiple inheritance class, i.e. a class linked with higher level classes through two or more is-a relationships, can be employed also for the conceptual classes; this holds only when the classes which can be declared implicitly within other classes have other specific attributes.
E. Locuratolo and J. Palomaki / Extensional and Intensional Aspects of Conceptual Design
169
7. Conclusion A partition method is designed to achieve the flexibility of semantic models in reflecting changes occurring in real life and the efficiency on object systems. The method of partition has been applied only to the extensional aspects of concepts, i.e. set. In this paper a partition method is applied to the intensional aspects of concepts as well. A particular problem concerning the intensional negation of a concept is solved by defining a restricted intensional negation of a concept, which is important in all practical conceptual design situations all of which are confined in specific restricted part of the universe of discourse. Some further development of partitioning is presented as well. Acknowledgments This paper derives from the research undertaken by the authors during the CNR short term mobility program ( Pise – ’99). References: [1] Cardenas, A. F., and McLeod, D., 1990: Research Foundations in Object-Oriented and SemanticDatabase Systems. Prentice Hall, Englewood Cliffs, NJ 07632. [2] Elmasri, R., and Navathe, S.B., 2000: Fundamentals of Database Systems, Addison-Wesley. [3] Kauppi, R., 1967: Einführung in die Theorie der Begriffssysteme. Acta Universitatis Tamperensis. Ser. A. Vol. 15 Tampere: Tampereen yliopisto. [4] Locuratolo, E., 1998: “ASSO: Portability as a Methodological Goal”. Technical Report IEI B4-05-02, 1998. [5] Locuratolo, E., and Rabitti, F., 1998: “Conceptual Classes and System Classes in Object Databases”. Acta Informatica 35(3), 181-210. [6] Locuratolo, E., 2005: “Model Transformations in Designing the ASSO Methodology”. In Idea Group Inc Transformation of Knowledge, Information and Data: Theory and Applications, 283-302. [7] Nixon, B., and Mylopoulos, J., 1990: “Integration Issues in Implementing Semantic Data Models”. Advances in Database Programming Languages. ACM Press, New York and Reading: AddisonWesley, 187-217. [8] Palomäki, J., 1994: From Concepts to Concept Theory. Acta Universitatis Tamperensis. Ser. A. Vol.416. Tampere: Tampereen yliopisto.
170
Information Modelling and Knowledge Bases XIX H. Jaakkola et al. (Eds.) IOS Press, 2008 © 2008 The authors and IOS Press. All rights reserved.
Emergence of language: hidden states and local environments Jaak HENNO Tallinn Technical University, Tallinn 19086, Estonia
Abstract. Here is considered emergence of language in an environment, where agents communicate about issues which are not observable at the moment of communication, e.g. they search food and exits in a labyrinth and communication helps them to achieve their goals (food, exits). In such a situation message's receiver can not observe message's object at the moment when message is received. It is shown, that these non-observable objects can be mapped to observable objects using local environments. Agents do not build representations of the external word, instead of reasoning on representations of the world they access the world directly through perceptions and actions and their perceptions influence their behaviour. Messages from other agents, i.e. emerging language also change their behaviour and increases effectiveness of the whole population. Language is described via two mappings: syntax (i.e. syntactic objects, words) is interpreted in semantics by the meaning mapping; the speech mapping creates for semantic objects their syntactic denotations. Words in agents language gradually become mediated semantic objects, i.e. obtain the same significance (trigger the same actions) as the actual real-word situations which they denote.
1. Introduction In the last 10 years have appeared many papers where language emergence is investigated using computer simulations [1],[2],[3],[4],[5] etc. The topic is still not well understood and "classical"-style language researchers sometimes do not quite believe in the results obtained: "Reactions vary from fascination and incomprehension to scepticism or downright rejection" [6]. However, most of language researches believe in the method: "… emergent area of consensus is the growing interest in using computational modelling to explore issues relevant for understanding the origin and evolution of language"[7]. Especially popular in language emergence studies have become Sony's robot-dog Aibo [8],[9],[10] etc. In these studies robots research their environment and send and receive messages about objects, which they find, i.e. about objects, which both sender and receiver simultaneously can perceive (so called "word games"[11]). It is assumed, that both communication parties (sender and receiver) unambiguously can identify the topic, object of the message (e.g. using pointing [11]). But pointing is ambiguous (this problem has been discussed already e.g. in [12]; more realistic, probabilistic methods for topic determination in multi-topic environment was considered e.g. in [13]). Assumption of simultaneous perception (grounding) of message's object in the moment of communication is also rather unrealistic in real language acquisition. Children generally do receive little, if any, feedback while learning words [14]; most of the words what we know and use are not taught to us pointing to the corresponding object or event, we have learned them from context - picked them up from conversations,
J. Henno / Emergence of Language: Hidden States and Local Environments
171
from texts etc [15], [16]. When human language emerged, tribes were usually living in caves and the most important topics of communication (e.g. where to find food) where not simultaneously observable for communicating parties. How can agents create a common language, when they search exits and food in a labyrinth, i.e. the typical messages are "found food", "the passage was a dead end"; how can they develop a common language using messages about objects which are not directly observable at the moment of communication? We learn meanings of words from context. For every object and event there is a context, where it appears or occurs, and this context makes it possible to disambiguate the meaning of words. Context can be used also to map non-observable (at the moment when message is received) objects and events to real, observable (later) ones. Another feature of "real-word" language acquisition which is usually not considered in papers considering language emergence is pragmatics - the practical value of learned words. Learning colours as combinations of RGB values (without any practical significance) can give new insights to the design of learning (pattern-matching) algorithms and principles of learned perceptually grounded categories([17]). But it may not be very relevant to understanding the emergence of common vocabulary, since features (e.g. speed) of this process depend essentially on practical value of topics which these words denote. For instance, suppose that a community of agents is searching a labyrinth and the (only) perception what agents are able to receive is whether it is possible to continue along a passage or not (the passage turned out to be a dead end), i.e. the only topic of messages would be "dead end". If agents do not have any goal (they are just wandering) then clearly such a message (if they already understand it) does not change receiver's behaviour. But if they want to find an exit from the labyrinth, then such a message would make the receiver to turn around and search for another route. Thus the whole community will be more active (search more rooms); this increased activity will also increase the probability of encounters, i.e. exchanged messages and greater number of messages will increase the speed of emergence of common vocabulary. Here is an illustration to this. On the next time step agents (1,1) and (1,3) will meet in the room (1,2) and agent (3,1) will send to agent (1,1) message "dead end". If this message does not have any significance for agent (1,1), he would continue from the room (1,2) to rooms (1,3), (2,3) and also find himself in a dead end. But if his goal is to find an exit, he will change direction and continue into room (2,2) and meet there agent from the room (3,1), i.e. he got more possibilities to communicate and learn. All messages with practical value have similar effect - increase the speed of emergence of language. If the value is positive (e.g. food), agents tend to move towards their sources; the concentration of agents around (food) sources increases and they have more communication opportunities; if the value is negative (danger), agents try to avoid their sources and move in the opposite direction(s), the concentration of agents increases in nondangerous areas and probability of communication increases there. In the following is shown, how agents can learn and create a common system of denotations, language (actually only vocabulary) in a situation, where 1) objects and events, which are described in their messages can not be perceived at the moment when message is received; 2) agents have goals; messages are significant to agents and make them (when already understood) to change their behaviour (direction of movement); there are messages with
172
J. Henno / Emergence of Language: Hidden States and Local Environments
positive value (make agents to move forward) and negative value (make agents to move back); 3) agents start without any common language; language is created (learned) when agents exchange messages; when guessing agents do not receive any feedback about correctness of their guess. In some papers where word games are used as a tool to investigate emergence of language it is claimed that feedback is essential for language emergence (e.g. in [17]); however, experiments with word games have shown that feedback is not needed for language emergence (e.g. [19],[20],[21]). The set-up was investigated with a computational model. Used in experiments agents had very natural and simple properties: - agents move in 2D labyrinth; they have goals, which they want to achieve – find exit, find food, beware dangers; - agents perceive properties of their environment, which are essential to achieve their goals (semantic objects and events); they can not change their environment (e.g. mark rooms where they have been); - they can send and receive messages about objects and events; when creating a message, they use their speak function, which maps their experiences (inner state) to lexical symbols, words (agents can create sufficiently many different words); meaning function maps received message (words) into semantic objects and events; - agents are event-driven finite-state systems with inputs and outputs; agents do not build "world model", their outputs are functions of their environment and inner states(they implement "intelligence without representations"[18],[10]); - at the beginning agents do not have any understanding of each other's messages, i.e. do not have any common vocabulary, but in the process they create common vocabulary, where words gradually obtain the same meaning as external objects and situations which they denote and trigger the same changes in their behaviour (behavioural equivalence of denotations with external objects which they denote). 2.Objects and contexts The main idea in agent's reasoning (i.e. why it is possible that agent's can create a common system of denotations, vocabulary and later a language) is disambiguation of meanings when objects are presented in different contexts [19]. When an agent first sees two objects and another agent says two words: "õun, pirn" then certainly there is no way to understand, what is what, so in his vocabulary he has to use both words as possible denotations for both objects. But if he then sees one of these objects in different context and again hears one of the words which was also used earlier then he can place all words for corresponding objects: "õun = apple", "maja = house", "pirn = pear". The previous example (and most publications on language emergence) assumes that the message's topic (object) is unambiguously determined by both sender and receiver. But what if the object or event described in a message is not present (can not be pointed at), if speaker is describing something, what he has encountered earlier and what is now stored in his memory, i.e. his inner, directly non-observable (hidden) state?
J. Henno / Emergence of Language: Hidden States and Local Environments
173
To get a cue there should be some (recognizable by both, speaker and hearer) local context in their environment, where the message's object could be found. Agents act in space and time, so this local context what both agents can perceive should be a finite subspace, which is determined by some cues, which all agents can perceive and understand. This perceivable context allows mapping received words to external objects and events and thus disambiguating meanings. 3.Computational model For physical environment in computational modelling was used the well-known game "Hunt the Wumpus". In this game agents have to find an exit from a rectangular 2D-labyrinth. There are several dangers: a mysterious beast Wumpus, who scares agents and pits, where agents can fall. The dangers were somewhat soften: agents are not destroyed, but always could escape and later warn others about the danger; dangers only affect their movements, they try to avoid dangers. Agents have to feed themselves – in some rooms of the labyrinth they can find food (food sources renew themselves and will not be exhausted). Thus the semantic environment has four sorts of objects: - agents; agents are event-driven finite state machines, who live (i.e. move) in discrete time (all in parallel, i.e. in on one clock cycle all agents move to next room); - rooms; the attributes of a room is the list of its connections with other rooms (where one can go from here); - dangers: Wumpus and pits ; - food. Agent's state consists of components, which store different types of information: - current room attributes (in which directions it is open, i.e. where he can continue) and agent's current direction, i.e. in which direction he was moving when he arrived in this room; thus he can always say, whether a room is a dead end (the only opening is in the opposite direction with his movement); - hunger – how many moves have passed when he last time got food; - danger – there are dangers (Wumpus, pit) nearby. Generally agents always try to move forward into next room (default movement - they have to find an exit); they turn back (180o) only when they can not continue (dead end). Agents can exchange messages. When some agent learned something, he would try to inform others with following messages: "there was a dead end" (when the sender is returning from this dead end); "there is food in this passage"; "there is a danger ahead"; "I'm hungry!" - this message indicates, that there was no food in the last ht rooms (see below). These messages can not be analyzed (understood) at the time when they are received – usually the message's object is not perceivable when message is delivered. For instance, if an agent found a dead end and meets on his way back another agent , then he would tell him that
174
J. Henno / Emergence of Language: Hidden States and Local Environments
this is a dead end, but the dead end itself is not perceivable, there is (at the moment) no external information about the actual object (dead end). To recognize that message (word) produced by another agent means something about external environment (dead end, food, danger) and this object is somewhere nearby, there should be some finite area (messages context environment) which both the sender and receiver can recognize and where this particular object is situated. Message's sender describes in his message only these experiences which occurred in this local context and if messages receiver searches message's context he can find the message's object. For messages context could be used agents short-term memory of fixed length, e.g. all agents remember what they have seen in last five rooms. But this is artificial and nonefficient use of memory: if agent moves along an empty corridor (with only one possibility to continue), why should he remember all these empty rooms? More natural is to consider changes, i.e. where something happened or rooms where it was possible to select between more than one rooms to continue. Thus for first three types of messages the context environment is the corridor where sender and receiver currently are; for the last ("hungry!") message the context area is finite time interval - ht moves (the same for all agents). 4.Agent's data structure and algorithm 4.1 Inputs Agent's can perceive their environment. From a room where they currently are they can observe the following (i.e. meanings, the elements of the semantic space): r - how many other rooms are connected to the current room; they also know, which room the (just) come from and this (last visited room) is not counted, thus they can recognize a dead end (r = 0) and end of a dead-end corridor (r>1), these two inputs are denoted as de and de ; f - whether there is food in the room; dan - whether there is danger in the room (Wumpus or pit); ag - whether there is another agent (or several) in the current room; t – time (tick), entering the next room increases time counter t by one; time is used only to correct agents state of hunger: agent's time (state of hunger) is set to 0 if there is food in the room, on entering next room the time counter is increased until t ht ; after that agent becomes hungry, i.e. its state component H=1 (see below) and agent does not increase its time counter any more; m – message from another agent. 4.2 Agent's State Agent's state consists of six components. The first five correspond to one sort of meanings and are activated/deactivated by entering/leaving the corresponding external context environment; the last holds un-analyzed messages, received from other agents in the current context. The components and their state rules are: - DE - (Dead End) is activated ( DE 1 ), when agent finds itself in an dead end and has to move back; in this state agent can turn left or right ( 90D or 270D ), but not back ( 180D , to direction of the dead end); as soon as agent finds that there are more than one
J. Henno / Emergence of Language: Hidden States and Local Environments
175
possibilities to continue (the room where it come from, i.e. the dead end direction is not counted), DE 0 (this is also the initial value), e.g. on the picture (states are described in the order of movement) DE(1,4) = DE(1,3) = DE(1,2) = DE(1,1) = 0; on the next step agent encounters a dead end: DE(2,1) = DE(1,1) = DE(1,2) = 1; DE(1,3) = 0 (an exit there is more than one possibility to continue); - F (Food) – F 1 , when agent finds food in a room, but food is forgotten ( F 0 , this is also the initial value) as soon as agent exits the passage where food was found, i.e. in a room where there are more than 1 possibilities to continue (the same way as DE), e.g. on the picture F(1,3) = 0, F(1,2) = F(1,1) = F(1,2) = 1 (exit from the current context), F(1,3) = 0; - Dan – (Danger) – there is danger (Wumpus or pit) nearby – the behaviour of this state component is similar to DE, F; - T – time (steps) stores the number of steps, but only until the value ht , when T ht , the component H (Hunger) is activated (H=1) and T remains unchanged; T,H are reset (T , H 0 ) when food is found; - M – messages from another agents (un-altered); the messages will be kept until agent leaves the current context (passage with only one possibility to move forward, i.e. r<2); in the room where agent leaves the current context (a room with r>1) agent analyzes received messages using cues from environment what he stored (i.e. dead end, food, danger) when he was moving in the current context. For memory components DE,F,Dan,H there is a buffer for words which (possibly) can denote the corresponding meaning. For every word is stored also its use count – how many times this word (probably) was used to denote this entity. All these buffers have finite length. Since all agents can "invent" new words if they do not yet have any word for a meaning, it is possible that some agent receives # Ag 1 different words from some meaning ( # Ag is the number of agents), thus in maximally unrestricted simulation length of buffers should be # Ag 1 . However, every exchange of messages decreases probability for need to use a new word. In experiments speed of convergence with buffers of length # Ag / 4 was nearly the same as with length # Ag 1 (but with long buffers the model become very slow), thus the buffer length was usually set to 5. When a buffer gets full, the word with smallest use count is dropped. 4.3 State changes Agent's state changes are triggered by their states and external inputs according to following functions: de l DE (input de turns memory state DE to true); f l F (input f turns F to true); dan l Dan ; t l T max(T 1, ht ) ;ag l speak (DE , F , Dan ) - when agent meets another agent, he speaks about what he knows (has stored in corresponding states DE,F,Dan,H) about dead ends, food, hunger and dangers; ag speak (ag ) l (M ) - if agent meets another agent who speaks, the received message is stored in memory M (un-altered, since there is yet possible make any decisions about its meaning);
DE de l (DE 0) - on leaving a dead-end corridor the value of DE becomes false;
176
J. Henno / Emergence of Language: Hidden States and Local Environments
F de l (F 0) - when leaving a corridor with food the value of F becomes false, agent "forgets" about the food; Dan de l (Dan 0) - when leaving a corridor with a danger agent forgets about danger; M de l M 0 - agent "forgets" everything what he encountered in the last context (but before that the content of M is used to update the function meaning).
4.4 Speak and meaning mappings For the function speak agent selects from the corresponding buffer a word with maximum use count; if the buffer is still empty, agent "invents" a new word, different from all other words which are in his buffers (in [20] were considered also some other tactics of word selection, e.g. agents can sometimes “stick” to their “own” word - what they invented earlier - etc.). The input de (exit from the current context) fires update process for the meaning function. Let the content of the message's short-term memory be M {w1,..., wn }, n 0 and let {B1,..., Bm }, B j {de, dan, f , h } be the elements of semantic domain, which correspond to agent's activated state components, i.e. his experiences in the current contexts; agent has to create a mapping from words to elements of the semantic domain (meanings). The update for the mapping meaning consists of following steps: 1. Receiver separates from the set {w1,..., wn } of received words the known words, i.e. words w k , which are already present in the list of denotations of some B j , j=1,…,m . Receiver increases use counts of these words in corresponding lists and removes these elements B j from the set {B1 ,..., Bm } (words for them are already found). Let {v1 ,..., vk } and {C1 ,..., Cl } , k n , l d m be the remaining words and objects. Since there is no additional information available (what means what), every word vi is mapped to every C j , i.e. m2 ({v1 ,..., vk } o {C1 ,..., Cl } {v1 ,..., vk }
{C1 ,..., Cl } (direct product) - all words vi , i=1,…,k are added to word lists of objects C j , j=1,…,l. 2. The "known" words (i.e. words, which were removed in step 1.) are also removed from word lists of all meanings, which do not occur in the message, i.e. B {B1 ,..., Bn } - it is assumed, that all words in the message are about objects B1 ,..., Bn . In [20],[21] it was shown, that with explicit objects (pointing - receiver gets together with message also an indication of the message's object) the algorithm converges (i.e. agents will create a common vocabulary); the modified version which is used here converges also. The version used here can be interpreted as a version of explicit pointing: all messages are received only on exit from the current context and their objects are the meanings which receiver got from the current context. 4.5 Movements Agent's movements are one of their outputs (another is messages). At the beginning, when they do not yet have any information from others they always try to move forward and turn back only if they encounter a dead end or danger. So there are four forces fs , fde , fh , fdan , which affect their selection of the next move: search and hunger force forward, danger and
J. Henno / Emergence of Language: Hidden States and Local Environments
177
dead end – backwards (but hunger can also move backward, if agent decides that he received message where food was not mentioned and the message is reliable). When an agent already has some words in his word buffers and receives a message from another agent, he decides the next move using the information contained in the message. He decodes the message, i.e. evaluates meaning and reliability of every word w i of the message: m(wi ) is this element from {de, f , dan } , which has the word w i in the corresponding words buffer with highest use count; its reliability r (m(w i )) is 1/ N , where N is the number of states {de, f , dan } which have w i in their word buffer. If the current room is not a dead end (there is at least one possibility to continue), agent decides whether to continue or to move back using the following rules: - if H 1 (agent is hungry) and r ( f ) 0 (i.e. he has a message and for some word from the message m(wi ) f ), then agent continues; - in other cases probability to continue is 1 (r (de ) r (dan )) . 5.Results In the experiments m agents were put randomly into a computer-generated n q n labyrinth with random dangers and food sources. At the beginning all agents were "tabula rasa", i.e. did not have any vocabulary. When they all had found exit, they (in their current vocabulary-learning state) were put into another computer-generated labyrinth and so on, until they managed to create some level of mutual understanding, i.e. common language. The length of the experiment was measured as the number of communications they had. Since they all moved in parallel, the number of encounters in a situation, where somebody had to say something to another (had found food or a dead end), was rather small (especially when the society was small). On the picture 5 agents who had developed some understanding (rate of understood one-word messages > 50%) searched exit in a 10 q 10 labyrinth. A situation where one agent warned another about a dead end can be recognized as agent turning back without any obvious reason and these are marked by small circles (in all in this case there were 15 such communications). In the experiments were measured several parameters: speed of development of common words (all agents used the same word for some meaning), understanding of messages (understanding was always better than the rate of common words), influence of the size of society and the size of labyrinth etc. Very essential was the size of society (number of agents). If there were few agents, they met (i.e. had an opportunity to communicate) seldom if ever, so they could not develop common vocabulary. And even if they did (they were successively placed in many labyrinths, so they had to meet and communicate), they nevertheless could not use each other effectively, could not pass the learned information to others. On the next picture are results of series of experiments, where societies of agents of different sizes were searching a series of 30 q 30 labyrinths. Measured was the average number of rooms which agents passed before they found exit, depending on the level of their common vocabulary (how many messages from others they already understood). Small number of agents (societies of 5 and 20 agents) could develop common vocabulary (after many trials), but could not use it effectively, the vocabulary nearly did not have any effect on their efficiency. Societies of
178
J. Henno / Emergence of Language: Hidden States and Local Environments
200 and 300 agents developed common vocabulary nearly in the same time, but for them the common vocabulary made their search more than two times more efficient.
% of visited rooms
140.00 120.00 100.00
5 agents
80.00
20 agents
60.00
200 agents
40.00
300 agents
20.00 0.00 5 15 25 35 45 55 65 75 85 95 % of understood messages
Figure 1. How size of society increased society's effectiveness
The goal - why agents developed common vocabulary - was to decrease the labyrinth search time, therefore the best parameter for estimating the use of communication skills is the total number of moves agents made before they all found the exit. At the beginning, when agents could not understand each other they wandered more or less randomly and the percentage of rooms searched (from all the possible rooms) was always around 110-130%. When they had learned to understand each other and there was sufficiently many of them (number of agents > number of rooms/10), the percentage of passed rooms decreased rapidly. On the following pictures is the total number of passes what 50 agents made in 30 q 30 labyrinth presented as line thickness; the first presents the situation in the beginning of series (agents did not understand each other and visited nearly all rooms), on the second – at the end, when they understood already > 95% of each other messages – most of them moved along the path which took them to the exit.
Figure 2. How common vocabulary increases society's effectiveness
6.Conclusions and ideas for continuation Here was investigated, how a society of agents could develop common vocabulary in a situation, where agents have goals and try to communicate with each other about external situations and topics which are essential for them (help to achieve their goals), but these objects and situations may not be observable to message's receiver at the moment when they meet and communicate. Thus in a message sender has to describe his previous experiences, his inner state, which is not observable to receiver. It was shown, that agents
J. Henno / Emergence of Language: Hidden States and Local Environments
179
can nevertheless develop common language (here actually only vocabulary) when they use local environment (message's context), which they can recognize. Message's sender describes in his message only these experiences which occurred in this local context and receiver searches this local context in order to understand, what the message was about. The emerging common language has practical value and allows the society to act (to achieve their goals) more efficiently. The presented above agent's structure is very elementary and could be improved in several ways. For example, currently they communicate only if they both are in the same local environment (the first image). But it would be natural, that agents could communicate also when sender has just left the local environment (the second image). For that the receiver should get the direction of the last move of the sender – either sender should be able to communicate this (but for this the vocabulary should be enlarged, they should learn new words) or the receiver should be able to perceive this from sender's state. Another issue which should be investigated further is how they use what they have learned. In the described above simulation agents use the emerging language only to improve their output, to find exit. They do not change their inner structure, the "thinking" algorithm. But learning could also improve their decision-making. In the above model the "natural" context for the agent's state component H is not the current passage, but the time interval of length ht and this could be also used to modify the agent's decision-making algorithm. The discussed algorithm, where agents search in parallel and learn to communicate to each other their findings could be applied in many practical situations. For instance, the presented algorithm works on every non-cyclic graph, e.g. a tree. On the picture are presented results of an experiment, where 3 agents who already understood each other's messages (80%) were searching a tree (target was the root); line thickness indicates, how many edges agents passed; their efficiency was ca 10 times greater than at the beginning of the experiment (when they did not understand each other). References [1] Christiansen, M.; Kirby, S. Eds. (2003) Language Evolution. Oxford University Press: New York. [2] L. Steels (2006). Experiments on the emergence of human communication. Trends in Cognitive Sciences 10(8), pp. 347-349 [3] Baronchelli, A., Felici, M., Loreto, V., Caglioti, E. and Steels, L. (2006) Sharp transition towards shared vocabularies in multi-agent systems. Journal of Statistical Mechanics P06014 [4] Galantucci, B. (2005) An experimental study of the emergence of human communication systems. Cogn. Sci. 29, pp 737–767 [5] Cangelosi, A., Riga, T., Giolito, B., and Marocco, D. (2004) Language emergence and grounding in sensorimotor agents and robots. In First International Workshop on the Emergence and Evolution of Linguistic Communication. May 31- Jun 1 2004, Kanazawa, Japan [6] L. Steels (2006) How to do Experiments in Artificial Language Evolution and Why. In Cangelosi, A., Smith A. and Smith K., editor, Proceedings of the 6th International Conference on The Evolution of Language (EVOLANG6), London [7] M.H. Christiansen, S. Kirby (2003). Language evolution: consensus and controversies. Trends in Cognitive Sciences, 2003, 7:7, pp 300-307
180
J. Henno / Emergence of Language: Hidden States and Local Environments
[8] L. Steels, F. Kaplan (2001) AIBO’s first words: The social learning of language and meaning. Evolution of Communication, 4(1) [9] J. Poudade, L. Landwerlin, P. Paroubek (2006). Cognitive Situated Agents Learn to Name Actions. ECAI 2006 pp 51-55 [10] Deb Roy. (2005). Grounding words in perception and action: Computational insights. Trends in Cognitive Sciences, 9(8) [11] L. Steels, P. Vogt (1997). Grounding adaptive language games in robotic agents. In C. Husbands and I. Harvey, editors, Proceedings of the Fourth European Conference on Artificial Life, Cambridge MA and London, 1997. The MIT Press [12] Quine, W. V (1960). Word and Object. MIT Press: Cambridge, MA [13] P. Vogt (1998). The Evolution of a Lexicon and Meaning in robotic agents through self-organization, In : C. Knight and J.R. Hurford (eds.). The evolution of Language (selected papers from the 2nd International Conference on the Evolution of Language, London, April 6-9). [14] Bloom, P. (2000) How children learn the meaning of words. Cambridge, MA: MIT Press. [15] Sternberg, R.J. (1987). Most vocabulary is learned from context. In M.G. McKeown & M.E. Curtis (Eds.) The nature of vocabulary acquisition. Hillsdale, NJ: Erlbaum. [16] Nagy, W.E. & Herman, P.A. (1987). Breadth and depth of vocabulary knowledge: implications for acquisition and instruction. In M.G. McKeown & M.E. Curtis (Eds.) The nature of vocabulary acquisition. Hillsdale, NJ: Erlbaum [17] L. Steels, T. Belpaeme (2005). Coordinating Perceptually Grounded Categories through Language: A Case Study for Colour. Behavioral and Brain Sciences, 28:4, pp 469-89 [18] R. A.Brooks (1991) Intelligence without representation. Artificial Intelligence 47, 139–159. [19] Andrew D. M. Smith (2002) Evolving Communication through the Inference of Meaning. University of Edinburgh, September 2003. pp 1-418 [20] J. Henno (2002). Emergence of communication and creation of common vocabulary in multi-agent environment. Proceedings of the 12th European-Japanese Conference on Information Modelling and Knowledge Bases. Krippen, Swiss Saxony, Germany. May 27-30, 2002, pp 229-233 [21] J. Henno (2006). Mathematical Model of Natural Languages. ICCC 2006 IEEE International Conference on Computational Cybernetics, Tallinn, Estonia August 20-22, 2006, Proceedings, pp 275-281
Information Modelling and Knowledge Bases XIX H. Jaakkola et al. (Eds.) IOS Press, 2008 © 2008 The authors and IOS Press. All rights reserved.
181
Frameworks for Intellectual Property Protection on Multimedia Database Systems Hideyasu Sasaki Attorney-at-Law, New York State Bar, the Third Judicial Department, Albany, N.Y., U.S.A. Ritsumeikan University, 6-4-10, Wakakusa, Kusatsu, Shiga 525-0045, Japan
[email protected] Yasushi Kiyoki Keio University, 5322 Endo Fujisawa, Kanagawa, 252-8520 Japan
[email protected] Abstract. In this paper, we discuss the issues and future trends on patentable parameter setting components implemented in multimedia database systems. Multimedia databases in the applications of parameter setting components consist of copyrightable metadata. The data-processing processes are patentable in the forms of parameter setting components. The current techniques in parameter setting components enclose a variety of numerical parametric information which inventors would like to cover as trade secret. We present the conditions of copyrightability on the multimedia databases and the patentability on the parameter setting components with the directions for protecting numerical parametric information as trade secret.
1 Introduction The principal concern of this paper is to present the conditions of copyrightability on the multimedia databases and the patentability on the parameter setting components with the directions for protecting numerical parametric information as trade secret. Our secondary concern is to provide researchers and practitioners in information modeling and knowledge bases with legal references on the concepts, issues, trends and frameworks of intellectual property protection regarding “multimedia database systems” in engineering manner. A multimedia database system, as an information system, consists of digital contents in databases and retrieval mechanisms. The intellectual property protection of multimedia database systems is a critical issue in the multimedia database community that demands frameworks for recouping their investment in database design and system implementation. Intellectual property law gives incentive to advance appropriate investment in database design and implementation with two conventional types of intellectual property protection: copyright and patent [1, 2]. Multimedia digital contents take a variety of forms including text, images, photos and video streams, which often commingle in multimedia databases. Nevertheless, present legal studies are not satisfactory as the source of technical interpretation of the intellectual properties regarding multimedia databases. The intellectual property protection of the multimedia databases demands clear and concise frameworks.
182
H. Sasaki and Y. Kiyoki / Frameworks for Intellectual Property Protection
2 Background In this section, we discuss two main issues on the intellectual property protection regarding multimedia database systems. The first issue is the copyright protection of databases to which the multimedia digital contents are stored in multimedia databases. The second issue is the patent protection of the retrieval mechanisms of multimedia database systems. 2.1 Copyright on Multimedia Databases U.S. Copyright Act [3] defines that a compilation or assembling of individual contents, i.e., preexisting materials or data, is a copyrightable entity as an original work of authorship. Gorman and Ginsburg [4], and Nimmer, et al. [5] state that a compilation is copyrightable as far as it is an “original work of authorship that is fixed in tangible form”. Multimedia database systems consist of multimedia digital contents which are indexed and stored in databases for appropriate retrieval operations and the retrieval mechanisms which are optimized and applied to object domains of those databases. The entire multimedia database is copyrightable in the form of a component of “contents-plus-indexes” while static indexes or metadata are fixed to multimedia digital contents in a tangible medium of repository, i.e., database. Static indexes or metadata represent a certain kind of categorization of the entire content of each database (See Fig. 1). The originality on the categorization makes each database copyrightable as is different from the mere collection of its individual contents. What kind of categorization should be original to constitute a copyrightable compilation on the database? The court of American Dental Ass’n v. Delta Dental Plan Ass’n [6] determined that minimal creativity in compilation sufficed this requirement of originality on databases. Any standard or framework on the requirement is not clear in the technical or engineering meanings. A uniform framework on the categorization regarding indexes or metadata of databases must be formulated in engineering manner. The European Union has legislated and executed a scheme for protecting a database including its content per se, known as the sui generis right of database protection [7, 8, 9]. That European scheme shares the same issue on the originality regarding the categorization of multimedia digital contents in databases. 2.2 Patent on Multimedia Database Systems U.S. Patent Act [10] defines that a data-processing process or method is patentable subject matter in the form of a computer-related invention, i.e., a computer program. The computer program is patentable as far as the “specific machine . . . . . . produce(s) a useful, concrete, and tangible result . . . for transforming . . . ” physical data (“physical transformation”) [11]. The computer-related inventions often combine means for data-processing, some of which are prior disclosed inventions. A retrieval mechanism in a multimedia database system consists of a number of “processes”, i.e., methods or means for data processing in the form of combination of computer programs. A set of programs focuses on image processing, while another set of programs operates text mining, for example. Meanwhile, the processes in a retrieval mechanism of a multimedia database system comprise means or components for parameter setting which is adjusted to retrieve specific kinds of multimedia digital contents, for example, images in certain domains. The problem is that which process is to realize technical advancement (nonobviousness) on its combination of the prior arts and is to be specific/enable on its parameter setting. These two issues are emerging
H. Sasaki and Y. Kiyoki / Frameworks for Intellectual Property Protection
183
KEYWORD-BASED RETRIEVAL DATABASE DESIGNER
DB CONTENTS
INDEXES/
+ METADATA DICE 4
DICE 6
CREATE
INTEGRATE (STATIC COMPILATION)
CONTENT CREATOR Figure 1: Formulation for copyrighting multimedia databases.
problems in the advent of multimedia database systems. Uniform frameworks on the novel combination and the specific parameter setting must be formulated in engineering manner, respectively. 3 Frameworks for Intellectual Property Protection In this section, we outline the frameworks for intellectual property protection regarding multimedia database systems: copyrightable database and patentable retrieval mechanism. 3.1 Multimedia Database as Copyrightable Entity Our framework for copyrighting the multimedia database determines which type of database should be copyrightable in the form of a component of contents-plus-indexes [12, 13, 14]. The collection of static indexes and individual contents forms a component of contents-plusindexes. That component identifies the entire content of each database, as is a static and copyrightable compilation. Copyrightable compilation is to be of sufficient creativity, i.e., originality in the form of a component of contents-plus-indexes. The set of conditions on the original categorization regarding indexes or metadata is formulated as below [13, 14]: A categorization regarding indexes or metadata is original only when 1. The type of indexes or metadata accepts discretionary selection in the domain of a problem database; otherwise, 2. The type of taxonomy regarding indexes or metadata accepts discretionary selection in the domain of a problem database. A typical case of non-original categorization is a photo film album database which has indexes of consecutive numbers. That case does not accept any discretion in the selection
184
H. Sasaki and Y. Kiyoki / Frameworks for Intellectual Property Protection
I. PATENTABLE SUBJECT MATTER OF PROCESSES FOR CBIR FUNCTIONS + TECHNICAL CONTRIBUTION (E.U.) (ENTRANCE TO PATENT PROTECTION)
DO CLAIMS COMPRISE THE MEANS no FOR PARAMETER SETTING ? yes
PATENTABLE OR NOT AS THE PROCESSES RELATED TO CBIR FUNCTIONS (out)
II. NONOBVIOUSNESS OR INVENTIVE STEPS (TECHNICAL ADVANCEMENT) (1) DO THE PRIOR ARTS PREDICATE A COMBINATION OF THE MEANS FOR PERFORMING A FUNCTIONAL PROCESS OF CBIR ? no (2) DO THE FUNCTION REALIZE QUANTITATIVE AND/OR QUALITATIVE ADVANCEMENT ?
(1-a) DO THE DESCRIPTIONS SPECIFY THE FORMULAS FOR PARAMETER SETTING ?
(1-b)DOES THE DISCUSSED INVENTION HAVE A CO-PENDING APPLICATION THAT SPECIFIES THE ABOVE FORMULAS ?
no (out)
yes DO THE PROCESSES HAVE THE IMPROVED FORMULAS FOR PARAMETER SETTING BASED ON PRIOR DISCLOSED MEANS FOR CBIR ? yes no (2-a) DO THE PROCESSES REALIZE A NEW FUNCTION BY COMBINING THE PRIOR DISCLOSED MEANS ?
yes
OBVIOUS NOT AS no PATENTABLE
no
yes
no (out)
yes (2-b) DO THE DESCRIPTIONS DISCLOSE EXAMPLES OF VALUES ON PARAMETER SETTING ?
no (out)
yes (I) WORKING OR PROPHETIC EXAMPLES OF INITIAL VALUES OR WEIGHTS ON PARAMETER SETTING
yes
III. ENABLEMENT OR CLARITY OF CLAIMS (CLEAR SPECIFICATION) PATENTABLE AS A DOMAINSPECIFIC APPROACH OF CBIR
(ii) WORKING EXAMPLES OF THE RANGE OF VALUES ON PARAMETER SETTING
PATENTABLE AS A DOMAINGENERAL APPROACH OF CBIR
Figure 2: Formulation for patenting the retrieval processes.
of the type of indexes or metadata, or the type of taxonomy. The photo film album database uses its respective film numbers as indexes for its retrieval operations. The taxonomy of the indexes is only based on the consecutive numbering without any discretion in its selection of the type of indexes or taxonomy regarding a multimedia database. Meanwhile, the discretionary selection of the type of indexes or metadata, or taxonomy constitutes copyrightable compilation of minimal creativity, i.e., originality on the categorization regarding indexes or metadata. A typical case of discretionary selection of the type of indexes or metadata is the web document encyclopedia as a multimedia database. Suppose that a database restores pictures of starfish which are manually and numerically numbered by day/hour-chronicle interval that is based on their significant life stages from birth to death. That database is to be an original work of authorship as a copyrightable compilation in the form of a component of contents-plus-indexes. That database of discretionary type of numbering or indexing is an original, i.e., copyrightable database. 3.2 Multimedia Database System as Patentable Mechanism Our framework for patenting the retrieval mechanisms of multimedia database system determines which type of retrieval mechanism should be patentable in the form of a component of novel combination of prior disclosed processes and/or a component of specific parameter setting (See Fig. 2) [15, 16, 17, 14]. The frameworks focus on the following three requirements for patentability: “patentable subject matter” (entrance to patent protection), “nonobviousness” (technical advancement) and “enablement” (specification) [18]. The requirement for nonobviousness on the combination of the processes for data-
H. Sasaki and Y. Kiyoki / Frameworks for Intellectual Property Protection
185
processing as the retrieval mechanism in a multimedia database system is listed as below [17]: 1. The processes for performing a retrieval mechanism must comprise the combination of prior disclosed means to perform certain mechanism which is not predicated from any combination of the prior arts; in addition, 2. The processes for performing a retrieval mechanism must realize quantitative and/or qualitative advancement. Otherwise, the discussed processes are obvious so that they are not patentable as the processes for performing a retrieval mechanism. First, a combination of prior disclosed means should not be “suggested” from any disclosed means “with the reasonable expectation of success” [19]. Second, its asserted function on the discussed mechanism must be superior to the conventional functions which are realized in the prior disclosed or patented means in the field of the retrieval mechanism of multimedia database system. On the latter issue, several solutions for performance evaluation are proposed and applicable. Another general strategy is restriction of the scope of problem claims into a certain narrow field to which no prior arts have been applied. This claiming strategy is known as the local optimization of application scope. A component for parameter setting realizes thresholding operations in the form of a computer program with a set of ranges of parametric values. In retrieval mechanisms, parametric values determine, as thresholds, which candidate image is similar to an exemplary requested image by computation of similarity of visual features [20, 21, 22, 23]. That parameter setting component is to be a computer-related invention in the form of computer program as far as that parameter setting is sufficiently specified to enable a claimed invention or retrieval mechanism [24]. The requirement for enablement on the parameter setting component of the retrieval mechanism in a multimedia database system is listed as below [17]: (1-a) The descriptions of the processes for performing a retrieval mechanism must specify the formulas for parameter setting; otherwise, (1-b) the disclosed invention of the processes should have its co-pending application that describes the formulas in detail; in addition, (2-a) the processes must perform a new mechanism by a combination of the prior disclosed means; otherwise, (2-b) the processes should have improved formulas for parameter setting which is based on the prior disclosed means for performing a retrieval mechanism, and also should give examples of parametric values on parameter setting in descriptions. For 2-b, the processes must specify the means for parameter setting by “giving a specific example of preparing an” application to enable those skilled in the arts to implement their best mode of the processes without undue experiment [25, 26]. U.S. Patent and Trademark Office [24, 27] suggested that the processes comprising the means, i.e., the components for parameter setting must disclose at least one of the following examples of parametric values on parameter setting: (i) Working or prophetic examples of initial values or weights on parameter setting; (ii) Working examples of the ranges of parametric values on parameter setting. The “working examples” are parametric values that are confirmed to work at actual laboratory or as prototype testing results. The “prophetic examples” are given without actual work by one skilled in the art.
186
H. Sasaki and Y. Kiyoki / Frameworks for Intellectual Property Protection
The retrieval mechanisms of multimedia database systems are patentable in the form of components of novel combinations of prior disclosed processes and/or components of specific parameter settings while they are to satisfy the above conditions. 3.3 A Simulation Example for the Formulated Procedural Diagram The proposed formulation in Fig. 2 should be clear with its application to an exemplary multimedia database system. We apply it to “Virage Image Retrieval”(VIR), which was developed in the early 1990s as a typical content-based retrieval of visual objects stored in digital image database systems. VIR is an indexing method for an image search engine with “primitives”, which compute similarity of visual features extracted out of typical visual objects, e.g., color, shape and texture of images. VIR evaluates similarity of images with ad hoc weights, i.e., parametric values, which are given to the parameter setting components for correlation-computation, by user-preference. Its claims consist of “function containers” as means-plus-functions for feature extraction and similarity computation. Its first claim, as described below, constitutes the primitives as the means-plus-functions. Those primitives realize a domain-general approach of CBIR by the formulas on parameter setting. VIR Claim # 1. A search engine, comprising: a function container capable of storing primitive functions; . . . a primitive supplying primitive functions . . . . . . , wherein the primitive functions include an analysis function . . . . . . of extracting features from an object . . . . First in Fig. 2, on its patentable subject matter, its retrieval processes consisting of the formula for parameter setting are to be determined as patentable subject matter in the form of computer programs. Those data-processing processes generate physical transformation on a specific machine, i.e., a computer memory with certain classification results. Second, on its nonobviousness, those data-processing processes are inventive steps that consist of combinations of the prior arts on thresholding functions as implemented in the integration of classification based on similarity computation, visual feature extraction and automatic indexing techniques. Those combinations are not predicated from any conventional keyword-based retrieval technique. Third, on its enablement, VIR’s description of preferred embodiments gives its clear specification on the formulas for parameter setting that realizes a domain-general approach of CBIR that was a brand new technology at the time. VIR Description . . . . . . For primitives having multiple dimensions, . . . . . . , An equation for an exemplary Euclidean metric is as follows. Primitive design. A primitive encompasses a given feature’s representation, extraction, and comparison function. . . . . . . The constraints are as follows: Primitives, in general, map to cognitively relevant image properties of the given domain. The formulation should take advantage of a threshold parameter (when available),. . . . . . . The retrieval mechanisms of multimedia database systems are patentable in the form of components of novel combinations of prior disclosed processes and/or components of specific parameter settings while they are to satisfy the above conditions.
H. Sasaki and Y. Kiyoki / Frameworks for Intellectual Property Protection
187
3.4 Trade Secret in Parameter Settings An emerging problem is discussed on the parameter settings of retrieval mechanisms. Patent application on the parameter setting components demands applicants as developers to make public the detailed know-how on the best range of parametric values in practice. The discovery of those parametric values needs considerable pecuniary investment in research and development. That kind of knowledge should be kept covered in the form of trade secret but not be open in public via patent application. The multimedia database community demands a scheme that determines which parameter setting component should be patentable or kept secret regarding multimedia database systems. 3.5 Embedding Trade Secret in Parameters It is necessary to prepare a scheme that determines how and which part of parameter setting components should take the form of trade secret. The problem is how to interpret the “working examples” of initial values or weights on parameter setting and the ranges of parametric values. The requirement for patenting parameter setting components as computer-related inventions demands inventors to make public their discovered “working examples” on those parameter values: initial values or ranges. The practice in patent application, nonetheless, does not always force applicants to disclose to examiners complete and perfect evidences on those initial values or ranges of parametric values, but those values as should work in their best mode at the present art. In the reality of application practice, inventors have three choices for embedding trade secrets on their know-how of parametric values in the forms of patentable parameter components: 1. On the initial values, their prophetic examples should be disclosed in patent application, instead of working examples; 2. On the ranges of parametric values, those ranges should be widened as possible at the best but not complete mode; 3. Otherwise, the ranges of parametric values should be replaced with several initial values of prophetic examples. The issue is when those patentable parameter setting components should be allowed to embed trade secrets on their parametric values. The framework or set of conditions to realize that problem depends on application cases. 4 Conclusions and Future Works In this article, we have discussed issues on intellectual property protection regarding multimedia database systems which consist of indexed multimedia digital contents in databases and retrieval mechanisms. We have presented the frameworks for copyrighting the database of multimedia database systems in the form of a component of contents-plus-indexes, and for patenting the retrieval mechanism of multimedia database systems in the form of a combination of processes and/or a component of parameter settings. We have also pointed out an emerging problem on the trade secret of parameter setting components and the possible directions for its solution.
188
H. Sasaki and Y. Kiyoki / Frameworks for Intellectual Property Protection
We are working to formulate a framework when certain patentable parameter setting components embed trade secret on their parametric values in several industrial fields of application software: authentication and encryption in the music file share industry and the embedded systems and software in the automobile electronics industry. In the field of visual information retrieval, the multimedia database community faces a variety of problems. Especially, a problem demands urgent solutions for the future progress of multimedia database systems: a framework for protecting a database as a whole. In addition, databases contain multimedia information including images and videos. Portable information devices allow people to easily access to a large amount of downloadable multimedia files stored in distributed databases around the world. Network technology for efficient data transactions often triggers unauthorized misappropriation of those multimedia files that are important intellectual assets. Even the sui generis right of database protection discussed in Europe is not to protect any database as a whole in the present legal system. It is necessary to prepare a framework for protecting entire databases including their contents. That framework should determine how and which type of database is to be protected as a whole. Acknowledgements This study is supported financially in part by the Grant-in-Aid for Scientific Research (“KAKENHI”) of the Japanese Government: No. 18,700,250 (FY 2006-2009). This study is also supported financially in part by the Microsoft Grant on Intellectual Property Research Promotion for the Year of 2005. References [1] J.M. Jakes and E.R. Yoches, Legally Speaking: Basic Principles of Patent Protection for Computer Science, Communications of the ACM 32(8) (1989) 922–924 . [2] C. Junghans and A. Levy, Intellectual Property Management: A Guide for Scientists, Engineers, Financiers, and Managers, Hoboken, NJ: John Wiley & Sons. (2006) . [3] U.S. Copyright Act, 17 U.S.C. Sec. 101, & 103, (2005) . [4] R.A. Gorman and J.C. Ginsburg, Copyright: Cases and Materials (6th ed.), University casebook series. Charlottesville, NC: The Michie Company. (2002) . [5] M.B. Nimmer, P. Marcus, D.A. Myers, and D. Nimmer, Cases and Materials on Copyright & Other Aspects of Entertainment Litigation Including Unfair Competition (7th ed.), Dayton, OH: LexisNexis. (2006) . [6] American Dental Assfn v. Delta Dental Plan Assfn, 126 F.3d 977 (7th Cir. 1997) . [7] J. Reinbothe, The Legal Protection of Non-creative Databases. In: Proc. of the Database Workshop of the International Conference of Electronic Commerce and Intellectual Property, (WIPO/EC/CONF/99/SPK/22-A), WIPO. Geneva, Switzerland, September 14–16, 1999. [8] P. Samuelson, Legally Speaking: Legal Protection for Database Content, Communications of the ACM 39(12) (1996) 17–23. [9] T. Aplin, Copyright Law in the Digital Society: The Challenges of Multimedia, Oxford, U.K.: Hart Publishing. (2005) . [10] U.S. Patent Act, 35 U.S.C. Sec. 101, 103, & 112, (2005) . [11] In re Alappat, 33 F.3d 1526, 31 U.S.P.Q.2d 1545 (Fed. Cir. 1994) (en banc) . [12] H. Sasaki and Y. Kiyoki, A Proposal for Digital Library Protection. In: Proc. of the 3rd ACM/IEEE-CS Joint Conference on Digital Libraries, Los Alamitos, CA: IEEE Computer Society Press. Houston, TX, May 27–31, 2003, p. 392 .
H. Sasaki and Y. Kiyoki / Frameworks for Intellectual Property Protection
189
[13] H. Sasaki and Y. Kiyoki, Copyrighting Digital Libraries from Database Designer Perspective. In: Proc. of the 7th International Conference on Asian Digital Libraries (ICADL), Lecture Notes in Computer Science, 3334. Berlin: Springer-Verlag. Shanghai, China, December 11–14 (2004) pp. 626–629 . [14] H. Sasaki and Y. Kiyoki, Multimedia Digital Library as Intellectual Property. In: Design and Usability of Digital Libraries: Case Studies in the Asia Pacific, Idea Group Press. (2005) 238– 253 . [15] H. Sasaki and Y. Kiyoki, Patenting Advanced Search Engines of Multimedia Databases. In: S. Lesavich (Ed.), Proc. of the 3rd International Conference on Law and Technology , International Society of Law and Technology (ISLAT). Anaheim, Calgary, Zurich: Acta Press. Cambridge, MA, November 6–7 (2002) pp. 34–39 . [16] H. Sasaki and Y. Kiyoki, Patenting the Processes for Content-based Retrieval in Digital Libraries. In: E.-P. Lim, S. Foo, C. Khoo, H. Chen, E. Fox, S. Urs, & T. Costantino (Eds.) Proc. of the 5th International Conference on Asian Digital Libraries (ICADL), Lecture Notes in Computer Science, 2555. Berlin: Springer-Verlag. Singapore, December 11–14 (2002) pp. 471–482 . [17] H. Sasaki and Y. Kiyoki, A Formulation for Patenting Content-based Retrieval Processes in Digital Libraries, Journal of Information Processing and Management 41(1) (2005) 57–74 . [18] R.P. Merges and J.F. Duffy, Patent Law and Policy: Cases and Materials (3rd ed.), Dayton, OH: LexisNexis. (2002) . [19] In re Dow Chemical Co., 837 F.2d 469, 473, 5 U.S.P.Q.2d 1529, 1531 (Fed. Cir. 1988) . [20] Y. Rui, T.S. Huang, and S.F. Chang, Image Retrieval: Current Techniques, Promising Directions and Open Issues, Journal of Visual Communication and Image Representation 10(4) (1999) 39–62. [21] A.W.M. Smeulders, M. Worring, S. Santini, A. Gupta, and R. Jain, Content-based Image Retrieval at the End of the Early Years, IEEE Trans. on Pattern Analysis and Machine Intelligence 22(12) (2000) 1349–1380. [22] A. Yoshitaka and T. Ichikawa, A Survey on Content-based Retrieval for Multimedia Databases, IEEE Trans. on Knowledge and Data Engineering 11(1) (1999) 81–93. [23] S. Deb, Multimedia Systems and Content-based Retrieval, Hershey, PA: Idea Group Inc. (2004) . [24] U.S. Patent and Trademark Office, Examination Guidelines for Computerrelated Inventions, 61 Fed. Reg. 7478 (Feb. 28, 1996) (“Guidelines”). Available: http://www.uspto.gov/web/offices/pac/dapp/oppd/patoc.htm. (1996) . [25] Autogiro Co. of America v. United States, 384 F.2d 391, 155 U.S.P.Q. 697 (Ct. Cl. 1967) . [26] Unique Concepts, Inc. v. Brown, 939 F.2d 1558, 19 U.S.P.Q.2d 1500 (Fed. Cir. 1991) . [27] U.S. Patent and Trademark Office, Examination Guidelines for Computerrelated Inventions Training Materials Directed to Business, Artificial Intelligence, and Mathematical Processing Applications (“Training Materials”). Available: http://www.uspto.gov/web/offices/pac/compexam/examcomp.htm. (1996) .
190
Information Modelling and Knowledge Bases XIX H. Jaakkola et al. (Eds.) IOS Press, 2008 © 2008 The authors and IOS Press. All rights reserved.
Wavelet and Eigen-Space Feature Extraction for Classification of Metallography Images Pavel Praks a,1 , Marcin Grzegorzek b , Rudolf Moravec c , Ladislav Válek d , and Ebroul Izquierdo b a Dept. of Applied Mathematics, Technical University of Ostrava, Czech Republic b Multimedia & Vision Research Group, Queen Mary University of London, UK c Research and Development, Mittal Steel Ostrava plc, Ostrava, Czech Republic d QT - Production Technology, Mittal Steel Ostrava plc, Ostrava, Czech Republic Abstract. In this contribution a comparison of two approaches for classification of metallography images from the steel plant of Mittal Steel Ostrava plc (Ostrava, Czech Republic) is presented. The aim of the classification is to monitor the process quality in the steel plant. The first classifier represents images by feature vectors extracted using the wavelet transformation, while the feature computation in the second approach is based on the eigen-space analysis. Experiments made for real metallography data indicate feasibility of both methods for automatic image classification in hard industry environment. Keywords. Measurement, hard industry, human factors, content-based image retrieval, wavelet transformation, statistical classification, numerical linear algebra, partial symmetric eigenproblem, iterative solvers.
1. Introduction Any meaningful human activity requires perception. Under perception realization, evaluation, and interpretation of sensory impressions is understood. It allows the human to acquire knowledge about the environment, to react to it, and finally to influence it. There is no reason in principle why perception could not be simulated by some other matter, or instance, a digital computer [6]. The aim of the simulation is not the exact modeling of the human brain activities, but the obtainment of similar perception results. Research activities concerned with the mathematical and technical aspects of perception are the field of pattern recognition. One of the most important perceptual abilities is vision. The processing of visual impressions is the task of image analysis. The main problem of image analysis is the recognition, evaluation, and interpretation of known patterns or objects in images. 1 Dept. of Information and Knowledge Engineering, University of Economics, Prague, Czech Republic. Correspondence to: Pavel Praks, Dept. of Applied Mathematics, VŠB – Technical University of Ostrava, 17. listopadu 15, CZ 708 33 Ostrava, Czech Republic. Tel.: +420 59 732 4181; Fax: +420 59 691 9597; E-mail: [email protected].
191
P. Praks et al. / Wavelet and Eigen-Space Feature Extraction
2n × 2n Gray Level Image f κ,ρ ..
xm
.
l
k
s=0
s = −1
bs,0,0 bs,1,0 bs,2,0 bs,3,0 bs,0,0 bs,1,0
d2,s
bs,0,1 bs,1,1 bs,2,1 bs,3,1 bs,0,1 bs,1,1
2|bs|
..
bs,0,2 bs,1,2 bs,2,2 bs,3,2 .
d0,s
s = s = −2 bs
d2,s
d2,s+1
d0,s d1,s d1,s
d0,s+1
d1,s+1
bs,0,3 bs,1,3 bs,2,3 bs,3,3
Figure 1. 2D signal decomposition with the wavelet transformation for a local neighborhood of size 4× 4 pixels. The final coefficients result from gray values b0,k,l and have the following meaning: b−2 : low-pass horizontal and low-pass vertical, d0,−2 : low-pass horizontal and high-pass vertical, d1,−2 : high-pass horizontal and high-pass vertical, d2,−2 : high-pass horizontal and low-pass vertical.
In this paper, the problem of automatic pattern classification in real metallography images from the steel plant of Mittal Steel Ostrava plc (Ostrava, Czech Republic) is addressed. The objective is to monitor the process quality in the steel plant. For this reason two different image classification algorithms are used and compared in this contribution. The first one computes feature vectors with the wavelet transformation, while in the second one the eigen-space analysis is applied. The paper is structured as follows. Section 2 describes the theoretical background of the wavelet-based image classifier. In Section 3, intelligent image retrieval using the partial eigen-problem is presented. Experimental comparison of these two approaches for image classification follows in Section 4, while Section 5 closes the paper with some conclusions.
2. Statistical Wavelet-Based Classification In this section, a statistical wavelet-based approach for image classification is presented. Section 2.1 describes the training of statistical models for different image concepts. These models are then used for image classification, which is presented in Section 2.2. 2.1. Training of Statistical Concept Models Before images can be classified in the recognition phase (Section 2.2), statistical models Mκ for all image concepts Ωκ considered in a particular classification task are learned in the training phase. The concept modeling starts with the collection of training data. In this work real metallography images from a cooking plant are used for this reason. Subsequently, the original training images are converted and resized into gray level images of size 2n × 2n (n ∈ N) pixels. In all these preprocessed training images 2D local feature vectors cκ,m are extracted using the wavelet transformation [4]. Training images are divided into neighborhoods of size 2|bs| × 2|bs| (in Figure 1, 4 × 4 pixels). These neighborhoods are treated as 2D discrete signals b0 and decomposed to low-pass and high-pass coefficients. The resulting coefficients bsb, d0,bs , d1,bs , and d2,bs are then used for feature vector computation
192
P. Praks et al. / Wavelet and Eigen-Space Feature Extraction
cκ,m (xm ) =
ln(2sb|bsb|) ln[2sb(|d0,bs | + |d1,bs | + |d2,bs |)]
.
(1)
The feature vectors cκ,m are modeled by normal density functions pκ,m = p(cκ,m |μκ,m , σ κ,m ) .
(2)
Due to the large number of training images for each concept Ω κ , it is possible to estimate the mean value vector μκ,m and the standard deviation vector σ κ,m for all image locations xm , i. e., all feature vectors cκ,m . Finally, statistical models Mκ for all image concepts Ωκ are created and ready for use in the classification phase (Section 2.2) 2.2. Image Classification Once the concept modeling (Section 2.1) is finished, the system is able to classify images taken from a real world environment. First, a test image f is taken, preprocessed, and local feature vectors cm are computed in it in the same way as in the training phase (Section 2.1). Second, the classification algorithm based on the Maximum Likelihood (ML) Estimation is started. The task of the image classification algorithm is to find the concept Ω κb , (or just its index κ ) of the test image f . In order to do so, the density values for all concepts Ω κ have to be compared to each other. Assuming that the feature vectors c m are statistically independent on each other, the density value for the given test image f and concept Ω κ is computed with pκ =
m=M
p(cm |μκ,m , σ κ,m ) ,
(3)
m=1
where M is the number of all feature vectors in the image f . All data required for computation of the density value pκ with (3) is stored in the statistical concept model Mκ . These density values are then maximized with the Maximum Likelihood (ML) Estimation [10] κ = argmax pκ κ
.
(4)
Having the index κ of the resulting concept the classification problem for the image f is solved.
3. Latent Semantic Indexing In this section, we present the intelligent image retrieval using the partial eigen-problem. The numerical linear algebra is used as a basis for the information retrieval in the retrieval strategy called Latent Semantic Indexing, see for instance [1], [2]. LSI can be viewed as a variant of a vector space model, where the database is represented by the document matrix, and a user’s query of the database is represented by a vector. LSI also contains a low-rank approximation of the original document matrix via the Singular Value Decomposition (SVD) or the other numerical methods. The SVD is used as an automatic tool
P. Praks et al. / Wavelet and Eigen-Space Feature Extraction
193
for identification and removing redundant information and noise from data. The next step of LSI involves the computation of the similarity coefficients between the filtered user’s query and filtered document matrix. The well-known cosine similarity can be used for a similarity modeling. Recently, the methods of numerical linear algebra are also successfully used for the face recognition and reconstruction [5], image retrieval [8,7], as a tool for information extraction from internet data [9] and for iris recognition problem [7]. The "classical" LSI application in information retrieval algorithm has the following basic steps: i) The Singular Value Decomposition of the term matrix using numerical linear algebra. SVD is used to identify and remove redundant noise information from data. ii) The computation of the similarity coefficients between the transformed vectors of data and thus reveal some hidden (latent) structures of data. Numerical experiments pointed out that some kind of dimension reduction, which is applied to the original data, brings to the information retrieval following two main advantages: (i) automatic noise filtering and (ii) natural clustering of data with "similar" semantic. 3.1. Image Coding In our approach [8,7], a raster image is coded as a sequence of pixels. Then the coded image can be understood as a vector of an m-dimensional space, where m denotes the number of pixels (attributes). Let the symbol A denote a m × n term-document matrix related to m keywords (pixels) in n documents (images). Let us remind that the (i, j)element of the term-document matrix A represents the colour of i-th position in the j-th image document. 3.2. Implementation Details of Latent Semantic Indexing In this section we will describe the possible software implementation of the Latent Semantic Indexing method. Let the symbol A denotes the m × n document matrix related to m pixels in n images. The aim of SVD is to compute decomposition A = U SV T ,
(5)
where S ∈ Rm×n is a diagonal matrix with nonnegative diagonal elements called the singular values, U ∈ Rm×m and V ∈ Rn×n are orthogonal matrices1 . The columns of matrices U and V are called the left singular vectors and the right singular vectors respectively. The decomposition can be computed so that the singular values are sorted in decreasing order. The full SVD decomposition (5) is memory and time consuming operation, especially for large problems. Although the document matrix A is often sparse, the matrices U and V have a dense structure. Due these facts, only a few k-largest singular values of A and the corresponding left and right singular vectors are computed and stored in memory. The number of singular values and vectors which are computed and kept in memory 1A
matrix Q ∈ Rn×n is said to be orthogonal if the condition Q−1 = QT holds.
194
P. Praks et al. / Wavelet and Eigen-Space Feature Extraction
can be chosen experimentally as a compromise between the speed/precision ratio of the LSI procedure. We implemented and tested LSI procedure in the Matlab system by Mathworks. Following [2] the Latent Semantic Indexing procedure can be written in Matlab by the following way. Procedure Original LSI [Latent Semantic Indexing] function sim = lsi(A,q,k) % Input: % A ... the m × n matrix % q ... the query vector % k ... Compute k largest singular values and vectors; k ≤ n % Output: % sim ... the vector of similarity coefficients [m,n] = size(A); 1. Compute the co-ordinates of all images in the k-dimensional space by the partial SVD of a document matrix A. [U,S,V] = svds(A,k); % Compute the k largest singular values of A; The rows of V contain the coordinates of images. 2. Compute the co-ordinate of a query vector q qc = q’ * U * pinv(S); % The vector qc includes the co-ordinate of the query vector q; The matrix pinv(S) contains reciprocals of non-negative singular values (an pseudoinverse); The symbol ’ denotes the transpose superscript. 3. Compute the similarity coefficients between the co-ordinates of the query vector and images. for i = 1:n % Loop over all images sim(i)=(qc*V(i,:)’)/(norm(qc)*norm(V(i,:))); end; % Compute the similarity coefficient for i-th image; V (i, :) denotes the i-th row of V . The procedure lsi returns to a user the vector of similarity coefficients sim. The i-th element of the vector sim contains a value which indicate a "measure" of a semantic similarity between the i-th document and the query document. The increasing value of the similarity coefficient indicates the increasing semantic similarity. 3.3. Partial Eigen-problem The image retrieval process can be powered very effectively when the time consuming Singular Value Decomposition of LSI is replaced by the partial symmetric eigenproblem, which can be solved by using fast iterative solvers [7]. Let us assume the following relationship between the singular value decomposition of the matrix A and the symmetric eigenproblem of the symmetric square matrices A T A:
P. Praks et al. / Wavelet and Eigen-Space Feature Extraction
195
Figure 2. An example of results of the wavelet (left) and partial eigen-problem based image retrieval (right). The query image is situated in left up corner and it is related to the query LCT52XP1010229_a.jpg. The well-classified images are at positions 1, 3 and 5 for the wavelet method and at positions 1 – 3 for the partial eigen-problem method.
A = U SV T
(6)
AT = (U SV T )T = V S T U T
(7)
AT A = V S T (U T U )SV T = V S T SV T
(8)
Moreover, let us assume the SVD decomposition (5) again. Because of the fact that the matrix V is orthogonal, the following matrix identity holds: AV = U S.
(9)
Finally, we can express the matrix U in the following way: AV S + ≈ U
(10)
Here the symbol S + denotes the Moore-Penrose pseudoinverse (pinv). Let us accent that the diagonal matrix S contains only non-negative singular values for real cases; The singular values less than tol ≈ 0 are cut off by the Matlab eigs(A’*A, k) command. There is no exact routine for the selection of the optimal number of computed singular values and vectors [3]. For this reason, the number of singular values and associated singular vectors used for the partial symmetric eigenproblem was estimated experimentally, but it seems that k < 10 is suitable for real image databases. For example, we choose k = 8 for the large-scale NIST TRECVID 2006 data [11]. In contrast to the SVD approach, the size of the partial symmetric eigenproblem (the size of AT A matrix) does not depend on the number of pixels (keywords) at all. Since the number of computed singular values k << n for real problems and k is small, the image retrieval using the partial symmetric eigenproblem is more efficient [7] than the "classical" SVD approach [2].
196
P. Praks et al. / Wavelet and Eigen-Space Feature Extraction
Figure 3. An example of results of the wawelet (left) and partial eigen-problem based image retrieval (right). The query image is situated in left up corner and it is related to the query LDT70SP1010023_a.jpg. The well-classified images are at positions 1 – 3 and 7 for the wavelet method and at positions 1 – 4, 6 and 7 for the partial eigen-problem method.
Figure 4. An example of results of the wawelet (left) and partial eigen-problem based image retrieval (right). The query image is situated in left up corner and it is related to the query SCK60U12.jpg. The well-classified images are at positions 1 – 4 and 7 – 9 for the wavelet method and at positions 1 – 3, 5 and 7 for the partial eigen-problem method.
4. Experiments and Results 4.1. Experimental Data We experimented with real metallography images taken from the steel plant of Mittal Steel Ostrava plc, Ostrava, Czech Republic. In fact, we deal with sample images of con-
P. Praks et al. / Wavelet and Eigen-Space Feature Extraction
197
Figure 5. An example of results of the wawelet (left) and partial eigen-problem based image retrieval (right). The query image is situated in left up corner and it is related to the query SDK53M27. All retrieved images (except image no. 2) are well-classified by the wavelet method. All retrieved images (except image no. 4) are well-classified by the the partial eigen-problem method.
tinuously cast steel from billet device for continuous steel casting. This device produces billets of 180 mm square, 160 and 210 mm round. The closer parameters are stated in the Table 1. Steel samples from the cast billets are taken away for the device for continuous steel casting. These are crosscuts of the cast billets. These samples are conveyed into metallography lab where they are mechanically adjusted. In order to stress a sample macrostructure, crosscut etching is done. Consequently, photographs of these etched crosscuts are being taken. Commissioned on:
7 December 1993
Type:
billet, radial, two-point alignment
Heat volume:
205 tons
Casting method:
closed, through submerged nozzles and stoppers
Casting arc radius:
10.5 ; 21m (two-point alignment)
Cooling of semi-product:
water (single component)
Cutting of semis:
torch cutting
Slab marking: punching, 10-character code Table 1. Basic chosen parameters of the device for continuous casting No. 1.
Evaluation time of one image
0.36 secs.
Local feature vectors from neighborhoods
8 × 8 pixels
Type of wavelet transformation 8 TAB Johnshon Wavelet Table 2. Properties of the Statistical Wavelet-Based Classification method.
198
P. Praks et al. / Wavelet and Eigen-Space Feature Extraction
Properties of the document matrix A Number of keywords Number of documents Size in memory
458×480 = 219 840 40 67.089 MB
The SVD-Free LSI processing parameters Dim. of the original space 40 Dim. of the reduced space (k) 6 Time for AT A operation 0.64 secs. Results of the eigensolver 0.219 secs. The total time 0.859 secs. Table 3. Image retrieval using the partial eigen-problem method; Properties of the document matrix (up) and LSI processing parameters (down).
Photographs from the verification of electromagnetic steel mixing in the crystallizer have been used for verifying of the described methods. The total number of images in the image database was 83. The number of images in the training set was 20. 4.2. Experimental Results The results of image retrieval experiments are presented in Fig. 2 - Fig. 5. The resulted images are presented by decreasing order of similarity. The query image is situated in left up corner. The similarity of the query image and the retrieved image is written in parentheses. In order to achieve well arranged results, only 9 most significant images are presented. The presented shape of the crosscuts does not respond to a reality completely (they were slightly deformed at the photograph evaluation). It can be stated that these are the first results for billets 180mm square and 210mm round. The evaluated subject was the whole crosscut of billet samples. 4.3. Conclusions for Experiments Our results indicate that the both methods can automatically recognize the shape and the type of images found in our image database. The behaviour of both methods is close to the classification of a human expert. Moreover, the results of Table 2 and Table 3 indicate a possibility of real-time analysis. The first results point out that the discussed methods can be also used for the evaluation of crosscut macro structure of billet samples. In order to achieve more precise evaluation results, individual areas of a sample crosscut images should be deeply analyzed in the future work. This deeper image analyze is also important for searching metallurgical relations in images, which are hidden in the image database.
5. Conclusion In this paper, a comparison of two approaches for automatic pattern classification in images taken from a real world environment has been presented. The experimental data in the form of metallography images has been provided by the Mittal Steel Ostrava plc (Ostrava, Czech Republic). The objective of this research activity is to monitor the quality
P. Praks et al. / Wavelet and Eigen-Space Feature Extraction
199
process in the steel plant. The first classifier used for this reason (Section 2) represents image patterns by feature vectors extracted with the wavelet transformation, while the second one (Section 3) is based on the eigen-space analysis. Classification results for experiments presented in Section 4 prove a very high performance of both approaches in a real world environment. In the future, the statistical wavelet-based approach (Section 2) will be combined with the eigen-based analysis (Section 3). One can imagine that a fusion of these two methods will bring a significant improvement in terms of classification rates. Acknowledgments The research has been also partially supported by a program "Information Society" of the Academy of Sciences of the Czech Republic, project No. 1ET401940412. The work leading to this contribution has been partially supported by the European Commission under contract FP6-027026-K-SPACE. References [1] W. M. Berry, Z. Drmaˇc, and J. R. Jessup. Matrices, vector spaces, and information retrieval. SIAM Review, 41(2):336–362, 1999. [2] D.A. Grossman and O.Frieder. Information retrieval: Algorithms and heuristics. Kluwer Academic Publishers, Second edition, 2000. [3] Berry W. M., Dumais S. T., and O’Brien G. W. Using linear algebra for intelligent information retrieval. SIAM Review, 37:573–595, 1995. [4] S. Mallat. A theory for multiresolution signal decomposition: The wavelet representation. IEEE Transactions on Pattern Analysis and Machine Intelligence, 11(7):674–693, July 1989. [5] N. Muller, L. Magaia, and B. M. Herbst. Singular value decomposition, eigenfaces, and 3d reconstructions. SIAM Review, 46(3):518–545, 2004. [6] H. Niemann. Pattern Analysis and Understanding. Springer-Verlag, Berlin, Heidelberg, Germany, 1990. [7] Praks P., Machala L., and Snášel V. On SVD-free Latent Semantic Indexing for iris recognition of large databases. Multimedia Data mining and Knowledge Discovery. Ed. V. A. Petrushin and L. Khan, Springer-Verlag London Limited, 2007. [8] P. Praks, J. Dvorský, and V. Snášel. Latent semantic indexing for image retrieval systems. In SIAM Conference on Applied Linear Algebra. The College of William and Mary, Williamsburg, USA, http://www.siam.org/meetings/la03/proceedings/Dvorsky.pdf, July 2003. [9] V. Svátek, M. Labský, P. Praks, and O. Šváb. Information extraction from html product catalogues: coupling quantitative and knowledge-based approaches. In Dagstuhl Seminar on Machine Learning for the Semantic Web. Research Center for Computer Science, Wadern, Germany, http://www.smi.ucd.ie/Dagstuhl-MLSW/proceedings/labskysvatek-praks-svab.pdf, February 2005. [10] A. R. Webb. Statistical Pattern Recognition. John Wiley & Sons Ltd, Chichester, UK, 2002. [11] P. Wilkins, T. Adamek, P. Ferguson, M. Hughes, G. J. F. Jones, G. Keenan, K. McGuinness, J. Malobabic, N. E. O’Connor, D. Sadlier, A. F. Smeaton, R. Benmokhtar, E. Dumont, B. Huet, B. Merialdo, E. Spyrou, G. Koumoulos, Y. Avrithis, R. Moerzinger, P. Schallauer, W. Bailer, Q. Zhang, T. Piatrik, K. Chandramouli, E. Izquierdo, L. Goldmann, M. Haller, T. Sikora, P. Praks, J. Urban, X. Hilaire, and J. M. Jose. K-space at trecvid 2006. In Proceedings of the TRECVid Workshop, Gaithersburg, Maryland, USA, 2006. NIST, http://wwwnlpir.nist.gov/projects/tvpubs/tv6.papers/k-space.pdf.
200
Information Modelling and Knowledge Bases XIX H. Jaakkola et al. (Eds.) IOS Press, 2008 © 2008 The authors and IOS Press. All rights reserved.
Semantic knowledge modeling in medical laboratory environment for drug usage: CASE study Anne TANTTARI, Kimmo SALMENJOKI Department of Computer Science, University of Vaasa Box 700, 65101 Vaasa, Finland Lorna UDEN Department of Computer Science, University of Staffordshire Beaconside Staffordshire, UK Abstract. In this paper we will consider the usage of knowledge modelling and design with conceptual design and analysis in improving the knowledge description of medical data. There exists many XML and information system (IS) based approaches for producing easy to use, reliable, trustworthy and coherent medical systems and services, like the concepts of HL7. In computer applications web, web services and semantic web based technologies and tools are being used in the business and ecommerce scenarios to produce user-focused, (business) process oriented services in future medical IS. Our case is focused on laboratory exams and their knowledge analysis for predicting and analyzing the usage of various drugs by patients. We show with our case the role of semantic knowledge modelling in coordinating various medical services in a real hospital setting, and hope to extend our work in the future towards semantic web based applications that improve and ease the patient treatments in the IS of the future hospitals.
1. Introduction Traditionally medical care is highly technology and information technology (IT) oriented. In Finland the coordination and collaboration between medical organizations have been advocated and increased in the last ten years. In last years, the overall process of patient treatment and its integration with various existing IS has caught the main attention of development. In this paper we will focus on the knowledge modeling and design of laboratory diagnosis. We describe the usage of semantic knowledge design for this case. We will describe more generally the role and possibilities of present knowledge modeling and design, mainly originating from the software development and system theoretical scenarios [6].
2. Satellite model for the Combined Information of the Laboratory Exams and Drugs “A conceptual model is a model of a subject area or area of knowledge, sometimes called a domain, that represents the primary entities (the things of the domain), the relationships among entities, the attribute values (sometimes called properties and property values) of the entities and the relationships, and sometimes rules that associate entities, relationships, and attributes (or all three) in more complicated ways.”[2]. Semantic web with enhanced knowledge modelling provides the generic setting of refining the information granularity for any domain
A. Tanttari et al. / Semantic Knowledge Modeling in Medical Laboratory Environment
201
of applications [1]. Using databases which are traditionally modelled with ERD diagrams as sources of data, we will use information analysis in describing the inherent properties and interrelations of laboratory data with the de facto standard of Finnish medicine descriptions in Pharmaca Fennica [12]. Pharmaca Fennica is an encyclopaedia produced by the Pharmaceutical Information Centre containing all the drugs existing in the Finnish market. For each drug, this contains its ingredients and the related diseases and traumas that can be cured with that specific medicine. The information of this book is reliable and fully covering all materials as well as available online. Ultimately, the improved knowledge awareness will lead towards shared ontologies and unified knowledge sources [9, 11]. In defining the properties of our information entities with their relations, we will form the taxonomy for laboratory test usage. This analysis was done informally and intuitively using the domain expertise of the real world experts. The outcome of this analysis was a formal model of our problem. Although there are various approaches for concept analysis, we believe that the most suitable model to be used is the satellite model, which is based on the inherent structure and consumption models of the laboratory processes and treatment approaches. This is because the satellite model is good in embracing various roles and models into a unified knowledge description. In our case, the knowledge contained in the model will be shared and utilized by various persons and medical process in daily work. There are varying ways of using satellite model in practice. When we look at the medical descriptions of laboratory tests they share several features. By features we mean not only fixed facts but also the connections arising from the domain context. In looking at the documents we found, for example, that in certain laboratory tests the bentsodiatsepin is an element of attention in the test description. However, the present descriptive documents do not contain this concept based either in its universal nature or in its typical property classification when comparing it with other related features. Hence these laboratory tests share the common property that they are a group of tests that are used for finding the usage of bentsodiatsepin as a drug. Of course, these tests also have a multitude of other knowledge and information properties, which will be here left out due to complexity as well as to maintain clarity of our generic approach [3, 4]. Using the medical drug table and therapy group classification of Pharmaca Fennica [12] we have developed the detailed satellite model. In this paper we only give a demonstrative excerpt from the overall model in [16]: Neurological disease
Temesta Wyeth BENTSODIATSEPIN
dizziness
Loratsepam Xanor Pfizer
anxiety
psychopharmaceutical drug depression
Diatsepam
Alprox Orion Pharm
Oksatsepam
Alpratzolam Generics Merck NM
Xanor Depot Pfizer
Figure 1.
Alpratsolam
Opamox Diazepam Desitin Desitin
Oxamin Oxepam
Except from the satellite model, full model available in [16]
In the higher part of the medical hierarchy of Pharmaca Fennica the drug therapy main class and therapy subclasses have been used as a basis for the classificaion. Using the produced satellite model of Figure 1 we have collected a list of all those laboratory tests addressing
202
A. Tanttari et al. / Semantic Knowledge Modeling in Medical Laboratory Environment
bentsodiatsepin that appear in the Vaasa Central Hospital [5]. Using this model, we have come up with a following category of the laboratory exams related to bentsodiatsepin [15]: Laboratory test Medicine and poison test Bentsodiatsepin S-Diatsepam S-Alpratsolam S-Oxatsepam S-Nitratsepam S-Klobatsam S-Klonatsepam B-Drug_monitoring U-Drug_screening (qual) U-Drug_screening U-Drug, check and screening Figure 2.
Categories for bentsodiatsepin in lab exams
Conceptually this is a single heritance based concept system. When exams are ordered from the laboratory in the real life cases the measurement of bentsodiatsepin can be requested by various exam names. To cover these cases we have to use a multiple heritance description, where various super classes inherit two or more properties from the higher order concepts in the system, as shown in Figure 3. Laboratory test Medicine and poison test Bentsodiatsepin Diatsepam Name of the medicine Diapam Orion Pharma Diazepam Desitin Desitin Gastrodyn Comp Leiras Medipam Ratiopharm Relapamil Orion Pharma Setsolid Alpharma Setsolid Novum Alpharma Setsolid Prefill Alpharma Vertipam Orion Pharma Name of the test S-Diatsepam B-Drug_monitoring U-Drug_screening (qual) U-Drug_screening U-Drug, check and screening Figure 3.
Inherited categories for bentsodiatsepin in lab exams
The building of these hierarchical systems can either proceed as top down development process or as bottom-up development process. In the top down approach classes are refined into smaller, more specific classes. The higher concept hence processes less features than the lower ones. Depending on the case, we might also take an intermediate approach, where we develop the classification starting from the middle of the hierarchy with an origin that is highly meaningful from the domain context point of view. In general, this is done by a combination or scenario based development, see [8, 10].
A. Tanttari et al. / Semantic Knowledge Modeling in Medical Laboratory Environment
203
3. Using Semantic Knowledge Approach with Laboratory Documentations As mentioned earlier, various semantic web approaches have permeated most IS application, such as the use of XML as data layer in classical IT applications [14]. In the next section we describe how the satellite model of Figure 1 is translated to its technical implementation using RDF and its related expression formulations. Next we will illustrate two simple examples of utilizing our satellite model of previous chapter in real life medical documentation: 3.1 Changes in patient treatment process by doctor In the medical laboratory, finding the most relevant test from a multitude of tests often creates practical problems for doctors. Using the topic or test name based classification requires additional information in order to be successful. Because of the continuous changes in naming, it is critical to keep the system up to date all the time. Doing incorrect or unnecessary exams is bad, both from the patient’s and economical points of views. The following diagram exemplifies the approach when the additional knowledge of the model is being used: Disease or trauma
Figure 4.
Name of the medicine
Active drug
Name of the laboratory
The process how the doctor can find a name of the correct laboratory test
How ever to maintain this kind of system is complicated due to the constant changes in the data. Besides new drugs and those withdrawn from the market the tests themselves are continuously improved and some tests become obsolete. To ease this maintenance the shared knowledge between the Pharmaceutical Information Centre and laboratories is crucial. Here, again, semantic web provides means of sharing and combining the information sources. This approach is well accepted and used in the research of advanced knowledge retrieval on the web in general as well [21]. 3.2 Finding a lab exam related to a specific medicine When treating a patient the doctor has to know the link between the patient’s disease and related drugs. In reality, finding this link is a manifold communicative and systematic process between the domain expert and a patient with his/her own needs and ideas. Overall this process can be generically described as below: Name of the medicine
Figure 5.
Name of the laboratory test
The correct laboratory test related to a specific medicine via the model
When one wants to enhance this process with computer based knowledge, the conceptual model has to be both simple, easy to maintain and apt for automated processing. Here the key role of semantic metadata is most important. From the practical point of view the span of the visible concepts should not be too wide or deep. From the patient records the doctor will possess the knowledge of all the drugs that the patient is using at that moment. So, in most simple case, it would be sufficient to combine this knowledge with its specifically related laboratory tests to reduce the complexity and improve the quality of the medical care by simple automation.
204
A. Tanttari et al. / Semantic Knowledge Modeling in Medical Laboratory Environment
Various expert systems have been already developed using this approach (see [13]), but in our case, we want to improve the explicit usability of the deeper knowledge provided by the higher interlinking of these two variable information sources using semantic web methodologies and tools. In practice, this would require a search engine that would recommend a suitable drug based on the patient’s blood test and medical history as demonstrated in Figure 5. Of course, there exists a multitude of other viable medical cases for using semantic web in medical IS beyond these two cases mentioned here. 4. Using RDFS/RDF with Satellite Model for Practical Laboratory Work 4.1 Presenting the knowledge of laboratory data for knowledge processes For implementing the case of section 3.2, we next turn to the RDFS/RDF based knowledge descriptions of the laboratory documentation. As an example we show how RDFS/RDF statements related to finding the laboratory test will be written. Statements of this kind will form the basis for our systemized medical ontology: Table 1. Thesaurus linking medicines and lab exams
Name of the medicine Diapam Orion Pharma
Laboratory test S-Diatsepam
Frisium Sanofi-Aventis S-Klobatsam
This is directly derived from the satellite model of Figure 1. Here from the resource, property type and its values we can form the following sentences in English: - Diapam Orion Pharma named drug is researched by an exam named Sdiatsepam in the laboratory - Frisium Sanofi Aventis drug is researched by an exam named S-klobatsam Using N3 (Notation 3), which is a non-technical presentation of these so called RDF triples spells as: - [<#drug>”Diapam Orion Pharma”;<# exam >”S-Diatsepam”] - [<#drug >”Frisium Sanofi Aventis”;<# exam >”S-Klobatsam”], where # identifies a URI-address. For ultimate clarity these can be automatically reformatted as diagrams:
http://www.Pharmaceutical_Information_Centre/Na of_the_medicine/ Diapam Orion Pharma
S-Diatsepam
laboratory_handbook="http://www.vshp.fi/laboratory_handbook/Name of the laboratory test# http://www.Pharmaceutical_Information _Centre/Name_of_the_medicine/ Frisium Sanofi Aventis
S-Klobatsam
laboratory_handbook="http://www.vshp.fi/laboratory_handbook/Name of the laboratory test# Figure 6. Two triplets for finding exam names based on medicine names
A. Tanttari et al. / Semantic Knowledge Modeling in Medical Laboratory Environment
205
The above three presentations are used in the application domain by the experts as well as providing knowledge aware applications for describing the domain and case related processes for chapter 3. In building semantic applications various programming tools, like Jena, Joseki and JADE, allow building agent based systems that use the knowledge in dynamic manner beyond the primitive searching discussed in sections 3.1 and 3.2. 4.2 Example of using RDF for the knowledge of laboratory data and processes Using the previously discussed top down approach in detailing the information of the satellite model, we will next give the technical details of RDF as a practical example. When RDF bag elements are applied for our example data, we obtain the following RDF graph: rdf:bag
rdf.type
S-Diatsepam
rdf:_5
examines Diapam Orion Pharma
Name of the laboratory test
B-Drug_monitoring
rdf:_1 rdf:_2
U-Drug_screening_(qual) rdf:_3 rdf:_4
U-Drug_screening
U-Drug, check and screening
Figure 7. Description of a laboratory exam with RDF:bag as a graph
This diagram shows that the effect of the drug Diapam Orion Pharma can be examined by five different toxicity tests. The presented example gives hints as how the RDF- based knowledge could be used in the real medical cases with processes of section 3, see [19] for more details. 4.3 Presenting more complicated relations with RDFS graphs In XML usage the evolutionary improvement of information abstraction leads to the use of XML Schemas. Likewise in the semantic web setting, the growing of the RDF information will enable the developers of the documentation system to see the deeper relations between drugs and their treatment with related medical processes. For systematically describing these relations we will use RDF Schemas (RDFS) analogously to XML Schemas [17, 18]. The most important conceptual relation that RDFS represents is hypogyny. Hypogynies can be described using classes, instances and properties in RDFS (RDF-Schema). Figure 8 shows an example of multiple inheritance between the concepts in our case study with RDFS:
206
A. Tanttari et al. / Semantic Knowledge Modeling in Medical Laboratory Environment
rdf:Property Medicine and oison test
rdf:type
rdf:type
rdfs:domain rdfs:subClassOf
rdf:type rdfs:domain
Name of the laboratory test
name
rdfs:range rdfs:subClass
drug
Bentsodiatsepin
Diatsepam
person in charge _name rdfs:subClassOf
Alpratsolam
rdfs:range rdf:type
Oxatsepam
Name_of_the_medicine
rdf:type rdf:type
rdf:type rdf:type Rdfs:Class
Figure 8. RDF-schema for the ontology of bentsodiatsepin
Above bentsodiatsepin is defined as a subclass of ”Name of the laboratory test”. Each class typically contains one or many instances, for example this test is related to several other “Diatsepam” related tests. This is enabled as any resource can be an instance of several classes. With this feature any bentsodiatsepin related tests can be found with related medicines. The role of other properties is to describe attributes or relations to other resources. In RDFS we specify the scope or domain for these attributes with rdfs:domain values. This domain is also a resource by itself, which can thus appear as a subject for another property. Range specified with rdfs:range can appear as an object of an entity. For example, in our case, any medicine can appear as a subject for a toxic lab test and brand of medicine can appear as an object in related domain knowledge sentences written in RDFS. These will form the basis for the logical processing of the RDF statements of section 4.2. Ultimately, the domain inherent rules and processes (contained in the classical information systems) could be technically spelled out as OWL rule based sentences, providing a basis for the dynamic operation of the knowledge agents in knowledge processing systems beyond the classical expert systems, see [7, 21].
5. Conclusions In this paper we have demonstrated our approaches in using semantic web in the context of medical laboratory tests. It shows the principal advantage of knowledge analysis and provides a technological basis for developing knowledge supported and intensive medical treatment processes and systems. In subsequent papers we will address the medical treatment processes in general and their IS system using semantic web based tools and technologies. As the frameworks and approaches of shared knowledge get more popular, we will continue the research in systemizing the medical processes via their extended functional usage using these semantic web based knowledge models and descriptions together with web services on the software components. After this, our focus will move towards expert assistance, hospital process automation and later agent based approaches in consuming the already vast existing digital knowledge of medicine and its meaningful sharing in various medical cases.
A. Tanttari et al. / Semantic Knowledge Modeling in Medical Laboratory Environment
207
References [1] [2] [3] [4] [5] [6] [7] [8]
[9] [10] [11] [12] [13] [14] [15]
[16]
[17]
[18] [19] [20]
[21]
Berners-Lee, Tim (2003): Web Services - Semantic Web. Architectural layers to International WWW Conference.
208
Information Modelling and Knowledge Bases XIX H. Jaakkola et al. (Eds.) IOS Press, 2008 © 2008 The authors and IOS Press. All rights reserved.
Towards Automatic Construction of News Directory Systems Bin Liu, Pham Van Hai, Tomoya Noro, and Takehiro Tokuda {ryuu, hai, noro, tokuda}@tt.cs.titech.ac.jp Department of Computer Science, Tokyo Institute of Technology Meguro, Tokyo 152-8552, Japan
Abstract. Currently news sites and news index sites basically provide streams of news with simple classifications or keyword based searching with publication data. When we try to search for news involving a number of potential keywords or unknown keywords, then the task becomes manually tedious or almost impossible. We present automatic methods for constructing news directory systems which contain collections of news index information with flat or hierarchical classification structures. This directory structure enables us to reach the news articles without knowing the keywords exactly. We implemented and evaluated one sample news directory system.
1 Introduction On the Web it is not difficult for an ordinary user to manually access to a small number of domestic news sites in various countries/regions, such as New York Times, Guardian Unlimited and Straits Times, or to access to a number of global news sites covering various parts of our earth from their eyes, such as CNN International, BBC world, and Reuters. Also news index sites such as Google News US, Google News Canada and Google News Australia, tailored to concerns of intended audience countries/regions, provide index information to a large number of related news source sites. However, these news sites and news index sites provide two basic methods for access. Namely they provide streams of news with simple classifications such as Asia/Pacific section and Business section, or keyword based searching with publication data such as publication date and publisher name. Hence, if we would like to ask a global question, not tailored to particular audience countries/regions, involving a number of potential keywords or unknown keywords, keyword based searching for news articles may be manually tedious or almost impossible. For example, the question may be ”what kind of country/region names are frequently mentioned together with particular disease names at news sites in our world?” We present an automatic approach to these questions based on the idea of news directory systems. A news directory system is a collection of news index information automatically retrieved from various news sites on our earth and automatically classified into a flat or a hierarchical directory system. Users can customize directory systems, design the structures, and give definitions to directories for themselves. Index information would be classified into these directories automatically so that users can find information they need more quickly with these systems. We implemented and evaluated one sample news directory system. This
B. Liu et al. / Towards Automatic Construction of News Directory Systems
209
news directory systems allow us, for example, to take a look at news articles containing typical disease names such as Bird flu or AIDS in each instance subdirectories under Disease directory. The directory structure also enables us to reach the news articles without knowing the keywords exactly. The organization of the rest of this paper is as follows. In Chapter 2 we explain the overview of news directory systems. In Chapter 3 and 4 we respectively give methods for automatic construction of directory structures, methods for directory definitions and automatic placement. In Chapter 5 and 6 we explain our experimental evaluation, concluding remarks and future work respectively.
2 A News Directory System A news directory system has following subsystems for a user to take a look at news article index information automatically placed according to the given directory structures and the definitions of articles to be contained in each directory.
Figure 1: System Structure
Subsystem 1 Automatic retrieval of news article index information from news sites. Currently, we get the index information in two ways. To RSS news sites, we use RSS to retrieve news article pages. To those not providing RSS, we utilize the URL structure of news sites to retrieve news article pages. Subsystem 2 Definition of a directory structure with one level flat classification or multilevel tree classification. We can freely construct the directory structure as we wish or we can also use directory structures similar to that of WordNet [6] or Wikipedia [5]. Subsystem 3 Handling of definitions of news articles to be contained in each directory. We need to give each directory a definition with keyword expressions to specify articles.
210
B. Liu et al. / Towards Automatic Construction of News Directory Systems
Subsystem 4 Automatic placement of news article index information in each directory. Once the directories’ structures and definitions are given, the system will classify those index information collected and place them into corresponding directories automatically. Subsystem 5 A subsystem for query processing and visualization. Users can search for news articles they are interested in by following the directory structures or give search keywords directly. The search results can be also shown in a map. The structure of our news directory system is shown in Fig. 1.
3 Directory Structures In news directory systems we use one-level flat directory structures or multi-level tree directory structures. Typical examples of one-level flat directory structures may be as follows. • Classification of natural disasters such as typhoon and earthquake. • Classification of human diseases such as diabetes and malaria. Typical examples of multi-level tree directory structures may be as follows. • Classification of locations such as countries/regions on the earth and outside of the earth. • A small classification tree constructed from the large classification tree such as WordNet or Wikipedia classification.
Figure 2: Directory Disease
Figure 3: Directory Countries/Regions
An example of one-level flat directory structure is shown in Fig. 2 and an example of multi-level tree directory structure is shown in Fig. 3. Users can also build their original directory structures manually. Here we will give some methods to bulid directory structures with exsiting resources. Method 1 We use open knowledge collection of classifications by humans, such as Wikipedia and WordNet, to build an initial collection of instance names belonging to one category. Method 2 Our method of building multi-level tree directory is as follows. We need a small set of basic words. Such a set of basic words may be subject words in New York Times Topics Index or a subset of Longman defining vocabulary [4] or a subset of Oxford defining vocabulary [3]. For a given set of basic words we construct a small classification tree as follows.
B. Liu et al. / Towards Automatic Construction of News Directory Systems
211
1. We retrieve full paths of all basic words in the WordNet tree. 2. We construct the initial small tree using the full paths obtained in the step 1. 3. We construct the small tree by deleting all non-basic words having exactly one son node from the initial small tree. A process of construction of a multi-level tree directory is shown in Fig. 4.
Figure 4: Composition of a small classification tree
4 Directory Definition and Automatic Placement We need definitions of news articles to be contained in each directory and an automatic method of placing those news article index information into corresponding directories. 4.1 Directory Definitions Our default definition of a news article A to be contained in a directory B is that the article A has an occurrence of the word B. In addition to default definitions of single word occurrences, we may use explicit definitions of a news article in a directory using the expressions defined by following extended context-free syntax rules with repetition operator {} representing zero or more times of repetitions. expression → (term) {OR (term)} term → factor {AND factor} factor → (phrase)|(NOT phrase) phrase → word {SPACE word} word → character {character}
212
B. Liu et al. / Towards Automatic Construction of News Directory Systems
This expression allows us to define news articles having slightly more complicated word occurrences. For example, we may write a definition using the following expression. ((football) AND (NOT american football)) OR ((soccer)) This expression means that an article A is to be contained in the directory, if A contains the word ”football” but not ”american football” or A contains the word ”soccer”. The same expression may be written briefly as follows. 1. football AND (NOT american football) 2. soccer 4.2 Automatic Placement The task of automatic placement consists of two phases. In the first phase, we construct two finite-state automata M1 and M2 . In the second phase we actually classify news articles into directories using automata M1 and M2 . The automaton M1 recognizes each phrase using the transition by one character. The automaton M2 classifies a news article into corresponding directories according to expressions with the acceptance/non-acceptance result of each phrase by the automaton M1 . 4.2.1 The First Phase
We construct two automata M1 and M2 as follows. M1
1. We collect all defined phrases d1 , d2 , ..., dn consisting of characters and construct corresponding finite-sate automata M11 , M12 , ..., M1n , which have transition labels of one character and recognize defined phrases d1 , d2 , ..., dn respectively. 2. We construct a finite-state automaton M1 by applying subset construction method to the set of automata M11 , M12 , ..., M1n .
M2
1. We collect all expressions e1 , e2 , ..., en consisting of defined phrases and decompose each expression to terms t11 , t12 , ..., tn1 , ..., tnm 2. For each term, we construct a sequence consisting of sorted factors. Using the sequences, we construct corresponding finite-sate automata M21 , M22 , ..., M2k whose transition labels are phrase or NOT(phrase). 3. We construct a finite-state automaton M2 by applying subset construction method to automata M21 , M22 , ..., M2k . For the sample expressions of Section 4.1, we can construct M1 and M2 as follows.
B. Liu et al. / Towards Automatic Construction of News Directory Systems
213
4.2.2 The Second Phase
In the second phase we actually classify news articles into directories using automata M1 and M2 . Recognition of a phrase by M1 We run the automaton M1 with the initial control point in the initial state of M1 as follows. • If a phrase consists of one word, then the behavior of M1 is same as an ordinary automaton. • If a phrase consists of two or more words separated by spaces or other delimiters, then the behavior of M1 is as follows. – Each time we meet with a delimiter, then we introduce one more control point for recognizing the remaining postfix of the phrase from the initial state of M1 . Classification of news articles by M2 We run the automaton M2 with the initial control point in the initial state of M2 as follows. The input string consists of sorted phrases accepted by the automaton M1 . Each term of expressions has corresponding directories. If a control point reaches the end of the final phrase of a term by looking at the entire input string, then we associate the article with the corresponding directories. Otherwise, no corresponding directories exist. The basic behavior of M2 is as follows. If a state S has the control point for the first time, and the state S has a number of transition labels L1 , ..., Ln and corresponding next states N(L1 ), ..., N(Ln ), then we create one copy of the control point in each next state of S and delete the control point of S . Additional behavior is determined according to the transition label L and the first phrase p of the input string as follows. • If the transition label L is p, then the input string becomes the rest of the input string and the control point is in N(L) as above. • If the transition label L is NOT p, then we delete the control point in the next state N(L). • If the sorting ordering of the phrase of the transition label L is smaller than that of p, then we delete the control point in the next state N(L). • If the sorting ordering of the phrase of the transition label L is larger than that of p, then the input string becomes the rest of the input string and we move the position
214
B. Liu et al. / Towards Automatic Construction of News Directory Systems
of the control point from the next state N(L) to S . This control point may go to the next state N(L) when the first phrase of the input string becomes L after deletion of first phrases.
5 Experimental Evaluation 5.1 Construction of Directory Structures We constructed a sample news directory system having country/region directories, disease directories, natural disaster directories, energy resource directories, and sport directories. For our future use we also constructed a small classification tree of 885 nodes with 624 basic words from Longman defining vocabulary and 261 non-basic words from WordNet. Parts of WordNet tree and our constructed small tree near ”animal” are shown in Fig. 5 and 6. This small tree may serve us as a small classification tree of news articles.
Figure 5: A WordNet tree near ”animal”
Figure 6: A small classification tree near ”animal”
5.2 Automatic Placement We automatically classified 5,657 news articles collected from June 2006 to July 2006 from 21 news sites of 17 countries/regions into a countries/regions directory structure. We manually evaluated the precision rate and recall rate of our automatic placement method using country/region classification of 500 news articles as shown in Table 1. Of automatically classified 500 articles, 453 articles are appropriately placed. 12 articles mentioning country/region names are not classified into any country/region, because our definition of country/region names was primarily based on UN list of country names. 35 articles not mentioning country/region names are classified into countries/regions, because company names, event names, and news source names may contain country/region names. 5.3 Visualization and Analysis Out news directory system has a visualization subsystem so that users can understand the result visually. For example, we can represent the frequency level of co-occurrence of country/region names and particular words such as Bird flu on a world atlas using Google Maps
B. Liu et al. / Towards Automatic Construction of News Directory Systems
215
Table 1: Result of Automatic Classification 500 articles articles classified appropriately inappropriate articles not classified misclassified 453 12 35
as shown in Fig. 7. The co-occurrence of country/region names and Bird flu in a news article does not necessarily mean that Bird Flu epidemic is taking place in that country/region. However, this map shows that some country/region names are more frequently mentioned together with Bird flu than other country/region names.
Figure 7: A visualization map
Based on 5,657 news articles, disease name such as Bird flu is most frequently mentioned in countries/regions such as Indonesia and China. While disease name such as Cancer is most frequently mentioned in countries/regions such as United States and Australia. Table 2 shows the frequency of country/region names with some of human disease names and natural disaster names. Comparisons of our approach with existing approaches are as follows. For the classification of news articles, Bayesian classifications [2] may be used. However, the result of Bayesian classification is not deterministic or not predictable in general. We need to make our system’s behavior predictable. String matching algorithms such as Aho-Corasick algorithm [1] may be used for the classification of articles into directories. However string matching algorithms usually detect the occurrence of, for example, ”pen” in the word ”pencil” of a text. We need to avoid this partial matching in our system.
6 Conclusion We have presented automatic methods for constructing news directory systems. Our news directory systems allow us to search for news involving a number of potential keywords or unknown keywords.
216
B. Liu et al. / Towards Automatic Construction of News Directory Systems
Table 2: Frequency of country/region names together with particular words Category Bird flu Country Code IDN CHN THA VNM USA ESP AUS MYS IND LAO GBR NER HUN ZMB MMR KOR Article Count 72 35 28 14 12 7 7 5 5 4 4 3 3 3 3 2 Category Cancer Country Code USA AUS GBR KOR CHN ITA SGP THA FRA JPN LBN AUT VNM SWE SYR IRL Article Count 33 26 11 6 6 6 5 4 3 3 3 3 3 2 2 2 Category Tsunami Country Code IDN AUS THA JPN USA SGP DEU MYS VNM NLD GBR PHL IND CHN LKA FRA Article Count 147 7 7 4 4 4 3 3 3 2 2 2 2 1 1 1 Category Earthquake Country Code IDN CHN PHL USA JPN PAK SGP IND IRN TUR KWT AUS FRA TON EGY GIN Article Count 142 17 15 15 14 12 8 7 6 5 5 5 2 2 2 1
As our future work, we will extend our approach as follows. Improvement of precision for automatic classification According to the experiment, the rate of appropriate classification is 90%, we could improve it if we make the system recognize proper nouns, and we can also have more precision results if we give more definitions to directories in the system. Automatic definitions Currently, we define every directory in the system manually, it is really a costly and tedious work. But it is also one of the most important steps which will affect the automatic classification directly. If we use the relations between words in WordNet, it will help us in giving directories definitions. Multi-lingual searching We can construct a multi-lingual news directory system containing English, French, Chinese and Japanese news index information with the same classification structures, so that we can reach French news articles with the help of directory structures, and get ideas of the original contents with the help of Google Language Translation Tools from French to English.
References [1] Alfred V. Aho and Margaret J. Corasick, Efficient string matching: an aid to bibliographic search, CACM, 18(6), 333-340, June 1975. [2] Jennifer Hoeting, David Madigan, Adrian Raftery and Chris Volinsky, Bayesian Model Averaging, Statistical Science 14, 382-401, 1999 [3] A. S. Hornby and Michael Ashby, editors. Oxford Advanced Learner’s Dictionary of Current English. Oxford University Press, 2005. [4] Paul Proctor, editor. Longman Dictionary of Contemporary English. Longman, 2005. [5] Wikipedia, http://en.wikipedia.org/wiki/Main Page [6] WordNet, http://wordnet.princeton.edu/
Information Modelling and Knowledge Bases XIX H. Jaakkola et al. (Eds.) IOS Press, 2008 © 2008 The authors and IOS Press. All rights reserved.
217
A System Architecture for the 7C Knowledge Environment Teppo RÄISÄNEN, Harri OINAS-KUKKONEN Department of Information Processing Science, University of Oulu, Finland [email protected], [email protected] Abstract. This paper presents an information system architecture for the 7C model for organizational knowledge creation and management. The architecture is derived from the requirements that the 7C model posits. The architecture presented here comprises three layers: the conceptual layer, which discusses fundamental principles of the model, the technology layer, which tackles potential implementation technologies for the environment, and the application layer, which describes possible applications in the environment.
1 Introduction Knowledge management has received great attention both among practitioners’ and researchers’ literature for a longer period of time (see, e.g. [1][19][22][26][30][41]). More recently, collaborative approaches for managing knowledge have been proposed [44], suggesting that new knowledge is being created in group-efforts among many people instead of a few experts only [44]. This paper approaches knowledge management through a conceptual framework known as the 7C model [37]. This model suggests that knowledge is produced through the interaction of individual and social knowledge, as well as explicit and tacit knowledge. As the 7C model puts special emphasis on the social aspects of the knowledge management we will try to identify and analyze those new technologies that offer support for them. The research approach adopted for this paper is design science [25][20], in which IT artifacts are build and evaluated. March and Smith [25] recognize four types of design science products, namely constructs, models, methods and implementations. This paper describes a construct, namely an overall information system architecture for the 7C model. More specifically, systems development as a research methodology consists of five parts [31]: 1) constructing a conceptual framework, 2) developing system architecture, 3) analyzing and designing the system, 4) building the (prototype) system, and 5) observing and evaluating the system. In line with this definition, the research described in this paper is part of a larger system development research effort. According to March and Smith constructs “form the vocabulary of a domain”, and “they constitute a conceptualization used to describe problems within the domain and to specify their solutions” [25]. The 7C conceptual framework has been originally described in [37]. The contribution of this paper lies in the system architecture, which together with the conceptual framework, may be regarded as a whole construct [25]. Later, following the framework presented here the 7C knowledge environment will be implemented and experimented. Nunamaker et al. [31] define that system architecture is supposed to: 1) define a unique architecture design for extensibility, modularity, etc., and 2) define functionalities of system components as well as interrelationships between them. They also state that careful system requirements definition should be made and that the requirements should be
218
T. Räisänen and H. Oinas-Kukkonen / A System Architecture for the 7C Knowledge Environment
measurable. For this reason, we aim at identifying the requirements for the overall 7C architecture, and then present the architecture using layers, integrating the functionalities and interrelationships of system components within the architecture. The paper is organized as follows. Chapter 2 describes the 7C conceptual framework. Chapter 3 analyses the framework in order to define requirements for the 7C information system architecture. Chapter 4 summarizes the requirements to recognize such concepts that the architecture must implement. Chapter 5 presents possible implementation technologies which are able to meet the concepts. Chapter 6 and 7 discuss example applications and the contribution of the paper. And finally, chapter 8 concludes the paper.
2 The 7C Model in a Nutshell The 7C model [37] for understanding organizational knowledge creation suggests that the following seven Cs play a critical role in the creation of organizational knowledge: Connectivity, Concurrency, Comprehension, Communication, Conceptualization, Collaboration, and Collective intelligence. Technologically, the benefit is realized through the fluent connectivity that the Internet technology provides with information and people for potentially several concurrent users (the 1st and 2nd Cs). The World Wide Web and its hypertext functionality to promote options and allow freedom of choice with contextual support provides users with a rich environment for comprehending (the 3rd C) and communicating (the 4th C) the information they find. Knowledge is conceptualized (the 5th C) as knowledge artefacts, which serve as a collaboration vehicle through interaction between information producers and consumers, within a team of co-workers or among other stakeholders (the 6th C). All of these six preceding Cs contribute to the growth of “togetherness” or collective intelligence (the 7th C) [37]. The creation of organizational knowledge is not a linear process, but rather a multicycle spiral process [37]. See Fig. 1. The framework assumes that connectivity of all stakeholders with the joint information space and people potentially concurrently is provided in a technologically sound manner, e.g. through the Web, Internet, wireless, mobile and other technologies. The 7C model follows Nonaka and Takeuchi [30] in that the integration of individual and social orientations (in their terminology individual and organizational) are emphasized, and that knowledge is assumed to be created through interaction between tacit and explicit knowledge. The model follows Engelbart [13] in the outcomes of the Comprehension, Communication and Conceptualization sub-processes. Individual
Social
Kw transfer Communication Tacit knowledge Comprehension Explicit knowledge
Kw creation
Collective intelligence
Kw creation
Conceptualization
Kw Collaboration application
Figure 1: Organizational knowledge creation [37].
T. Räisänen and H. Oinas-Kukkonen / A System Architecture for the 7C Knowledge Environment
219
The four most central sub-processes in the knowledge creation are [37]: x Comprehension – a process of surveying and interacting with the external environment, integrating the resulting intelligence with other project knowledge on an ongoing basis in order to identify problems, needs and opportunities; embodying explicit knowledge in tacit knowledge, “learning by doing”, re-experiencing. x Communication – a process of sharing experiences between people and thereby creating tacit knowledge in the form of mental models and technical skills; produces dialog records, which emphasize the needs and opportunities, integrating the dialog along with resulting decisions with other project knowledge on an ongoing basis. x Conceptualization – a collective reflection process articulating tacit knowledge to form explicit concepts and rationale and systemizing them into a knowledge system; produces knowledge products of a project team, which form a more or less comprehensive picture of the project in hand and are iteratively and collaboratively developed; may include proposals, specifications, descriptions, work breakdown structures, milestones, timelines, staffing, facility requirements, budgets, etc.; rarely a one-shot effort. x Collaboration – a true team interaction process of using the produced conceptualizations within teamwork and other organizational processes. Each of the sub-processes may also be regarded as the building of an artifact and reasoning why it has been built the way it has, i.e. capturing the knowledge rationale. Repeatedly going through these phases in a seamless and spiral-like way leads into the growth of collective intelligence. Support for capturing deep individual thinking and recording the dialog between team members may help create truly innovative knowledge products. The learning involved in the comprehension and communication processes is closely related to the attitudes of the participants, i.e. whether they understand their weak points in the sense of individual learning styles, for example. In spite of receiving a lot of attention recently among practitioners, relatively little organizational knowledge management research has discussed the evaluation of the suggested solutions [38]. Evaluating may be carried out at the individual, work unit (group, team, or department), or overall organizational levels. The 7C model shares the view of King and Ko [22] to knowledge in that knowledge surpluses data and information, and thus even if it emphasizes knowledge content, it also addresses the link from knowledge back to re-shaping data and information (cf. [41]). The increase of sharing and dissemination of information and the increase in varied interpretations are obvious and, as a matter of fact, by no means the most important measures for the success of knowledge management solutions [38]. The truly important measure is the identification of underlying non-obvious, complex problems and issues. This may help better formulate the problems and issues the organization is facing, has faced or will face. Naturally, means for solving these problems are urgently needed. By emphasizing the identification of the key organizational issues and focusing more clearly on solving these instead of something else, the organization also becomes less dependent on its individuals. At the same time the corporate or collective intelligence grows by the transfer of ideas, experience and best practices, and the individuals become more confident at their daily work [38]. An example of these, in particular Collaboration and Conceptualization, is the role of argumentation or design rationale in systems development (cf. [34]).
220
T. Räisänen and H. Oinas-Kukkonen / A System Architecture for the 7C Knowledge Environment
3 Requirements for the 7C Information System Architecture The purpose of this paper is to develop an information system architecture that follows the 7C conceptual framework. According to Nunamaker et al. [31], system architecture should be designed for extendibility and modularity. This is supported by presenting the architecture in three layers: application layer, technology layer and conceptual layer. Extendibility is supported by separating possible applications and technologies from the key concepts presented in the conceptual layer. The key concepts provide those underlining principles upon which the architecture, and the 7C model, builds upon. Modularity is supported by defining the structure of each layer. As new technologies are developed, they can be included in the technology layer if they support the identified key concepts. First, to recognize the key concepts for the system architecture, we will aim at identifying requirements posed by the 7C model itself.
3.1
Connectivity
The fluent connection provided by Internet technology is the basis of the 7C model. The users must have access to the system whether they are working at home or in the office. For example, language context processes of Communication and Comprehension rely heavily on Internet technology to provide a connection. This connection can be to people (Communication) or to knowledge (Comprehension). The connection to the Internet provides users a space in which they can communicate and interact regardless of time or place. Connection may also be improved through multiple access points in the system (e.g. mobile access) in such a manner that users are able to stay connected even when on the move.
3.2
Concurrency
Concurrency refers to the fact that the system may have several concurrent users, which, in some cases, may be interested to work with the exact same knowledge artifacts. Thus, proper concurrency control must be taken care of. Internet technology provides a good start for the Concurrency. However, Concurrency may be supported to a greater extent through providing another access point to the system. For example, a mobile access to the system for those on the move may enhance their participation for the knowledge creation processes. Providing mobile access should require no client application to be installed for the mobile device (and, in a matter of fact, for the desktop either). The system should be able to be used with any device that has a modern browser.
3.3
Comprehension
Comprehension is a process of “surveying and interacting with the external environment, integrating the resulting intelligence with other (…) knowledge” [37]. It is the process of embodying explicit knowledge into tacit knowledge. Different knowledge artifacts are created and stored for gaining collective intelligence. The user must be able to browse these artifacts and organize them as (s)he sees fit. Through browsing and organizing existing explicit knowledge, the user is able to “identify problems, needs and opportunities”, and thus learn by doing [37]. This interaction should go deeper than just browsing and organizing. The user should be able to ‘play with’ the existing knowledge. For example, the user should be able to integrate and link different pieces of knowledge, to edit or highlight
T. Räisänen and H. Oinas-Kukkonen / A System Architecture for the 7C Knowledge Environment
221
texts and graphics, or to take an audio file and embed it within a video. In any case the interaction should go deeper than just browsing of static Web pages or the generation of dynamic Web pages through user defined queries. Another way to support deeper understanding would be to allow users to see (potentially any kind of) similarities between knowledge artifacts, in particular between different pieces of knowledge. An associative link [6] between two knowledge objects would explain the user that these objects are somehow related or that they have something in common. Providing this information may trigger the user to understand something totally new. Links may also be typed and they may have attributes [6]. Typed links may help users organize information more effectively and, more importantly, “lend context for readers” to boost Comprehension [6]. Guided tours or paths [6] are examples of providing such a context.
3.4
Communication
Communication is the process of “sharing experiences between people and thereby creating tacit knowledge in the form of mental models and technical skills” [37]. Tacit knowledge an individual possesses may be transferred to other individuals or to a group of individuals. While the transfer of codified knowledge (electronic documents or pictures, for example) is easy to support with computerized information systems, supporting the transfer of tacit knowledge is much more difficult. Asynchronous communication must be supported: users are not always online at the same time, but they must still be able to discuss issues through the knowledge support system. In the 7C model controlling concurrency means supporting the co-presence of users in the virtual space. Even though the knowledge workers may be located in different places, they can still be connected to the same work processes. Co-presence may require some support for synchronous communication, in which knowledge transfer may be enhanced through real-time communication. Marwick [26] argues that in text-based chats, people use such a kind of informal dialog that can help the emergence of new tacit knowledge. Another aspect that speaks for text-based communication is the fact that we can relatively easily search, navigate, and visualize previous text-based communications. We can also add structure for text-based conversations: summarize, highlight, link and annotate them [14]. For example, discussion stored in a XML file may include meta-information on it (i.e. metadata such as date, topic, participants etc.), aw well as the actual content of the discussion. Annotating a certain part of the conversation can be done simply by adding a new tag into a specific spot in the file. With structure visualizations these discussions become relatively lively, and annotations and links between discussions may be displayed when needed. For communications stored in video or voice, this becomes much more difficult. Nevertheless, video, voice and pictures are important in tacit knowledge transfer. Tacit knowledge is often deeply rooted in visual and other bodily senses [29]. According to Nonaka [29] tacit knowledge can be acquired without language, through observations, imitations and practice. Tacit knowledge gained through visual observation may be impossible to articulate and transfer without some visual stimuli to trigger and help the transfer process. Thus it should be possible to use different kinds of multimedia objects (video, sound, pictures etc.) to enrich text-based discussions. After all, the things that are communicated are more important than how they are communicated. The 7C model states that Communication is a process of sharing tacit knowledge, particularly experiences. Typically, information communication technologies provide a means for communication, but they also have an effect on what users
222
T. Räisänen and H. Oinas-Kukkonen / A System Architecture for the 7C Knowledge Environment
communicate [40]. According to the 7C model users should be encouraged, or even persuaded, to share their knowledge and experiences with co-workers in organizational settings. For persuasive purposes, information systems can be regarded as tools, social actors, or as media [16]. As a tool, an information system may persuade by making the sharing of knowledge easier. As a social actor, it may reward the user and provide social acceptance. And as a medium, it may provide people vicarious experiences that motivate them to share information. One problem for tacit knowledge sharing and formation is the potential lack of trust among participants [26]. Especially in virtual environments where the lack of past or future association (face-to-face meetings, for example) decreases the potential existence of trust [21]. One solution for building trust online is to create online communities [4]. Virtual environments may help share some experiences. If a past experience was “learned the hard way” (which may have seen an embarrassing or even humiliating personal experience) sharing such a lesson requires not only trust, but personal courage as well. If no past or future connections among participants exist, sharing such experiences might be easier. On the other hand, if the users know each other, there should be a way to share experiences anonymously. Even though the Communication process is probably the easiest C to support, there are still potential problems with it. Understandably, the sharing of tacit knowledge is more complicated than the sharing of explicit knowledge. In a matter of fact, instead of only supporting the sharing of knowledge for other stakeholders, a support environment should also support the acquisition of knowledge by individual users. A critical, social requirement for the environment such as discussed in this paper is to ensure that users end-up sharing their knowledge and experiences. Special emphasis should be put on such knowledge and experiences that other users do not know. Another important requirement is that the communications are stored in a well-defined, text-based format, such as XML or its variants. In this manner, the communications can best support the full 7C knowledge creation cycle, and information may be reused in the Comprehension and Conceptualization phases more easily and to a larger extent than if they were in some other formats, such as audio.
3.5
Conceptualization
Conceptualization is the “collective reflection process articulating tacit knowledge to form explicit concepts and systemizing the concepts into a knowledge system” [37]. It is the process of transforming tacit knowledge into explicit, and it is probably the least researched area of the 7C processes. This may also be why the existing systems and tools offer little support for it. According to Nonaka [29], the first step in transforming tacit knowledge into explicit knowledge is the use of metaphors. Moreover, the use of metaphors “constitutes an important method of creating a network of concepts which can help to generate new knowledge about the future by using existing knowledge” [29]. It is a creative, cognitive process which relates concepts that are far apart in an individual’s memory. When two concepts are presented in a metaphor, “it is possible to (…) make comparisons that discern the degree of imbalance, contradiction or inconsistency involved in their association” [29]. Nonaka also states that contradictions incorporated in metaphors may be harmonized through the use of analogies. Association of meaning by metaphors is mostly driven by intuition and involves images, whereas association of meaning through analogy is more structural and functional, and is carried out through rational thinking.
T. Räisänen and H. Oinas-Kukkonen / A System Architecture for the 7C Knowledge Environment
223
Conceptualization is a collective process and it requires some sort of consensus about the explicit concept being formed and systemized. This might mean that people have different opinions and ideas about the concept at hand. In that case, reaching a consensus (or compromise if the ideas are too far apart) might need a strong argumentation. If we are to get others to accept a radical idea (or at least to accept the existence of differing opinions) we must show why they should do so. Capturing design rationale in systems development may be used to accomplish just this. Design rationale means the understanding of why an artifact has been designed the way is has [34]. Capturing the rationale behind explicit concepts may lead to “clarity of thinking and augmentation of (…designer’s…) memory” as well as to better communication [34]. With argumentation, we may try to understand the specific elements of each others’ concepts, and perhaps even try to persuade others into accepting our viewpoints, or in other words to conceptualize “knowledge rationale”. If we can argue the explicit knowledge created in the Conceptualization process, we then have a chance of understanding the tacit knowledge behind it. In this way the arguments behind the knowledge help us in Comprehension, Communication and Conceptualization, making knowledge rationale one of the key concepts of the 7C architecture. The outputs of the Conceptualization process are the explicit concepts (basically this can be any explicit knowledge object) backed up with rationale arguing (against or for) the concepts. Visualizing and linking these concepts to each other may help in Comprehension and Collaboration. In the 7C model Conceptualization is a collective process, and the use of metaphors and analogies could facilitate the formation of explicit concepts. Visualization of metaphors and linking them through analogies may provide a way for new concepts to emerge. By utilizing knowledge rationale one may help others to understand his/her reasoning, thus helping Comprehension, Communication, and Conceptualization (and indirectly also Collaboration).
3.6
Collaboration
7C Collaboration process is a “true team interaction process of using (…) conceptualizations within teamwork” [37]. As discussed in the Communication process, a shared virtual environment must be provided for the team to work in. The most important aspect of Collaboration process is to support the coordination and distribution of work. Users should be able to know who is doing what and with whom. An essential aspect for the Collaboration process is that it must provide ways to utilize the produced conceptualizations. Thus, the users should be able to decide who works with whom and with what conceptualization. The actual outcomes of the Collaboration may vary depending on the job at hand but the shared virtual environment provides a good starting point for teamwork. Browsing previous cases, e.g. conceptualizations in use, and reusing the work already accomplished should also be possible.
3.7
Collective intelligence
Going through the Conceptualization, Communication, Conceptualization and Collaboration phases several times in a seamless spiral-like way leads into the growth of Collective intelligence [37]. While organizations create new knowledge, they also forget it [1][3][11]. That is why the storage, organization, and retrieval of organizational knowledge are important [42].
224
T. Räisänen and H. Oinas-Kukkonen / A System Architecture for the 7C Knowledge Environment
In the 7C architecture, it is important that all the knowledge artifacts created in any subprocess are stored. These knowledge artifacts can be anything between discussions in the Communication process and metaphors in the Conceptualization process. Equally important is that the stored knowledge artifacts can be retrieved whenever needed.
4 From the requirements to the architecture – key concepts In the discussion above, 22 requirements for the 7C architecture were identified. The requirements are summarized in Table 1. With these requirements we aim at capturing the essence of the 7C model and identifying the key concepts underlying the C’s. Requirements R1 and R3 state that the 7C architecture must be designed as a Web information system that also takes into count possible mobile users. In this way the architecture can provide the best possible support for Connectivity and Concurrency. Without it much of the potential of 7C may be lost. Multiple users working on the same knowledge artifact requires concurrency control (requirement R2). Comprehension requires that the users must be able to interact (browse, search, read, requirements R4-R6) with existing knowledge artifacts and their metadata in order to comprehend or learn from them. This is essential for new tacit knowledge to emerge as merely providing static information is not enough to truly support Comprehension. Table 1. Requirements for the 7C information system architecture. 7 C’s Connection Concurrency
Comprehension
Communication
Conceptualization
Collaboration
Collective Intelligence
Requirements R1: must be designed as a Web information system R2: must provide concurrency control for managing simultaneous users working with the same knowledge artifacts R3: should be designed mobile aware R4: must provide a way to interact, browse and search the knowledge artifacts and metadata concerning the knowledge artifacts R5: must provide a way to reorganize stored knowledge artifacts R6: should provide a way to interact with the knowledge rationale R7: must enable the sharing of knowledge and experiences R8: must support asynchronous text-based communication R9: should support synchronous communication R10: should support user communities and increase of trust among users R11: should be able to share experiences anonymously R12: must support the definition of knowledge concepts R13: must support the capture of rationale behind the explicit concepts R14: should support the use of metaphors to recognize contradictions R15: should support the use of analogies to resolve the contradictions R16: should support the visualization of concepts R17: should support the linking of concepts R18: must provide a shared virtual working environment R19: must support the coordination and distribution of work R20: should support the use of visual conceptualizations R21: must store all knowledge artifacts created in any 7C process R22: must provide a way to retrieve stored knowledge artifacts
Communication requires that the users can share their experiences or tacit knowledge (R7). Without it no transfer of knowledge will take place. The feeling of community could be used to further enhance this (R10). Much of the knowledge transfer should be text-based so that the previous communications may be easily stored, visualized and searched (R8). To further increase tacit knowledge transfer synchronous communication may be used (R9). Sharing of past experiences “learned the hard way” could be facilitated by allowing users to do it anonymously (R11).
T. Räisänen and H. Oinas-Kukkonen / A System Architecture for the 7C Knowledge Environment
225
Conceptualization means definition of explicit concepts (shift from tacit to explicit). The architecture should support this by defining the knowledge concepts (R12). Resolving between differing opinions or ideas about concepts becomes important (R13). Argumentation behind explicit knowledge is also vital for Comprehension (reading the argumentation might help the reader to understand the tacit knowledge behind the argumentation). Conceptualization might also be enhanced with the use of metaphors and analogies (R14, R15). All this might be facilitated by allowing the visualization and linking of concepts (R16, R17). Collaboration requires a shared working environment (R18). Without it doing any collaborative work is impossible. Collaborative work also requires coordination and distribution of work tasks (R19), so that work is efficient and users know what they should be doing. Also the users should be able to collaborate by using the conceptualizations in their work (R20). Finally, the processes produce knowledge artifacts that must be stored and retrieved as needed (R21, R22). Without the ability to store and retrieve the knowledge, there would be no organizational memory and the knowledge created would be quickly lost. From the requirements, we can recognize key concepts for the 7C architecture. The first is the knowledge rationale. As the 7C is a model for understanding organizational knowledge creation, knowledge and how it is represented is essential. Knowledge rationale means backing up the explicit knowledge objects with solid argumentation. The second is the use of hypertext functionality, i.e. features such as linking, and metadata. The third is the concept of mobile aware Web information system which supports the Concurrent Connection required by the 7C model. Because the 7C is a model for organizational knowledge creation and management, knowledge and how it is represented are crucial for it. This paper proposes that the rationale behind knowledge, i.e. knowledge rationale, should be treated equally important to the knowledge itself. This means that any produced concept of knowledge is stored with argumentation for it. This helps in many ways. For example, if another similar knowledge concept is being produced existing argumentation may be checked to understand why a certain knowledge concept is defined the way it is, or argumentation that has been found valid in one case may be found valid in the other case, too. It may also be possible to find knowledge traces in these argumentations, and this rationale might even include some of the tacit knowledge associated with the task at hand. This might help managing the organizational memory also. For example, the piece of explicit knowledge could be an important decision, e.g. whether or not a company should expand to new markets, based on a collection of facts, e.g. an analysis by consultants. If the question at hand is argued for and against, the ultimate decision will be easier to make. Often this argumentation holds much of the knowledge, and it is imperative for the organization that it is stored with the knowledge as it may be more important to trace the arguments than to know the exact decision. In the 7C model, knowledge rationale is embedded in Comprehension, Communication, Conceptualization, and Collaboration sub-processes. Each of these may produce new artifacts and new knowledge. For example, in Conceptualization, the produced concepts can be seen as explicit knowledge in the form of proposals, specifications, descriptions, work breakdown structure, etc., and the rationale behind the knowledge. The knowledge rationale is in the very heart of 7C architecture, and all of the processes deal with it in one way or the other. Knowledge rationale can be seen as an addition to Conversational Knowledge Management (CKM) [10][44]. In CKM knowledge is created and shared through questions and answers. This is typically done through email lists, discussion forums, or similar. Knowledge rationale adds the element of argumentation
226
T. Räisänen and H. Oinas-Kukkonen / A System Architecture for the 7C Knowledge Environment
to CKM. In Conceptualization the question-answer pair would not capture all relevant knowledge. While it is relatively easy to capture explicit knowledge in question-answer pairs, capturing tacit knowledge is more difficult. In knowledge rationale one question can have many answers and each answer can have arguments against or for made by different people [34]. In this way, conversations carried out in the Conceptualization become dynamic and natural, and the arguments may embed tacit knowledge regarding the question-answer pair at hand. CKM can also been as a way to transfer existing explicit knowledge to others, i.e. mainly the responder transferring his/her knowledge to the individual asking the question (and to others who read the questions and answers). In knowledge rationale, there is a better chance for new knowledge to emerge. New knowledge might emerge in the dialog between the arguments for and against as the users would have to come up with better arguments to counter other people's arguments. The same thing could also happen in CKM but knowledge rationale persuades users to do this through argumentation Besides linking and metadata discussed earlier the interaction capabilities provided by hypertext functionality are also important for the 7C model. They provide the means for "survaying and interacting with the external environment, integrating (...) intelligence (...), identify problems, needs and opportunities" [37]. Without the ability to interact with knowledge objects we loose some of the ability to "learn by doing" and re-experiencing [37]. As such the hypertext functionality is very important for the Comprehension. To allow the users to truly interact with the existing knowledge, hypertext must be provided in a richer way than with static Web pages or even with dynamic Web pages (i.e. Web pages are created according to the users actions). The users should be able to edit, comment, link and create the Web pages as they see fit. With this kind of functionality we may even further facilitate the Comprehension. Hypertext functionality is also useful for Conceptualization and Collaboration, too. We can use linking and annotation to help the use of metaphors, for example. As another example structure-based query can support knowledge rationale. As knowledge is saved with its reasoning, knowledge-based search is not enough: there also has to be the capability to investigate the rationale. Annotations [6] attached to knowledge can be used as the rationale. In Collaboration we can interact with the produced concepts to perform the work at hand and use them within teamwork [37]. The Concurrent Connection is realized through the concept of mobile aware Web information system [35]. A mobile aware Web information system (MAWIS) is a Web information system that has been designed with its potential usage through wireless interfaces in mind. Wireless interface refers to different mobile devices such as PDA’s, mobile phones, etc. With the concept of MAWIS, we can improve the connectivity as well as the number of concurrent users. In doing this, the separation of content from its presentation becomes essential.
Knowledge Rationale Explicit knowledge objects, rationale behind the objects
Hypertext functionality Linking, metadata, hypertext
Mobile aware Web information system Concurrent Connection to Information and People
Figure 2. Key concepts – conceptual layer of the 7C architecture.
T. Räisänen and H. Oinas-Kukkonen / A System Architecture for the 7C Knowledge Environment
227
To summarize, the key concepts of knowledge rationale, hypertext functionality and mobile aware Web information systems form the conceptual layer of the 7C architecture. It is shown in Figure 2. Knowledge rationale is perhaps the most important concept in the 7C system architecture. According to the 7C model, the outputs of Conceptualization are explicit concepts. In our architecture the explicit concepts consist of explicit knowledge and arguments behind them. The Connectivity and Concurrency suggest that the system should be designed as a mobile aware Web information system to increase the concurrent connection to information and to people. And lastly, the hypertext functionality serves as a basis for all 7C processes and it allows the use of linking and metadata, and user interaction with knowledge objects stored in the system thus helping Comprehension and Conceptualization.
5 Implementation considerations All key concepts of the architecture imply some specific technological needs for the implementation. Some technologies meet these needs better that others. For example, one core competency of Web 2.0 is to “harness the power of collective intelligence” [33]. This will go hand in hand with the 7C model. On the other hand, some other technologies seem to emphasize aspects that are not so suitable for the 7C model. We will first go through technologies that support hypertext functionality, in particular Web 2.0 technologies, as they will work also with the other concepts. Then we’ll look at the technologies that support the knowledge rationale, followed by technologies that support the mobile aware Web information systems. Finally, other possible technologies that might fit the overall 7C framework will be discussed.
5.1
Technologies supporting hypertext functionality
Web 2.0 [33] refers to a perceived or proposed second generation of Internet-based services, such as social networking sites, wikis, communication tools, and tagging, that emphasize online collaboration and sharing of knowledge between users. Web 2.0 is not a technical standard but rather a buzzword for innovative applications that are made possible by the ever growing number of Internet technologies and the novel use of combining existing technologies. Some characteristics for Web 2.0 have been defined [33]: 1) Web as platform, 2) Architecture of participation, 3) Rich user experience, and 4) Social networking The Web as a platform allows applications to be delivered and used through a Web browser. There is no need for software releases, licensing or porting to different operating systems [33]. For example, people can use www.google.com with just about any device that has a Web browser and they need no software updates or separate payments. In the Web 2.0 the importance and usefulness of a service is emphasized. This is mainly because the business value comes from delivering services over the Web platform [33]. A typical service could be a search engine or an on-line auction site. New Web services are also emerging in the form of mashups: a combination of existing Web services to form a new value-added service, e.g. combining Google Maps1 with apartment rental and home purchase services to create an interactive housing search tool [33]. Web as a platform improves the Concurrent Connection: users can run the service any time, anywhere without the need of client software. 1
http://maps.google.com
228
T. Räisänen and H. Oinas-Kukkonen / A System Architecture for the 7C Knowledge Environment
Architecture of participation refers to the success of Web sites that promote user participation. For example, Flickr2 not only stores your photos, but it allows you to share them with others. Weblogs and Wikis also provide an example of participation. Weblogs, or “blogs” are frequently updated Web pages with a series of archived posts, typically in a reverse-chronological order [28]. They are primarily textual, but often they also contain photos or other multimedia content. They also may include hypertext links to other Internet sites (often to other blogs). While personal homepages and Web publishing are nothing new as such, it is the user participation that gives weblogs an edge: the audience can read the blog, but they can also comment them. Blog entries, their comments and comments-oncomments enable better participation. It is interesting to note that the thing that made blogging truly participatory was not just the ability to comment on other's texts but the introduction of two types of links, namely the permalink and traceback [8]. The permalinks gave each blog entry a permanent location at which it could be referenced and this allowed a blogger (the writer of a blog) to cite exact blog entries. A traceback allows a blogger to ping other weblogs by placing a reciprocal link in the entry they have just referenced [8]. Together permalink and traceback allowed weblogs to become parcipatory: a blogger would know when other blogger would cite and comment his texts and he could write a reply. Participation is very important for Communication, Conceptualization and Collaboration in 7C. Another important Web 2.0 technology that supports participation and is important in knowledge management (see [44][39]) is a Wiki. Wikis are collaborative tools that enable groups to jointly create content [43], and they differ from plain discussion forums in collaborative aspects. In Wikis, users can edit any knowledge stored in it, not just their own writings as in discussion forums. Leuf and Cunningham [23] define Wiki as "a set of linked Web pages, created through the incremental development by a group of collaborating users". Wikis are found to be a good way to support question-answer pairs of CKM [44][10] and thus should also support knowledge rationale. The collaborative nature of Wikis allows Web documents to be authored collectively, which fits very well with the 7C model. There are many Wiki software systems available as open source. These Wiki software systems differ from each other mainly in their special features. Some useful features could be voting, workflow management and file and image galleria [43]. Wikis would also take care of concurrency and versioning issues to avoid conflict or inconsistencies arising from multi-user capabilities [43]. For the 7C model, Wikis could be used as a platform for Collaboration and Comprehension, vehicle for Communication, argumentation and Conceptualization and as a knowledge repository for all the knowledge created in 7C processes. As such, Wikis seem to provide a natural way to implement 7C tools. The term “Rich user experience” as well as “Rich Internet Applications” (see [2]) refers to fact that Web-based applications are starting to offer GUI-style application experiences to users [33]. An example of such user experience is Google Maps. Typically, in map-based Internet application a user has to click on a hyperlink to scroll the map. In Google Maps the user can click on the map and scroll it using the mouse, i.e. in a similar fashion as he would do on a desktop application. Google uses AJAX (Asynchronous JavaScript with XML) [17][27] and this collection of technologies has become one of the key components of the Web2.0 applications [33]. AJAX incorporates “standards-based presentation using XHTML and CSS, dynamic display and interaction using the Document Object Model, data interchange and manipulation using XML and XLST, asynchronous data retrieval using XMLHttpRequest, and JavaScript” [17]. While none of these 2
http://www.flickr.com
T. Räisänen and H. Oinas-Kukkonen / A System Architecture for the 7C Knowledge Environment
229
technologies are new in themselves, it is the novel use of them together that supports the provision of a rich user experience in Web 2.0 applications. For the purposes in the 7C model, rich user experience may facilitate the visual representation of concepts and knowledge objects. This might have a positive effect on Comprehension as the user experience would not hinder the work. The same applies to some degree to Collaboration, too, as users would apply the produced concepts in their work. With the Web 2.0, social networking has also found its way into Web applications. Typically, social networking sites allow users to create and maintain a network of close friends or business associates for social and/or professional reasons. An example of such a Web site is LinkedIn3. It allows members to look for jobs, seeking out experts or to make contacts with other professionals through a chain of trusted connections [32]. For the 7C purpose, social networking could be used to seeking out expertise (as in LinkedIn). Attaching metadata in the form of keywords (called tags) to content is a common way of organizing content for future navigation, filtering or search [18]. With Web 2.0, collaborative form of this process called tagging or folksonomy has gained popularity [18][33]. In tagging, people not only tag information for themselves but for others, too. This works best when there is no authority to control the tagging and people can use tags as they see fit [18]: Somebody might tag a video about a man breaking his arm as “man” and “funny” while another user can tag the same video with “accident”, or a photo of a puppy could be tagged “puppy” and “cute” and the photo could be retrieved using either tag correspondingly. This allows multiple and overlapping associative linking [7] imitating the human brain rather than a formal categorization [33]. Tagging can help the user in Comprehension because (s)he can browse, search and categorize explicit knowledge objects (s)he (and others) tagged, and in Communication because he can see how others have tagged knowledge and because (s)he can share his tags with others.
5.2
Technologies supporting knowledge rationale
Semantic Web is a project which tries to facilitate information exchange by bringing structure to the meaningful content of Web pages [5]. This is done by putting documents with computer-processable meaning (semantics). Semantic Web is not a separate Web but an extension of the current [5]. In Semantic Web XML (eXtensible Markup Language) and RDF (Resource Description Framework) are used to describe the structure (XML) and meaning (RDF) of the information. Ontologies are collections of information that define relations among terms [5] and they are created with OWL Web Ontology Language. Together, these techniques form the basis of the Semantic Web. According to Berners-Lee et al. [5], "the real power of the Semantic Web will be realized when people create many programs that collect Web content from diverse sources, process the information and exchange the results with other programs". These programs are called agents and their “effectiveness (…) will increase exponentially as more machine-readable Web content and automated services become available” [5]. Semantic Web as such seems to give more power to the computers, e.g. putting documents into computer-processable form for software agents, whereas Web 2.0 relies on users working collectively, e.g. through tagging and social networks. Since knowledge creation is a collective and social process Web 2.0 technologies seem to be more important for the knowledge management purposes than those of the Semantic Web. That is also why we do not represent knowledge artifacts through Semantic Web ontologies but rather through argumentation in conjunction with Web 2.0 technologies. Knowledge (be it in any 3
http://www.linkedin.com
230
T. Räisänen and H. Oinas-Kukkonen / A System Architecture for the 7C Knowledge Environment
computerized form – text, pictures, audio, video) is stored with reasoning concerning the knowledge. Typically, argumentation (or design rationale) means the understanding of why an artifact has been designed the way is has [15]. To argue for knowledge, we will use the Question-Answer-aRgument (QAR) method [34] and apply its concepts to knowledge rationale. QAR has been chosen because of its inherent support for hypertext functionality (linking, annotating, hyperdocument structure, etc.) and simplicity. In matter of a fact, it has originally designed in order to simplify the explicit rhetorical structure of rationale capture [34]. One suitable Web 2.0 technology for knowledge rationale is provided by Wikis. Knowledge objects can be argued within Wiki pages using QAR. In this way Wiki users would contribute collectively in the forming of the rationale. One of the most important steps in implementing the 7C Knowledge Environment is the integration of Wiki and QAR: the users must be able to interact with the argumentation stored in QAR and the knowledge stored in the Wiki pages.
5.3
Technologies supporting MAWIS
A mobile aware Web information system is a Web Information System that has been designed with potential usage with wireless interfaces in mind [35]. For successful construction of mobile aware Web information systems, content and presentation (functionality) should be separated from each other [36]. This enables information exchange with other information systems and also makes customization towards wireless devices easier, which further increases support for Connectivity and Concurrency. One way to separate the content and presentation is using XML to define the content and document structure and a stylesheet language to define the presentation [24]. Often Cascading Style Sheet (CSS) or Extensible Stylesheet Language (XSL) is used for presentation. As XML, CSS and XSL are integral parts of AJAX implementing 7C as mobile aware should be rather straightforward.
5.4
Other technologies for implementing 7C
Other solutions besides AJAX have emerged to support rich user experiences. One such solution is Adobe Flex. Typical Flex applications consist of interface elements build with MXML (Macromedia Flex Markup Language) and interactivity designed with ActionScript [9]. With Flex it is possible to create Flash-based applications with features such as chat, real-time dashboards, messaging and data push services [9] that run in a Flash player embedded in the browser. These Flash applications have excelled in recent years in streaming video on-demand [12]. An example of Flash for video streaming is YouTube4. Besides just streaming video Flash also lets users create layered visual effects by combining video with text, vector graphics, and other elements [12]. This could help users to comment certain interesting parts of the video instead of just commenting the whole video. This would imply that a Flash player would be suitable for playing the videos stored in the 7C environment. A challenge for using solutions such as Flex is that they require a plug-in5 to work with. This does not enable the best possible connectivity since all users will not install the plug-ins needed. Also the use of plug-ins in mobile settings is often impossible. Another way to provide richer use experiences is by extending the browser through user interface markup languages [45]. One such markup language is XUL (XML User Interface Language 4 5
http://www.youtube.com A plug-in is a program that interacts with a Web browser to provide a certain function on-demand.
T. Räisänen and H. Oinas-Kukkonen / A System Architecture for the 7C Knowledge Environment
231
[45]). An example of such an extension is to add triple-click functionality to the application (e.g. trible-clicking would select the clicked paragraph of text). The problem with such extensions is that they need browser support, which conflicts with the goals of concurrent connection and mobile use. As the 7C environment will require visualization of concepts, one implementation solution could be to use AJAX technologies to support the business logic and Flex to support the visualization. The simultaneous use of multiple new technologies may cause additional challenges for mobile use as the browsers in mobile devices in most cases do not support the latest technologies. Nevertheless, mobile access should be provided even with limited functionality, because even if mobile devices do not provide a rich user experience they may still help in the communication to a great extent.
6 System applications The key concepts and the technologies recognized implicitly suggest a set of tools which match with the requirements presented in chapters 3 and 4. Since Comprehension and Conceptualization are the processes that have received the least attention in research literature, we aim at putting special emphasis on them. A specific support tool for Comprehension should allow rich interaction with the existing knowledge: the users should be able to browse, search and categorize knowledge and the knowledge rationale stored in the 7C environment. Using personal and shared tags supports Comprehension by providing the kind of associative linking which enables the user to recognize similarities and possibly to identify specific needs and opportunities as well as potential problems. A richer user experience provided through AJAX or Flash may facilitate this interaction even more as the user is able to ‘play with’ the knowledge in a richer way than with the normal interaction capabilities provided by the static Web pages. In fact, the richer the interaction the better the chances probably are for comprehending something new. By tagging a user may share associative links with other users. This may facilitate Comprehension and Communication. For example, navigating through pieces of knowledge that have been tagged in a similar manner forms a path [6]. This may provide context for deeper Comprehension, e.g. through recognizing similarities. In a matter of fact, tags, as well as other ways to support link typing, are in the very heart of both Comprehension and Communication subprocesses, and for this reason the 7C Knowledge Environment should support flexible linking through different types of links. User1 blog Entry 1
User3 blog
Entry 2
Entry 1 Entry 2
Entry n
Concept 1 Knowledge Object1
User2 blog Entry 1
Entry n
Rationale (QAR)
Entry 2
Entry n
Figure 3. Users can link their blog entries to other users’ blog entries and to other objects within the system.
Users also need the capability to write down their own thoughts and ideas about different knowledge objects. This may be done with a tool such as a weblog. Weblog entries should be able to link with anything within the system (see Figure 3). Writing and
232
T. Räisänen and H. Oinas-Kukkonen / A System Architecture for the 7C Knowledge Environment
reading blog entries may facilitate Communication, in particular when users comment other users’ blog entries. Users should be able to modify one’s own blog entries, but they should only be allowed to read and comment other users’ blogs. A tool that would support Conceptualization should enable users to collectively articulate tacit knowledge in order to form explicit concepts. This paper approaches these concepts through the knowledge rationale. Each concept consists of an explicit knowledge object and the rationale behind it (see “Concept 1” in Figure 3). A Conceptualization tool should allow people to edit the explicit knowledge as well as argue for or against the question-answer pairs in the QAR and attach these debates into the knowledge objects at hand. Users should also be able to link concepts and knowledge objects together to show associations between them [6], e.g. concept1 could be linked to concept2 or to knowledge object1, and knowledge object1 could be linked to knowledge object2 or to concept1, etc. It should be also possible to form different concepts from one knowledge object. The same knowledge object may be used in different situations and each situation may require different arguments. Thus, we can create many concepts from one knowledge object (see Figure 4). Concept 1 Rationale1 (QAR)
Knowledge Object1
Rationale2 (QAR)
Concept 2
Figure 4. One knowledge object can belong to many concepts.
Storing and retrieving knowledge is important because without it organizations would not have a memory, and knowledge would be forgotten as soon as it was not used anymore. In the proposed architecture all explicit knowledge artifacts created in any subprocesses must be stored. This includes communications in the Communication process, knowledge rationale and the concepts produced in Conceptualization, and so on. A knowledge repository tool has two main features. It must allow the knowledge to be stored and retrieved, and it should enable removing unnecessary or gratuitous knowledge, when seen fit. The easiest implementation of the knowledge repository tool would probably be to make it a Wiki [43]. A tool support for Collaboration must allow the use of explicit concepts created in Conceptualization as well as the reuse of previous work carried out. The collaboration should be based on a shared virtual environment which would form a basis for the whole toolset. This may be done through a Wiki, where users may work collaboratively with the Concepts produced. The Wiki should also support visualizing the conceptualizations. In addition, a Wiki should handle the organization and distribution of work.
7 Discussion The application layer of the 7C architecture is represented in Figure 5. In its simplest form, 7C environment is a Wiki that consists of users’ blogs and concepts produced as knowledge rationale. Users blog for Communication purposes. To further facilitate real-time communication additional tools, such as VoIP-based tools, may be implemented. Blogs can also support Comprehension as the users may write down their thoughts and ideas. Yet, most of the Comprehension support is provided by browsing, searching and categorizing the concepts. Tagging is a key technology for Comprehension as it enables to define associative links between the concepts. Comprehension is also supported by allowing users to read the rationale behind knowledge objects. Conceptualization is
T. Räisänen and H. Oinas-Kukkonen / A System Architecture for the 7C Knowledge Environment
233
supported through a Wiki, where users collectively debate (argue) over produced knowledge objects using QAR. Wiki also works as a vehicle for Collaboration.
User’s Blogs
Concepts Knowledge Objects
Rationale
Application layer
7C Wiki
QAR
Flex & Flash
Tagging
Technology layer AJAX, XML, CSS, XSL, JavaScript, VoIP
Knowledge Rationale Hypertext Functionality
Conceptual layer
Mobile aware Web information system
Figure 5. 7C Information System architecture.
Integrating Blogs or at least blog-like features into Wikis should be rather straightforward. In a lightweight solution, normal Wiki pages could be used as personal blogs, but to take full advantage of the 7C model, at least permalinks and traceback of the blog functionality must be included in the implementation. The implementation of the 7C Wiki should use the technologies of Web 2.0. Especially tagging is essential, since it allows the kind of associative linking that could help both Comprehension and Conceptualization. QAR is suitably lightweight in supporting rationale related to knowledge objects. Too complicated a method of including rationale to knowledge objects might discourage users, and they might end-up not using the tool. AJAX and the technologies included in it offer the kind of rich user experience that might facilitate Comprehension and Collaboration. XML offers the technology to capture the content of all the knowledge produced in 7C Information System. Different stylesheet languages (CSS and XSL, in particular) provide a way to represent this knowledge in any required form, e.g. mobile device or desktop computer. On the conceptual level of the proposed architecture we have the key concepts that influence both of the layers above it. In a way the key concepts in the conceptual layer are a summary of the whole 7C environment: A mobile aware Web information system providing the needed Concurrent Connection on top of which hypertext functionality provides the means for users producing the knowledge rationale. Table 2 presents the key 7C subprocesses and how they are supported by the 7C Knowledge Environment. Concurrent Connection is provided by designing the system as a mobile aware Web information system (basically a Wiki). For the Comprehension subprocess the user’s interaction with the existing knowledge should be as rich as possible. The rich user experience delivered possibly by AJAX may be a key to success as richer experience may foster Comprehension. (S)he should also be able to use associative linking (tagging) to identify similarities. For the Communication, traceback and permalink features provide better participation and thus help users to communicate as they know better when and how someone comments their texts. Conceptualization is probably the least researched part of knowledge management. We propose that knowledge rationale can be used to better
234
T. Räisänen and H. Oinas-Kukkonen / A System Architecture for the 7C Knowledge Environment
support forming of explicit concepts required by the model. Wiki technology seems to be a natural way to support Collaboration. However, implementing all of the 7C features on top of a Wiki may be challenging. Table 2. Support of the proposed architecture for the subprocesses of the 7C model. 7 Cs Connection Concurrency Comprehension Communication Conceptualization Collaboration
Collective Intelligence
How they are supported The system is designed as a mobile aware Web information system Wiki handles concurrency control. Mobile access improves the chances for concurrent users. The users can interact with the knowledge and arguments stored in the environment, e.g. editing, linking (including tagging), commenting, combining existing knowledge Users can blog to communicate about their experiences and to read other users’ experiences. The users can use QAR to argue for and against a question to define the explicit concepts in a form of knowledge rationale The 7C Wiki can be used as a platform for collaboration where users divide the work among them and use the produced conceptualizations to perform collaborative knowledge work. All the created knowledge is stored in the environment and it can be retrieved whenever needed, e.g. in the Collaboration process
The continuous use of the proposed knowledge environment (in which all of the created knowledge is stored) should improve the efficiency and capabilities of its users, and thus in time also the Collective Intelligence of the organization. The most critical part of the environment is the Knowledge Rationale and how it can capture the concepts created in Conceptualization. 8 Conclusion This paper presented the 7C information systems architecture. The architecture consists of three layers, supporting extendibility and modularity as the role of extendibility and modularity are essential in IS architectures [31]. The conceptual layer composes of the key concepts posited by the 7C model. The technology layer presents possible technologies that could be used to implement the key concepts. And finally, the application layer presents the working applications of the 7C environment. The 7C environment must enable users to communicate with each other (using permalinks and traceback) and to interact with the knowledge stored in it. This interaction should go deeper than just browsing the knowledge: The user should be able ‘play with’ the knowledge. This richer interaction may provide a way for comprehending something new. The Wiki technology nicely supports Collaboration. An example of a 7C Knowledge Environment would be a Wiki that supports knowledge rationale using QAR. As a future work, a toolset following this architecture should be implemented and experimented with. The most crucial parts of the 7C model as well as the proposed architecture are Comprehension and Conceptualization. Special emphasis should be put on implementing and testing those, in particular the capture of knowledge rationale through the QAR method as the conceptualizations are used or they interact with many 7C subprocesses. One potential way to study this is to implement QAR either through wikis or blogs. Another important aspect that needs further investigation is the use of linking and link types in Comprehension and Communication subprocesses, e.g. using tags to recognize similarities or guided tours for sharing experiences.
T. Räisänen and H. Oinas-Kukkonen / A System Architecture for the 7C Knowledge Environment
235
Acknowledgements: We would like to thank Seamus Hickey for his comments on improving the presentation.
References: [1] Alavi, M. & Leidner, D. E. (2001). Review: Knowledge management and Knowledge Management Systems: Conceptual Frameworks and Research Issues. MIS Quarterly, Vol. 25, No. 1, pp. 107-136. [2] Allaire J (2002). Macromedia Flash MX-A next-generation rich client. Technical report, Macromedia, March 2002. [3] Argote, L., Beckman, S. L., & Epple, D. (1990). The Persistence and Transfer of Learning in Industrial Settings. Management Science, Feb 1990, Vol. 36, No. 2, pp. 140-154. [4] Ba, S. (2001). Establishing online thrust through a community responsibility system. Decision Support Systems 31, pp 323-336. [5] Berners-Lee, T., Hendler, J., Lassila, O. The Semantic Web. Scientific American, May 2001, Vol. 284, Issue 5, p3443. [6] Bieber M., Vitali F., Ashman H., Balasubramanian V. & Oinas-Kukkonen H. (1997) Fourth Generation Hypermedia: Some Missing Links for the World Wide Web. International Journal of Human Computer Studies, Vol. 47, No. 1, p. 31-65. [7] Bieber M., Oinas-Kukkonen H. & Balasubramanian V. (1999) Hypertext Functionality. ACM Computing Surveys, Hypertext and Hypermedia Electronic Symposium, Vol. 31 (4es), December 1999. [8] Blood, R. (2004). How Blogging Software Reshapes the Online Community. Communications of the ACM, December 2004, Vol. 47, No. 12. [9] Borck, J. R. (2006). Flex 2.0 Enriches App Dev Experience. InfoWorld; Aug 14, 2006, Vol. 28, No. 33, pg 41-42. [10] Cheung, K. S K., Lee, F. S. L., Ip, R. K. F. & Wagner, C. (2005). The Development of Successful On-Line Communities. International Journal of The Computer, The Internet and Management, Vol. 31, No. 1 (January-April, 2005), pp. 71-89. [11] Darr, E. D., Argote, L. & Epple, D. (1995). The Acquisition, Transfer, and Depreciation of Knowledge in Service Organizations: productivity in Franchises. Management Science, Nov 1995, Vol. 41, No. 11, pp. 1750-1762. [12] Emigh, J. (2006). New Flash player rises in the Web-video market. Computer, Vol. 39, Iss. 2, pg. 14-16. [13] Engelbart, D. (1992). Toward High-Performance Organizations: A Strategic Role for Groupware. In Proceedings of the GroupWare '92 Conference, San Jose, CA, August 3-5, 1992, Morgan Kaufmann Publishers. [14] Erickson, T. & Kellogg, W. A. (2000). Social Translucence: An approach to Design Systems that Support Social Processes. ACM Transactions on Computer-Human Interaction, March 2000, Vol. 7, No 1, pp 59-83. [15] Fischer G., Girgensohn A., Nakakoji K. & Redmiles D. (1992). Supporting Software Designers with Integrated Domain-Oriented Design Environments, IEEE Transactions on Software Engineering, Vol. 18, No. 6, pp. 511-522. [16] Fogg, B. J. (2003). Persuasive Technology: Using Computers to Change What We Think and Do, Morgan Kaufmann Publishers, San Francisco, 2003. [17] Garrett, J. J. Ajax: A New Approach to Web Applications. http://www.adaptivepath.com/publications/ essays/archives/000385.php (visited 1.1.2007). [18] Golder, S., A. & Huberman, B., A. (2005). Usage patterns of collaborative tagging systems. Journal of Information Science, Vol. 32, No. 2, 2006, pp. 198–208. [19] Grant, R. M. (1996). Toward a Knowledge-based Theory of the Firm. Strategic Management Journal, Vol. 17, Winter Special Issue, 1996, pp. 109-122. [20] Hevner, A. R., March, S. T., Park, J. & Ram, S. (2004). Design Science in Information System Research. MIS Quarterly, March 2004, Vol. 28 No. 1, pp. 75-105. [21] Jarvenpaa, S. & Leidner, D. E. (1999). Communication and Trust in Global Virtual Teams. Organization Science, Vol. 10, No. 6, November-December 1999, pp. 791-815. [22] King W. R., and Ko D. (2001) Evaluating Knowledge Management and the Learning Organization: An Information/Knowledge Value Chain Approach. Communications of the AIS, Vol. 5, Article 14, May 2001. [23] Leuf, B. & W. Cunningham (2001). The Wiki Way: Collaboration and Sharing on the Internet. Reading, MA: Addison-Wesley. [24] Lie, H. W., Saarela, J. (1999). Multipurpose Web publishing using HTML, XML, and CSS. Association for Computing Machinery. Communications of the ACM; Oct 1999, Vol. 42, No. 10, pg: 95. [25] March, S. T., & Smith G. F. (1995). Design and Natural Science Research on Information Technology, Decision Support Systems, Vol. 15, pp. 251-266.
236
T. Räisänen and H. Oinas-Kukkonen / A System Architecture for the 7C Knowledge Environment
[26] Marwick, A. D. (2000). Knowledge Management Technologies. IBM Systems Journal, 2001, Vol. 40, No. 4, pp 814830. [27] Mesbah, A. & van Deursen, A. (2007). An Architectural Style for Ajax. Proceedings of the 6th Working IEEE/IFIP Conference on Software Architecture (WICSA'07). IEEE Computer Society, 2007. [28] Nardi, B., A., Schiano D., J. & Gumbrecht, M. (2004). Blogging as social activity, or, would you let 900 million people read your diary? CSCW’04, November 6–10, 2004, Chicago, Illinois, USA. [29] Nonaka, I. (1994). A Dynamic Theory of Organizational Knowledge Creation. Organizational Science, Vol. 5, No. 1, February 1994, pp. 14-37. [30] Nonaka, I. & Takeuchi, H. (1995). The Knowledge-Creating Company — How Japanese Companies Create the Dynamics of Innovation, Oxford University Press. [31] Nunamaker, J.F. Jr., Chen, M. & Purdin, T.D.M. (1991). Systems Development in Information Systems Research, Journal of Management Information Systems, Vol. 7, Issue. 3, pp. 89-106. [32] O'Murchu, I., Breslin, J. G., Decker S. (2004). Online Social and Business Networking Communities. Proceedings of the ECAI 2004 Workshop on Application of Semantic Web Technologies to Web Communities, Valencia, Spain, August 23-27, 2004. [33] O'Reilly, T. (2005). What Is Web 2.0 - Design Patterns and Business Models for the Next Generation of Software. O'Reilly Network. http://www.oreillynet.com/pub/a/oreilly/tim/news/2005/09/30/what-is-web-20.html (retrieved 01/01/2007). [34] Oinas-Kukkonen H. (1998) Evaluating the Usefulness of Design Rationale in CASE. European Journal of Information Systems, September 1998, Vol. 7, No. 3, pp. 185-191. [35] Oinas-Kukkonen, H. (1999) Mobile Electronic Commerce through the Web. Second International Conference on Telecommunications and Electronic Commerce (ICTEC ’99). Nashville, USA, October 6-8, 1999, pp. 69-74. [36] Oinas-Kukkonen H., Alatalo T., Kaasila J., Kivelä H. & Sivunen S. (2001) Requirements for Web Information Systems Engineering Methodologies. In M. Rossi & K. Siau (editors): Information Modelling in the Next Millenium, Idea Group Publishing, 2001. [37] Oinas-Kukkonen H. (2004). The 7C Model for Organizational Knowledge Sharing, Learning and Management. Proceedings of the Fifth European Conference on Organizational Knowledge, Learning and Capabilities (OKLC ' 04), Innsbruck, Austria, April 2-3, 2004. [38] Oinas-Kukkonen H. (2005) Towards Evaluating Knowledge Management through the 7C Model. Proceedings of the European Conference on Information Technology Evaluation, (ECITE ’05), Turku, Finland, September 29-30, 2005. [39] Raman, M., Ryan, T. & Olfman, L. (2005). Designing Knowledge Management Systems for Teaching and Learning with Wiki Technology. Journal of Information Systems Education, Fall 2005, Vol. 16, No. 3, pp. 311. [40] Te’eni, D. (2001). Review: A Cognitive-Affective Model of Organizational Communication for Design IT. MIS Quarterly, Jun 2001, Vol. 25, No. 2, pp 251-312. [41] Tuomi I. (2000) Data is More than Knowledge: Implications of the Reversed Knowledge Hierarchy for Knowledge Management and Organizational Memory, Journal of Management of Information Systems, (16)3, pp. 103-117. [42] Walsh, J. P., & Ungson, G. R. (1991). Organizational Memory. Academy of Management Review, Vol. 16, No. 1, 1991, pp. 57-91. [43] Wagner, C. (2004). Wiki: a Technology for Conversational Knowledge Management and Group Collaboration. Communications of the Association for Information Systems, Vol. 13, 2004, pp. 265-289 [44] Wagner, C. (2006). Breaking the Knowledge Acquisition Bottleneck Through Conversational Knowledge Management. Information Resources Management Journal, Jan-Mar 2006, Vol. 19, No. 1, pp. 70-83. [45] Wusteman, J. (2005). About XML: from Ghostbusters to libraries - the power of XUL. Library Hi Tech. Bradford: 2005. Vol. 23, Issue. 1, pp. 118-129.
Information Modelling and Knowledge Bases XIX H. Jaakkola et al. (Eds.) IOS Press, 2008 © 2008 The authors and IOS Press. All rights reserved.
237
Inquiry Based Learning Environment for Children Marjatta KANGASSALO, Eva TUOMINEN Department of Teacher Education, Early Childhood Education, FIN-33014 University of Tampere, Finland
Abstract. This paper describes development work on children’s science learning environments that utilize an inquiry-based learning approach as well as modern technological possibilities. The emphasis in the paper is on the theoretical and pedagogical starting points of inquiry learning and their application to two multimedia learning environments. The modelling of children’s exploratory learning in these environments is also described and discussed.
1.
Introduction
Thinking skills, as well as individual and collaborative exploratory activities, have ascended to a significant position in the rapidly changing technological environments of today. The opportunities of computer technology in the pre-school and primary school learning environments, integrated into the children’s spontaneous activities, have been studied for the past 17 years in the different research projects by Kangassalo and her research partners (e.g., Kangassalo 1991, 1992, 1997, 1998c; Kangassalo and Kumpulainen 2003; Kangassalo et.al. 2005). The purpose of these different research projects has been to discover pedagogic practices in which the opportunities provided by the new technology would support pedagogical activities in a natural and justified manner. In this article, we describe developmental work where the principles and theoretical starting points of inquiry and exploratory learning have been applied to modern technological possibilities for children’s science learning environments. Inquiry-based technological learning environments open opportunities for both the individual and collaborative knowledge construction process. It builds on learning where the learners’ own earlier knowledge and deeper understanding of the phenomena can be achieved together with the development of the learners’ learning and exploration skills. In the article, the theoretical and pedagogical starting points of inquiry and exploratory learning are described along with the developmental work and main research results of the PICCO research program (e.g., Kangassalo 1992, 1997, 2001) and the pedagogical scaffolding approach and examples of the Proagents learning environment (e.g., Kangassalo et al. 2005). At the end of the article, the ongoing research work of the PICCO research program, in which the entire modeling and description of the children’s exploration processes will be the main aim, is presented and discussed (e.g., Kangassalo and Kumpulainen 2004).
238
2.
M. Kangassalo and E. Tuominen / Inquiry Based Learning Environment for Children
Theoretical and Pedagogical Background
2.1. Exploratory Learning Exploring, wondering, and asking questions are natural ways for small children to acquire knowledge of the objects and phenomena surrounding them. Children's exploration utilizes all of the senses, and, gradually, also encompasses spoken language. A small child explores her environment in every way possible. She feels, touches, throws, chews, and follows with her gaze. This multimodal exploration produces information about the object using various senses. Multisensory information builds a rich conceptual basis for understanding the objects, relations, and phenomena in the physical world. Interaction between the child and the physical environment gradually extends into interaction with other children. Exploration takes place alone, together with other children, and with adults. The progress and the phases of exploratory learning may vary considerably depending on the age of the children, whether there are children of the same age in the small group, whether exploratory learning is carried out within a strategy or curricula, the size of the groups, and the number of adults guiding the action.
2.2. Characteristics of Exploratory Learning The objectives of exploratory learning can be roughly divided into two types: 1) objectives that aim at learning the phenomenon or object in question, and 2) objectives that focus on the exploratory action per se. From the viewpoint of the phenomena to be studied, the essential objective in learning is to understand the phenomena in question. Understanding, in turn, includes comprehending causal relations and explaining and foreseeing the changes taking place in the phenomena. Reaching the deeper level of learning and understanding requires grasping the theory behind the phenomena. Thus, the objective of achieving deeper understanding is grasping the key concepts of the phenomenon and their mutual relations on the abstract level where it is possible to explain individual events and phenomena within the framework of the generated conceptual and theoretical knowledge (e.g., Ausubel 1965, 1968; Hakkarainen et. al. 2004; Kangassalo 1997.) The structuring of knowledge and conceptual constructions is a step-by-step process. In the first phase of learning, it is essential that the student recognizes and becomes aware of her previous conceptions and knowledge of the object to be studied. Against the background of one's previous knowledge, it is possible to recognize the gaps and deficiencies in one's knowledge and one's explanations of the phenomenon. These gaps, in turn, guide the learner in the acquisition of new knowledge and thus in the gradual perception of the big picture, one piece at a time. In the different phases of knowledge construction, we often run into situations in which we find the inconsistencies and contradictions of our own explanations and conceptions in relation to the views presented in the material used to support the learning. Inconsistencies are also detected in relation to the conceptions of the other students. In these situations, the child faces the need for a conceptual change, either in a very radical form or as a partial change in her own knowledge and conceptual structures. It is possible to reach conceptual change in students' conceptual constructions of the phenomenon via multimodal learning and teaching practices and the students' own active exploration. In teaching and guiding the learning, it is essential to guide the students towards seeking solutions and answers for why-questions as well as for how and what-questions. Exploring the phenomena begins, particularly with small children, with the observation of phenomena and objects, and it aims at making sense of what exactly happens in the phenomenon, what changes, how and in what kind of circumstances the change takes place. After this modeling phase, explaining
M. Kangassalo and E. Tuominen / Inquiry Based Learning Environment for Children
239
and foreseeing the phenomenon become gradually possible (e.g., Vosniadou 1999; Vosniadou et. al. 2001.). The objectives related to the research of exploratory learning have to do with learning research skills, developing research strategies, learning the skills of exploring together, and developing metacognitive skills, which means that the pupils learn to recognize and analyze their ways of learning and thinking, discovering alternative opportunities for research and learning available, and becoming aware of what they have previously learned on the objects and phenomena to be studied. Along with the development of metacognitive skills, the children also learn to pose questions themselves, which helps them proceed with learning by evaluating the gaps and deficiencies in their own knowledge. Guidance and support from the teacher are very important in all the phases of exploratory learning and in actions aimed at enhancing the deeper understanding of both the students' exploratory operations and the phenomena themselves. Exploratory learning combines the learning of the phenomena and objects and the developing of research operations together with the teacher and the other students in a natural way. At its best, exploratory learning is a research process that generates new understanding and new information about the object to be collectively studied.
2.3. Supporting Explorations in Classrooms Knowledge restructuring and comprehension activity that aims at understanding more deeply the phenomenon in question are deliberate processes, and in many cases some cognitive and sociocultural support is necessary. In schools, teachers play an important role in motivating and supporting students to engage in continuous efforts to seek understanding and to revise their prior knowledge. In other words, to amplify students’ motivation “a teacher has to create and maintain a sociocultural environment that favors comprehension activity” (Hatano & Inagaki 2003, 409). In exploration-based science instruction, one important condition for comprehension activity to occur is that greater emphasis is placed on students’ thinking processes rather than on the need for correct answers, and that enough time is given for the exploration of key concepts in one subject matter area (Vosniadou et al. 2001). It has also been considered important that students are provided with opportunities to work with phenomena instead of only watching teacher demonstrations. Some cognitive scaffolding should be available to help the students find new and alternative ideas (e.g., Andre & Windschitl 2003; Hatano & Inagaki 2003). In this connection, fostering students’ metacognitive awareness is essential, since students themselves aren’t often aware how the previous beliefs constrain their learning (Vosniadou et al. 2001). The enhancement of metaconceptual awareness is possible, for example, by providing students with opportunities to create verbal expressions of their ideas, and by guiding them to elaborate their explanations and previous knowledge with regard to the phenomena in question. It has also been considered essential that the order of the acquisition of concepts in a given subject matter area receive attention. The teacher can take this into account when scaffolding the explorations. Meaningful and theoretically relevant experiences as well as providing models and external representations are important in clarifying scientific explanations. (Vosniadou et al. 2001.) Models and external representations offer students opportunities to explore aspects of phenomena in other forms than the purely linguistic and they can facilitate the comprehension of complicated phenomena by providing visual presentations of interrelations in phenomena.
240
M. Kangassalo and E. Tuominen / Inquiry Based Learning Environment for Children
2.4. Children’s Conceptual Learning and Computer Simulations In the past few decades, cognitive research on science education has focused on the acquisition of science concepts and conceptual change. Several studies have shown that children form intuitive conceptions and explanations about phenomena based on their everyday experience. These conceptions are often very different from the current scientific knowledge and, in addition, can be very resistant to change. (E.g., Vosniadou1991, 1999.) In recent years, research has progressed from describing children’s initial conceptions to the analysis of the processes that instructional interventions can bring (Caravita 2001). Attention has been drawn to environmental factors. The research findings of instructional interventions have shown that “concepts are embedded in rich situational contexts, in the tools and artefacts of the culture, and in the nature of symbolic systems used during cognitive performance”. (Vosniadou 1999, 9.) The recent research on learning environments and instructional approaches that could facilitate conceptual change has emphasized variables such as the role of metaconceptual awareness and students’ preconceptions, social interaction among students, student self-regulation and autonomy, exploratory activities, the meaningfulness of educational tasks and the use of external representations (e.g., Caravita 2001, Diakidoy & Kendeou 2001; Kangassalo 1997; Vosniadou, Ioannides, Dimitrakopoulou & Papademetriou 2001). Computer simulations can provide children with an exploratory learning environment (e.g., Kangassalo 1991, 1998d). They have also been considered one way of addressing children’s intuitive conceptions and teaching for conceptual change. This is true especially when simulations allow learners to perceive what usually can’t be directly observed and provide visual representations for a set of interrelated concepts (e.g., Snirr, Smith & Grosslight 1995). Earlier research findings on children’s exploratory learning and the development of their conceptual thinking in a simulation environment have been encouraging. For example, the findings from Kangassalo’s (1996, 1997, 1998c) research indicate that children’s independent exploration with a computer simulation, at the stage when they are spontaneously interested in the phenomena in question, can facilitate knowledge construction in the direction of currently accepted scientific knowledge. According to the research findings, children’s exploration process in the simulation contained, for example, “wandering here and there, investigating and seeking for something and experimenting with aim.” (Kangassalo 1994, 296.) In examining the children’s exploration process in relation to their conceptual model, it seemed that the more developed the child’s conceptual model was, the more there was experimentation and investigation with a purpose (Kangassalo 1994, 1997).
2.5. Pictorial Computer Based Simulation PICCO as an Exploration Tool Kangassalo (1997, 1998c) has examined how a computer based multimedia simulation program could support children’s conceptual development in astronomy. The program children used in the research was PICCO (Pictorial Computer Based Simulation Program) (Kangassalo 1991, 1998d). The use of the program was designed so that children could use it based on their own interests, questions and their spontaneous and independent exploration. The program has been used in research experiments where children explored the phenomena by following their own interests without adult supervision and at school, which supported normal everyday school learning situations. (E.g., Kangassalo 1997, 1998c.) In the PICCO program, the selected natural phenomenon was the variations of sunlight and the heat of the sun as experienced on earth in relation to the positions of the earth and sun in space. On the earth level, children can explore their natural surroundings, its phenomena and events (such as day and night and seasons) in a natural and realistic way. On the space level,
M. Kangassalo and E. Tuominen / Inquiry Based Learning Environment for Children
241
these phenomena are represented with the help of an analogue model. (Kangassalo 1997, 1998a.) All events and necessary elements in the simulation are represented as pictures and familiar symbols. The program doesn’t include pathways or rules on how the children should proceed. Children can explore the phenomena according to their own interests, either alone or with friends. (Kangassalo 1997.) In Kangassalo’s research (1997, 1998c), there were thirty-three children aged between six to eight years. Eleven of these children had the simulation available to use in a day care center after school for a four-week period. The children could use the program spontaneously and independently. The children did not have any formal instruction about these phenomena either before or during the research period. Twenty-two children used the program at school while simultaneously having a teaching period concerning astronomical phenomena. Before and after the use of the simulation all of the children’s conceptual models were elicited. In the elicitation of a conceptual model, attention was paid to the order, continuity and regularity of events of the natural phenomenon on the earth, the interconnections of the earth and sun in space, as well as the interrelations of phenomena on the earth and in space. In addition attention was paid to the size, form and distance between the earth and the sun. The eliciting was achieved using procedures where children modeled the phenomenon through action, pictorially and verbally. (Kangassalo 1997.) Children’s conceptual models were at very different levels before the use of the simulation. Some children’s models were quite well developed, while others’ were still rather undeveloped. Only a few children’s models contained misconceptions. When the children used the computer-simulation, some changes occurred in their models. The most fundamental change that occurred was that the interconnections of different aspects and phenomena began to be constructed. The changes seemed to occur largely through the progression of different phases in the direction of the currently accepted scientific view. The extent of construction varied in children’s conceptual models. When comparing the results between the two groups – between children who didn’t receive any teaching and children who had a teaching period during the use of the program – some very interesting differences were discovered. The children who received teaching had more difficulties in the integration of the succession of seasons and the alternation of light and dark on earth in the relationship between the earth and sun, than the children who explored the phenomenon independently with the PICCO program. This difference was due to the fact that the children who received teaching tried to integrate these phenomena simultaneously, while in the conceptual models of children who explored the phenomena independently using PICCO by first exploring the causes of succession of seasons then the children started to integrate the alternation of lightness and darkness on earth in relation to the earth and sun in space (Kangassalo 1998c). The conceptual change in children’s conceptual models of the selected natural phenomenon in the PICCO environment followed Thagard’s (1992) classification: The modification of point of view takes place, the level of abstraction becomes higher, and the addition, deletion and reorganization of information occurred. Reorganization could be further divided into the replacement of existing relations into new relations, the joining together of separate relations or the discovery of new relations (Kangassalo 1997, 1998c). Before using the PICCO simulation, children’s conceptual models formed a starting point from which the exploration of the phenomenon was activated. Children’s exploration contained goal oriented and systematic action, wandering, seeking for something, investigating, experimenting, finding amusement with the space shuttle and making up stories. Goals and the intensity of exploration could vary, even during the same exploration situation. Furthermore, the more developed and integrated the conceptual model, the more the children’s exploration contained goal-oriented investigation and experimentation. Children’s
242
M. Kangassalo and E. Tuominen / Inquiry Based Learning Environment for Children
exploration strategies also developed at the same time when children’s conceptual models were developing. (Kangassalo 1996, 1997.) On the basis of the theoretical, methodological approaches and research findings of the PICCO research program, a new learning environment Proagents has been developed, which utilizes an exploratory learning approach. Additionally, a pedagogical support system has been constructed inside the system.
2.6. Question-Driven Inquiry as a Pedagogical Approach The theoretical foundations for the developed Proagents exploratory learning environment are derived from the inquiry-models that emphasize the role of questions as a starting point for the inquiry. One such a model is the interrogative model of inquiry. This model was originally developed for the purposes of the philosophy of science (see e.g., Hintikka 1985, 1988; Sintonen 1990, 1999), but it has also been used to represent knowledge-seeking in educational contexts (e.g. Hakkarainen et al. 2004; Hakkarainen & Sintonen 2002). In the interrogative model, the scientific procedure is viewed as information seeking by questioning. More specifically, inquiry is defined as a series of questions the inquirer poses during his/her inquiry process, either to nature or to some other source of information. The inquirer tries to derive an answer to his/her initial question or problem by using his/her existing knowledge and by formulating and seeking answers to smaller questions. The acquisition of new knowledge raises new questions that have to be examined. By choosing the questions, the inquirer can direct the course of the inquiry according to his/her own plans (Hakkarainen et al. 2004; Hintikka 1985). According to the interrogative model, an inquiry can be conceived as a dynamic, question-driven process of understanding (Hakkarainen & Sintonen 2002). Applying this model means that a child’s learning is viewed as an active process guided by his/her own questions and previous knowledge. The selection of the approach is largely based on Kangassalo’s earlier research findings (1997, 1998c) with the PICCO-computer environment where a child has been seen to progress in his/her exploration process step-by-step on the basis of his/her earlier knowledge foundation concerning the phenomenon in question. In this project, we intend to continue the PICCO-project by constructing support for children’s explorations. The support aims at encouraging the formation of questions in a child’s mind as well as the process of seeking for answers and explanations.
3.
Natural Phenomena for Simulations
3.1. Selecting Natural Phenomena When selecting the natural phenomena for the simulation applications it was essential that the phenomenon was important and significant in everyday life. The simulated phenomena have to awaken sufficient interest in the children and efficiently utilize the possibilities offered by the computer technology. The phenomena chosen are those that can in no other way be easily and illustratively presented, such as phenomena linked with space and elementary astronomy. An important selection criteria for the chosen natural phenomenon is that through its conformities to natural law, it forms a clear, well-organized knowledge structure and theory, and that these aspects lay a strong and well-defined foundation for the modelling of the phenomena in the simulation application. (See Kangassalo 1992, 1996, 1998a, 1998b.) The pictorial computer simulation PICCO concentrates on the variations of sunlight and heat of the sun as experienced on earth in relation to the positions of the earth and the sun in
M. Kangassalo and E. Tuominen / Inquiry Based Learning Environment for Children
243
space. On the earth level, the simulation concentrates on phenomena, which are close to the everyday experiences of children, such as day and night, seasons, changes in the life of plants and birds, etc. On the space level, it is possibilities to explore, for example, the earth, the earth and the sun, the solar system, the planets, and the dimensions of the universe. In the simulation it is possible, for example, to explore the variations in a natural environment on the basis of the interconnections and positions of the earth and the sun in space. The selected phenomena include multilevel interrelationships and central concepts. The whole phenomenon is rather complex and abstract, but the phenomena are clearly integrated with each other and form a coherent theory. The simulation program has been implemented in such a way that the knowledge structure and theory of the phenomenon are based on events appearing together with the phenomenon in question, and these events are illustrated. In the simulation all events and necessary elements are represented as pictures and familiar symbols. PICCO is very easy to use and it does not assume an ability to read or write. (See Kangassalo 1992, 1996, 1997.) From the basis of the PICCO simulation, the phenomena selected for the Proagents simulation were the change of night and day, the change of seasons, the earth, the earth's atmosphere, the layers of the earth, the earth and sun, and the solar system and its planets.
3.2. Cognitive Requirements for Modelling and Simulation The cognitive requirements for modelling and designing the natural phenomena for computerbased learning environments are based on theories and concepts of cognitive psychology, cognitive science, socio-cognitive approach and science learning. The main aim is that a constructed learning environment could support children in forming integrated abstract conceptual structures and models of the selected natural phenomena and support them in continuous knowledge construction process concerning the phenomena in question. Thus, cognitive requirements have to be taken into account when selecting the natural phenomena, writing the manuscript, modelling the phenomena, and simulating the phenomena onto the computer, as well as displaying the simulation on the screen and using the computer simulation. (See Kangassalo 1992, 1997, 1998a, 1998b.) The phenomena have to be modelled for the simulation applications according to the theory and existing knowledge of the phenomena. This means, for example, that information and knowledge on the screen and in the pedagogical agents’ descriptions, explanations and guidance have been designed and implemented according to the present scientific knowledge. Additionally, the sounds of natural phenomena, such as birdsong, the sound of wind and waves, are the natural sounds of nature. This is important because of children’s knowledge construction process and the formation of information and conceptual structures. These are significant because integrated and organized information, as well as knowledge structures in human memory at the general level of the phenomenon in question, is important to effective and demanding thinking, continuous knowledge construction and the theory formation. In these applications these requirements have been taken into account. (See e.g., Kangassalo 1992, 1996, 1997, 1998a, 1998b.) From the users’ perspective, it is important that the use of the application is based on the users’ own activity. Children can proceed according to their own interests and ideas. In these applications, there are no paths or rules on how to explore and go forward. Children can use as much time as they like each time. All this provides the children with possibilities to explore the phenomenon any time as long as they want and in the order they wish. When the program is under the user’s control, it is possible for the user to concentrate on the phenomenon in question. A child’s own activity, attention and interest, supports the development and construction of conceptual structures of the phenomenon within children. The more
244
M. Kangassalo and E. Tuominen / Inquiry Based Learning Environment for Children
complicated the phenomenon is, the more important is a child’s own activity and interest in analyzing and organizing information and its storing into the memory. (See e.g., Kangassalo 1997.)
3.3. Modelling the Phenomena Modelling the natural phenomenon for the computer simulations means constructing a presentation - a model - of the phenomenon. The modelling of the phenomenon is carried out by describing and presenting the core features and central events in the phenomenon such as they are in the phenomenon. By means of the model constructed onto the computer, the phenomenon can be imitated and simulated, that is to say that it is a simulated model of the phenomenon. (See e.g., van Gigch 1991, 122; Roberts, Andersen, Deal, Garet, and Shaffer 1983, 3; Rothenberg 1989, 75-82.) A simulated model of a phenomenon means that the phenomenon's events, objects, their characteristics and mutual time and space-related relations and changes within them are included in the model. The simulation model of the phenomenon is seen on the screen pictorially. The imitation of the phenomenon, by means of pictures, is constructed to be performed and manipulated via the computer. (Kangassalo 1997.) The modelling process for the PICCO simulation is described in detail in Kangassalo’s (e.g., 1997, 1998a, 1998b) articles and report. In the next chapters, the central features of the modelling process concerning the phenomena selected for the Proagents simulation will be described.
4.
The Proagents Learning Environment
4.1. Proactive Pedagogical Support System in the Environment In designing and developing the Proagents simulation system, the research done with the PICCO-research program has been continued. The aim was to design a multimodal computer simulation environment that would support pre- and primary school aged (6 to 8 years old) children’s exploratory and conceptual learning in the domain of astronomy. The computer simulation provides the children with an exploratory learning environment where they can explore the selected phenomena according to their own interests and questions. Children’s own questions are considered as a starting point for explorations. A child’s learning is viewed as an active process guided by his/her own questions and previous knowledge (see e.g., Lonka, Hakkarainen, Sintonen 2000). To achieve progress and deeper understanding, children need guidance and support for their exploration. In the system proactive pedagogical agents have been used to scaffold each child’s inquiries in the simulation by making questions and encouraging child’s own questioning and hypothesis formation as an aim to guide the child’s exploration process towards scientific inquiry. Agents’ operations in this system will be based mainly on auditory and haptic feedback, since the system has been developed both for sighted and visually impaired children. The interrogative model of inquiry as well as ideas from the progressive inquiry approach have been applied here as a pedagogical framework in creating and developing the proactive support for child’s explorations. As a principle, it is considered important that the support doesn’t replace the child’s own thinking, but rather guides the child to think and to explore the phenomena more deeply and extensively. Moreover, during the exploration process, the proactive agents scaffold the child by asking questions and posing problems that stimulates the child’s thinking and the formation of questions in child’s mind.
M. Kangassalo and E. Tuominen / Inquiry Based Learning Environment for Children
245
Proactive pedagogic agents support children’s explorations by encouraging a child’s own questioning, directing a child’s attention to objects and their relationships in phenomena, making questions and suggestions, guiding from familiar everyday phenomena, and progress gradually to more complicated topics and the causes and explanations of the phenomena. The agents scaffold each child with respect to his/her own capabilities and exploration paths. The agents don’t make any decisions for the child or force him or her into any particular exploration path. At any moment a child can choose either to listen to what the agents ask or suggest or to ignore. As a child’s explorations proceed, the agents’ support may decrease step by step. The agents’ support is based on a child’s explorations and user profiles. Very important in the agents’ action is the right timing and the form of the support and questions. The construction of the rules of proactive pedagogic agents has been designed and tested carefully. Children need guidance and support for their explorations for achieving progress and deepening their understanding. In the Proagents system, proactive agents have been constructed to support children’s questioning and explorations. Proactivity in this system can be defined as anticipative support. It takes into account the user and situation, predicts the users’ intentions, and acts accordingly. (See e.g., Tuominen, Peltola and Kangassalo 2003; Tuominen 2003, 2006.) The pedagogical agents have different imaginary characters and different names and voices. The entire learning environment has been constructed so that the narration and play are essential parts in children’s explorations. These elements form an important pedagogical support system for children’s exploration, science learning and thinking. This is because children’s thinking takes place in the form of continuing events, and fairy tales and stories support in recognizing and keeping in mind wholeness. In addition the meaning of different imaginary characters is to help children in analyzing and recognizing each individual themes in the application and this again helps children in navigation and the formation of interrelationships of different phenomena. (See e.g., Kangassalo 1997; Tuominen, Peltola and Kangassalo 2003.) The agent’s suggestions assist the child to find the central phenomena and concepts that are connected to the each application that the child is currently exploring. They also guide the explorations from one theme to another, and in this way support the finding of the relations and explanations in the selected phenomena. From the perspective of conceptual learning, the agents guide the explorations from familiar everyday observations towards the causes and scientific explanations of phenomena. For example if the child is exploring the solar system, and has already examined the different planets and their properties, an agent might suggest that the child if she or he seek out the planet earth. After that the agent may challenge the child to think why only earth has people and animals on it, and suggest if the child would like to explore the earth even closer. Furthermore, when exploring the earth an agent may direct the child’s attention to earth’s gravity (did you notice when you travelled with your space shuttle that you were pulled to the earth’s surface?), challenge the child’s thinking with arguments (what happens to different objects when you throw them into air or drop them?) and offer explanations on gravity. (See Kangassalo, Peltola, Tuominen 2003/2004.) In summary, the agents try to guide the children to elaborate their previous knowledge through their questions and suggestions, and encourage them to examine the properties and relations in phenomena. The proactive agents also aim to help the children become conscious of their own exploration and thinking. The agents allow children to explore and proceed in different ways, and the child him/herself can continuously choose either to listen to what an agent wants to say or to ignore him.
246
M. Kangassalo and E. Tuominen / Inquiry Based Learning Environment for Children
4.2. Proagents as an Exploration Environment In the Proagenst learning environment a child can explore the earth and its rotation around its axis, the solar system and its planets, the revolution of the earth around the sun and the atmosphere and core of the earth. These two latter ones were not yet at children’s use in the carried out research experiment. Because the environment has been constructed also for visually impaired children’s use, the exploration occures by using the stylus which gives haptic feedback (the Phantom stylus, http://www.sensable.com), see Figure 1. In the environment haptic feedback is supported by auditive feedback. The following description is based on the manuscript of the Proagents – learning environment (Kangassalo, Peltola, Tuominen 2003/2004; Tuominen 2006).
Figure 1. Phantom device in a child’s use
A child starts an exploration from the hexagonal research station, where in each corner is the door to the exciting research world, so-called mini-world. When a child would like to explore the solar system and its planets, at the door of the solar system the pedagogical, proactive agent Antti Astronaut welcomes an explorer. Antti Astronaut guides in the solar system from a planet to another planets, asks if a child would like to know more about planets or if she would like to follow the revolution of the different planet around the sun. As a child touches one of the orbits, the agent tells the child which planet's orbit it is. The orbit can be felt under the haptic stylus as a groove when a child is moving along the groove. When a child is exploring the orbit of the certain planet and she is just touching the planet, the program tells her what planet it is and where it is located in relation to the sun. It is also possible to listen to more information on each of the planets and the sun. The agent Antti Astronaut guides, gives more information and makes questions, if a child would like to choose the listening of Antti Astronaut. When a child would like to continue from the solar system to another mini-world, she presses a button in the system and returns to the research station and chooses a new mini-world. In the earth mini-world Earth Giant and his finger is guiding the exploration of the earth and its surface. The earth can be felt as a three-dimensional round object. A child can feel by using the finger of the Earth Giant the different forms of the surface. When touching the surface of the earth, it is possible to feel the differences between solid ground and the oceans. The ground feels hard and uneven, while the oceans are more even areas. It is also possible to feel the biggest mountain ranges. When touching the surface of the earth, a child can hear sounds of human inhabitation and the nature. At the sea it is possible to hear typical sounds to the ocean (waves, seagulls). A child may also wish to listen to which continent or ocean she is
M. Kangassalo and E. Tuominen / Inquiry Based Learning Environment for Children
247
currently exploring. The earth also has gravity that can be felt with the stylus as a light pull towards the earth. It is also possible to rotate the earth around its axis and to feel the moving of the earth under the finger of Earth Giant (stylus). As the visual feedback, it is possible to see the round globe and the Phantom stylus on the screen. In the mini-world, the earth and the sun, a child can follow the revolution of the earth around the sun. When a child is moving around the sun on the earths’ orbit she can listen typical sounds to different seasons in Finland. The sounds are located on the certain area on the orbit when the earth is revolving around the sun just at that seasonal time. The sounds of the seasons follow the time period of the seasons in Finland during the earth’s revolution on its orbit around the sun. Sunny Anneli, the pedagogical agent of the mini-world, tells about the seasons in Finland and asks, if a child would like to listen to more information about the things which cause the seasons in Finland during the earth’s revolution around the sun. Sunny Anneli also asks, if a child would like to listen some questions. Anneli asks, for example, how long the earth’s revolution around the sun takes in real time. She also tells about the sun’s light and warmth in each season. The sun, the earth, and the earth's orbit are shown as visual feedback on the screen To the mini-world called the research laboratory there is the collection of the different question and tasks for the child. The questions concern the things in the mini-worlds and the information and knowledge of their pedagogical agents. Next there are some examples of the questions, “if you throw a toy car from a tower, does it start floating in space” or “how long it takes when the earth is revolving around the sun”, which causes the variation of the seasons in Finland” and so on. If a child's answer is wrong, the program asks a child to think about it once more and listen to the question again. In the bowels of the earth mini-world a child can explore the insides of the earth. The various layers of the earth are represented as a cross-section of the northern hemisphere. The layers can be explored by touching them with the stylus. The top layer is hard and 'stony'; when descending to the interior, the layers become softer and softer. The haptic feedback inside the earth simulates the liquid core of the earth. As a visual feedback, the child may see a cross-section of the various layers and the Phantom stylus. In this mini-world the pedagogical agent Mr. Kairanen guides the child and they use a drilling machine for moving. In the atmosphere mini-world the child may study the earth's atmosphere from the surface of the earth to the upper layers of the atmosphere. The mini-world is presented similarly to the bowels of the earth mini-world, as a cross-section in which the bottom denotes the ground and the top of the screen (and the touching area) is the topmost border of the atmosphere. Exploring the mini-world is first and foremost based on auditory feedback. As a child is moving at the bottom of the touching area (”near the ground”), she can hear people's voices, birdsong, and the leaves of trees moving. As a child moves upwards, the sounds of airplanes and typical sounds to the wind can be heard. Further up in the atmosphere, the voices grow softer and disappear, until on the outside of the atmosphere there is silent space. The haptic feedback is almost unnoticeable and light, and it aims to create a tangible ”feeling of air”. A child can freely move using the stylus in the different layers of the atmosphere. The program also tells the child, when the child herself so desires, about the characteristics and the importance of the atmosphere. The moving in the mini-world occurs by using a space lift and the guide in this world is Iiro Ilmarinen.
4.3. Analyzing Children’s Inquiry: An Example An example of one child’s inquiry as he used the program is described in this section. The description is based on micro-level analysis of the inquiry process. It includes the analysis of the child’s exploration times in the micro-worlds and observations on the child’s questions
248
M. Kangassalo and E. Tuominen / Inquiry Based Learning Environment for Children
during the exploration. The agents’ messages and the researcher’s guidance during the use of the system are also examined. The child, whose inquiry is described next, used the system at the school for visually impaired children in Jyväskylä, Finland. He used the system twice, approximately 43 minutes on the first day and 37 minutes on the next day. The researcher sat next to the child and assisted the child as he used the program. Before and after the use of the system the child’s conceptual model was evaluated. The evaluation was based on Kangassalo’s (1997) study on the formation of children’s conceptual models of a natural phenomenon when using PICCO. On the basis of the first evaluation it was possible to say that the child had already formed models of spherical earth and sun before the use of the program. He knew all the planets by name and could arrange them in order starting from the sun. The order and the regularity of the times of day and the seasons were also organized on the level of the surface of the earth. The connections between the surface level of the earth and the mutual relations between the sun and the earth were very weak, however. For example, when thinking about the changing of day and night, the child thought that at night time the sun sets towards the equator, and during the day, the sun rises high up in the sky. When asked about causes of the seasons, the child first said that it’s because “the cold and warm waves hit”. He also explained that the earth ”turns to winter” and showed this by rotating the modelled earth back and forth against the table. The child was very active and concentrated while using the program. He thought long and hard about what he wanted to study next. He started his exploration in the solar system micro-world. This was also his most explored area, a total of 26 minutes, Table 1. The fact that the child started his exploration from the solar system micro-world corresponds to the previous studies (e.g., Kangassalo 1997) with the PICCO-environment where it was found that the children’s conceptual models formed a starting point from which the exploration of the phenomenon was activated. In this case, the child already knew the planets in the solar system while the mutual connections of the earth and the sun were quite disorganized.
Table 1. The exploration times in each mini-world Mini-world Solar system The earth The earth and the sun Research station
Time of exploration (total) 26 min 9 min 10 min 17 min
At the research station the child usually wandered around the different corners of the station and listened many times as the program told what could be explored in each micro-world. This also shows in the large amount of time he spent at the station (17 minutes), Table 1. He expressed many times that he would like to explore the atmosphere, but unfortunately that particular micro-world wasn’t yet constructed at the time of the research experiment. At the solar system micro-world the child’s exploration included wandering as he went through the planets, and sometimes he said aloud things like: ‘Let’s see if it would tell about Pluto’. During his exploration he also sought for different planets, and investigated the temperatures of the planets. He also actively investigated the surface features of the earth in the earth micro-world. The child was also interested in the program and its operations and experimented often what would happen if he, for example, pressed objects with the stylus. In the exploration the following strategies could be observed: wandering, seeking for something, investigating and experimenting (cf. e.g., Kangassalo 1997).
249
M. Kangassalo and E. Tuominen / Inquiry Based Learning Environment for Children
Some signs of reflection could be observed during the child’s exploring process as the child said, for example, ‘I don’t need to explore Pluto’. Reflection concerning the exploring process appeared also as the child started to rethink his earlier speculations about the duration of the earth’s revolution around the sun. The child had earlier, when exploring the earth’s revolution around the sun in the earth-sun micro-world, said that he thought that it takes the earth two months to go round the sun. The reflection started after hearing an agent’s message concerning the earth’s circumference: An Agent: You are now exploring the earth, the planet in which we live. The earth's circumference is approximately 40,000 kilometres. This means that if you could travel around the world by car, you would have to sit still for three weeks in a row. C: It takes less time than the round of the sun. I think I guessed a bit wrong. R: Well, what do you think about it now? C: Now I think if you took an airplane it'd take only two weeks. The example shows also that the duration of the earth’s revolution around the sun hadn’t yet been organized in child’s mind despite the active exploration of the earth-sun –micro world. It may be that the exploration of this micro-world is too confusing without a more developed model of the mutual relations of the earth and the sun in space and its connection to the seasons. As discussed in section 2.6, the pedagogical approach of the simulation program is based on interrogative model of inquiry, which means that inquiry is viewed as a series of questions the inquirer poses during his/her inquiry process. Next the questions the child asked during his inquiry are examined (see Table 2). The child asked a lot of questions during his exploration process. Most of the questions concerned the program and its operations. That is understandable because the child had to operate without visual feedback, and the equipment and the program was new to the child. There were eight (8) questions on the first exploration time and nine (9) questions on the second exploration time that clearly concerned about the phenomenon and its exploration. Most of these questions were fact-seeking questions (i.e. When have we explored the whole earth? Which planets have many moons? Are there two hundreds degrees?) that could be answered by providing factual information. During the exploration there were no questions concerning the phenomena that could be categorized as explanation seeking (how or why questions). A few of these questions could, however, be found in the evaluation situations that were conducted before and after the exploration. It seems that - on the level of questions that the child posed – the inquiry concentrated very much on the program and its operations.
Table 2. The child’s questions during the exploration The child’s questions during the exploration closed question: if/whether… what where, when how, why
Other
41
Exploration/ Phenomenon 8
13 6 8
4 5 -
5 1 -
The device
The program
6 2 1 4
7
During the child’s exploration there were a total of 16 messages from the agent, Table 3. The messages came mainly in the solar system micro-world (13 of the 16 messages). Of these 16 messages, there were only two messages that the child did not choose to hear. It may
250
M. Kangassalo and E. Tuominen / Inquiry Based Learning Environment for Children
be that one of the rejections was because the child wanted to test the agents. The other message that the child rejected came when the child was exploring the planets in the solar system. In most of the messages, the agent provided more information and explanations about the phenomena. The same message often came twice due to a failure in the system. Table 3. Agent’s messages and the child’s responses TIME (s) 0 508 1647
MICROWORLD
Research station Solar system
1659 1668
1947 1997
0 826
The Earth
890 1302
Solar system
1417 1426 1437
1457
1558 1709
1770
AGENT’S MESSAGE
1st exploration time Agent tells: Distances in solar system. Agent tells: Planets revolution around the Sun. Agent tells: The heat of the Sun The agent tells: The Sun’s distance from the Earth Agent tells: Distances in solar system Agent tells: The biggest planet is Jupiter. 2nd exploration time Agent tells: The circumference of the Earth
Message not received. Agent tells: The biggest planet is Jupiter. Message not received. Agent tells: The heat of the Sun. The agent tells: The Sun’s distance from the Earth. Agent tells: Planets revolution around the Sun. Agent tells: Distances in solar system. Agent tells: Planets revolution around the Sun. And asks: how long it takes the Earth to go round the Sun. Agent tells: Planets moons.
CHILD’S REACTIONS/ RESPONSES
CHILD’S EXPRESSION
-
-
-
-
-
-
The child comments the message.
“It would take a long time.”
-
-
-
-
Re-evaluating earlier thinking
“It takes less time than the round of the sun. I think I guessed a bit wrong.” “I look now at “no”.” “What would it say if I press “yes”?”
Testing the agents Testing the agents
-
-
-
-
-
An intentional choice to hear the agents. -
“I press “yes” so that I can hear information.” -
-
-
Elicitates a question from the child.
“What planets have many moons?”
M. Kangassalo and E. Tuominen / Inquiry Based Learning Environment for Children
251
When examining the messages in relation to child’s exploration process as a whole, it could be observed that there were during the second exploration time a phase where the child obviously tested the agents and afterwards, the using of agents changed intentional (“I press yes so that I can hear information”), Table 3. In addition, after hearing an agent tell about the circumference of the earth, the child started to rethink his earlier ideas about the duration of earth’s revolution around the sun. One of the agents’ messages also elicited a question from a child (“what planets have many moons?”) that could have been explored further. It seems that the child became more interested in the agents and their messages during the exploration, and was able to include them better as a part of his inquiry as the exploration proceeded. During his second exploration time the child even expressed a wish to hear an agent when he was in solar system micro-world and explored the sun. At that time he said: “If a message comes here, I will press “yes””. The researcher’s guidance focused largely on the using of the program. The researcher, for example, told the child how he could operate in the micro-world and what could be explored there. She also assisted in the use of the stylus, as it is important to keep the stylus in a right position to be able feel to objects properly. With regard to the phenomena, the researcher assisted the child in finding different objects in the micro-worlds and directed the child’s attention to central objects or features of micro worlds. She also offered explanations of phenomena and asked some questions. Table 4 presents the most essential aspects of the researchers guidance with regard to the phenomena and its exploration in each micro-world.
Table 4. Researcher’s guidance concerning the exploration of phenomena in different micro-worlds. SOLAR SYSTEM st 1 exploration time Assists in finding planet.
Assists the child find a planet. Guides the child to go along an orbit to find a planet Assists the child to find a planet.
Assists the child to find a planet. Guides the child to go along an orbit to find a planet Assists the child to find a planet. 2nd exploration time Asks: What would happen if the earth was there where mercury is? (The child doesn’t answer.) Explains the planets motions and asks if the child remembers how the earth moves in other micro worlds. (An agent’s message interrupts – the child doesn’t
THE EARTH st 1 exploration time Directs the child’s attention to the different sounds that can be heard when exploring the surface of the earth. Explains about the earth’s rotation Directs the child’s attention to feeling different surfaces. Asks about meaning of the “ticking” noise when rotating the earth. (The child doesn’t answer.)
nd
2 exploration time Asks about meaning of the “ticking” noise when rotating the earth. (The child doesn’t answer.) Asks once more about meaning of the “ticking” noise when rotating the earth. (The child doesn’t answer.)
THE EARTH AND THE SUN st 1 exploration time Directs the child’s attention to the sounds of seasons.
Assists the child to find the earth. Directs the child’s attention to the sounds of seasons. Assists the child to the orbit of the earth.
2nd exploration time Asks the child if he could find the orbit of the earth.
Asks the child to name the different seasons when the corresponding sounds are heard when circling the orbit.
252
M. Kangassalo and E. Tuominen / Inquiry Based Learning Environment for Children
answer.) Tells about planets distances.
Helps the child to find the earth.
Explains the meaning of the ticking noise as the child’s rotates the earth. Suggests that the child rotates the earth and examines the changes of day and night.
Asks about duration of the earth’s revolution around the sun. Guides the child to circle the orbit again and listen to the sounds of seasons. Asks again about the duration the earth’s revolution around the sun.
As can been seen in Table 4, in the solar system micro-world the researchers guidance mostly concentrated on assisting the child to find the planets he wished to explore. She also asked a few questions to which, however, the child didn’t answer. When the child was exploring the earth micro-world, the researcher directed the child’s attention to feeling the earth’s surface. She also explained the earth’s rotation around its axis, and tried to describe how in this program the ticking sound meant the child is rotating the earth. After that, she tried many times to get the child to think about why there was a ticking sound when the child rotated the earth. The child didn’t answer the researcher’s questions, maybe because the question itself can be considered a bit confusing. On the basis of the evaluation of the child’s conceptual model, it is also possible that the question was too difficult for the child as the child’s model concerning the changing of day and night was largely based on model where the sun sets and rises. In the earth-sun micro-world the researcher first directed the child’s attention to the different sounds of the seasons that could be heard when circling the sun. At the second exploration time she tried to direct the child’s thinking to the duration of the earth’s revolution around the sun on the basis of seasonal sounds heard when circling on the earth’s orbit. The duration of the revolution and its connection to the changing of seasons didn’t, however, outline for the child. Next, the changes that took place in child’s conceptual model after using the simulation program are briefly examined. There were some small changes that could be observed in the child's conceptual model concerning the phenomenon. The model that was based on the setting and the rising of the sun was left out or grew weaker: in the evaluation after using the program, the child no longer knew why the times of the day change and was not able to show this with the modelling clay. Another slight change that took place in the child's model is that the earth started to rotate around its axis. However, the child mostly associated this with the change of seasons. To summarize, it could be said that after using the program, some erroneous associations disappeared, and on the other hand, some still erroneous ones were being formed.
4.4. Environment Supporting Children’s Conceptual Thinking Studying how children explore the phenomena with the learning environment has been only one of the goals in our research. One of the main aims in the construction of the learning environment has been to support children’s conceptual thinking and learning with regard to the selected natural phenomenon. Our research has focused especially on the formation of visually impaired children’s conceptual models of the phenomenon in question in a situation where children are using the learning environment. Two 7-8-year-old visually impaired children participated at the research experiment at the school for visually impaired children and a third child in the usability laboratory a few months later. The children were interviewed both before and after using the program. The aim of the
M. Kangassalo and E. Tuominen / Inquiry Based Learning Environment for Children
253
interviews was to elicit the children’s conceptual models with regard to the chosen astronomical phenomena. The eliciting of conceptual models was based on earlier research conducted by Kangassalo (1997). The interviews were videotaped so that both operative and verbal expressions of the child could be taken into account when analyzing the data. The research data also included video recordings and log files of the children’s use of the program, and the questionnaire for the children’s parents. The video recordings of the children’s interviews were transcribed and analyzed in order to study the children’s conceptual thinking and conceptual change. The analysis of the exploration processes was based on log files and video recordings of the situations where children used the system. The log files provided information about the child’s exploration pathways: how long and in which order the child explored the micro worlds and what kind of things was explored. The child’s comments, questions, and other expressions as well as the researcher’s guidance could be observed from the videotapes. To obtain an accurate picture of the child’s exploration process, these video recordings were also transcribed next to the log files. With the collected data, it was possible to describe the children’s knowledge construction processes during the research period and also to examine the nature of children’s exploratory action in the constructed environment. (Tuominen 2006.)
5.
Modelling Children’s Exploration and Learning
The PICCO research program continues to study the development of the conceptual learning and thinking of 5- to 10-year-old children regarding selected natural phenomena. It examines children’s inquiry learning, reciprocal interaction as well as the potential of the learning environment in such activity environments where children can use information technology in addition to traditional materials and tools. With regard to the development of children’s conceptual thinking on natural phenomena, the specific research object is the formation of children’s conceptual models of natural phenomena, conceptual changes and the construction of knowledge (e.g., Kangassalo and Kumpulainen 2003, 2006.). In order to obtain a coherent picture of children’s conceptual and exploratory learning, social interaction and their meaning in children’s conceptual thinking and knowledge construction, it has been necessary to synthesize the collected empirical data into a coherent form. Specific description techniques will be developed for this purpose. The description techniques aim at describing children’s conceptual learning and thinking from the bases of the exploratory learning approach by taking into account children’s self-regulation and meta cognition, peer activities and adults’ (and agents’) scaffolding processes in the situated context. Due to the developmental age of the children as well as due to the tool rich learning context, the analysis techniques try to capture children’s activities from a holistic viewpoint by concentrating on modelling operative, non-verbal and verbal expressions. In our research project we would like to know how children’s knowledge construction process, social interaction and children’s exploration processes are developing and integrating and we try to find interrelationships social and cognitive activities in the development of understanding of the phenomena in question. The final aim will be to develop the description technique by which it is possible to simulate and animate dynamically children’s conceptual thinking and learning from multiple dimensions by taking into account the dynamic and situated nature of conceptual thinking and learning. (Kangassalo and Kumpulainen 2003, 2006.) In this article, the theoretical approach of the inquiry learning, its application to the simulation programs and first steps for analyzing and modelling children’s explorations has been described.
254
M. Kangassalo and E. Tuominen / Inquiry Based Learning Environment for Children
Acknowledgements We thank the children, their teachers and parents for participating in our experiments, the PICCO (115161) and Proagents (202179) research groups of Early Childhood Education, the Department of Teacher Education, University of Tampere for the theoretical and pedagogical developmental work, and the research group of the Department of Computer Sciences (202180) for the technical realization of the Proagents application. The mentioned projects are funded by the Academy of Finland.
References Andre, T. & Windshitl M. 2003. Interest, Epistemological Belief, and Intentional Conceptual Change. In G. M. Sinatra & P.R. Pintrich (Eds.) Intentional Conceptual Change. Mahwah, NJ: Lawrence Erlbaum Associates, 173197. Ausubel, David P. 1965. Introduction. In Richard C. Anderson and David P. Ausubel (Eds.) Readings in the Psychology of Cognition. New York: Holt, Rinehart and Winston. Ausubel, D. P. 1968. Educational Psychology: a Cognitive View. New York: Holt, Rinehart and Winston. Caravita 2001. A re-framed conceptual change theory? Learning and Instruction, 11, 421-429. Diakidoy, I-A. Vosniadou, S. & Hawks, J.D. 1997. Conceptual change in astronomy: Models of the earth and of the day/night cycle in American-Indian children. European Journal of Psychology of Education, XII (2), 159184. van Gigch, John P. 1991. System Design Modeling and Metamodeling. New York and London: Plenum Press. Hakkarainen, K. & Sintonen, M. 2002. The interrogative model of inquiry and computer-supported collaborative learning. Science & Education, 11, 25-43. Hakkarainen, K., Lonka, K. & Lipponen, L. 2004. Tutkiva oppiminen. Järki, tunteet ja kulttuuri oppimisen sytyttäjinä. 6. uudistettu painos. [Progressive Inquiry: Reason, emotions and culture as the initiators of learning.] Porvoo: WSOY Hatano, G. & Inagaki, K. 2003. When is conceptual change intended? A cognitive-sociocultural view. In G. M. Sinatra & P.R. Pintrich (Eds.) Intentional Conceptual Change. Mahwah, NJ: Lawrence Erlbaum Associates, 407-427. Hintikka, J. 1985. True and false logic of scientific discovery. In Hintikka, J. (Ed.) Logic of discovery and logic of discourse. New York: Plenum. Hintikka, J. 1988. What is the logic of experimental inquiry? Synthese, 74, 173-190 Kangassalo, M. 1991. PICCO. The pictorial computer simulation of a selected natural phenomenon for children’s use. Computer program. The grant of the Academy of Finland. Kangassalo, M. 1992. The pictorial computer-based simulation in natural sciences for children’s use. In Ohsuga, S., Kangassalo, H., Jaakkola, H., Hori, K. and Yonezaki, N. (Eds.) Information Modelling and Knowledge Bases III: Foundations, Theory and Applications. Amsterdam: IOS Press, 511-524. Kangassalo, M. 1994. Children’s independent exploration of a natural phenomenon by using pictorial computerbased simulation. Journal of Computing in Childhood Education 5 (3/4), 285-297. Kangassalo, M. 1996. Picco as a cognitive tool. In Y. Tanaka, H. Kangassalo, H. Jaakkola & A. Yamamoto (Eds.) Information modelling and knowledge bases VII. Amsterdam. IOS Press, 344-357. Kangassalo, M. 1997. The formation of children’s conceptual models concerning a particular natural phenomenon using PICCO, a pictorial computer simulation. Acta Universitatis Tamperensis 559. University of Tampere. 188 pp.
M. Kangassalo and E. Tuominen / Inquiry Based Learning Environment for Children
255
Kangassalo, M. 1998a. Luonnonilmiö kuvalliseksi tietokonesimulaatioksi. Valitun ilmiön mallintaminen, simulaation ja kuvaruudun näyttöjen suunnittelu. [A Natural Phenomenon into a Pictorial Computer Simulation: Modelling the Selected Phenomenon and Designing the Simulation and the Computer Screen]. Tampereen yliopiston julkaisusarja TAJU. Tampere. 129 p. Kangassalo, M. 1998b. Modeling a Natural Phenomenon for a Pictorial Computer-Based Simulation. In Kangassalo, H., Charrel, P-J, Jaakkola, H. (Eds.) Information Modelling and Knowledge Bases IX. IOS Press, Amsterdam, 239-254. Kangassalo, M. 1998c. Conceptual change in astronomical phenomena using PICCO. EARLI, Second European Symposium on Conceptual Change, Madrid, November 6-8, 1998. Kangassalo, M. 1998d. PICCO Day and Night, Seasons – the Earth and the Sun. Cd-rom. ISBN 951-98035-0-5. Piccos Programs Ltd. Kangassalo, M. 2001. Explorative learning in PICCO -environment. In Hannu Jaakkola, Hannu Kangassalo and Eiji Kawaguchi (Eds.) Information Modelling and Knowledge Bases XII. Amsterdam: IOS Press, 259-261. Kangassalo, M. and Kumpulainen, K. 2003. The Dynamics of Children's Science Learning and Thinking in a Social Context of a Multimedia Environment. In Hannu Kangassalo, Eiji Kawaguchi, Bernhard Thalheim and Hannu Jaakkola (Eds.) Information Modelling and Knowledge Bases XIV. Amsterdam: IOS Press, 188-197. Kangassalo, M. and Kumpulainen, K. 2004. Methodological Tools for Describing Children’s Knowledge Construction Process in Multimedia Environment. In Yasushi Kiyoki, Eiji Kawaguchi, Hannu Jaakkola, Hannu Kangassalo (Eds.) Information Modelling and Knowledge Bases XV. Amsterdam: IOS Press, 119-122. Kangassalo, M. and Kumpulainen, K. 2006. Investigating the ecological dynamics of children’s conceptual models in technology-enriched science classroom. A paper presented at the Annual American Educational Research Association Conference, San Francisco, April 2006, CA. Kangassalo, M., Raisamo, R., Hietala, P., Järvi, J., Peltola, K., Saarinen, R., Tuominen, E., Hippula, A. 2005. Proactive Agents That Support Children’s Exploratory Learning. In Kiyoki, Y., Wangler, B., Jakkkola, H., Kangassalo, H. (Eds.) Information Modelling and Knowledge Bases XVI. IOS Press, Amsterdam, 123-133. Lonka, K., Hakkarainen, K., Sintonen, M. 2000. Progressive inquiry learning for children – Experiences, Possibilities, Limitations. European Early Childhood Education Research Journal. Vol. 8, 7-23. Roberts, Nancy, Andersen, David F., Deal, Ralph M., Garet, Michael S. and Shaffer, William A. 1983. Introduction to computer simulation: The system dynamics approach. London: Addison-Wesley. Rothenberg, Jeff 1989. The Nature of Modeling. In Lawrence E. Widman, Kenneth A. Loparo and Norman R. Nielsen (Eds.) Artificial Intelligence, Simulation, and Modeling. New York: John Wiley & Sons, 75-92. Sintonen, M. 1990. How to put questions to nature. In D. Knowles (Ed.) Explanation and its limits. Royal Institute of Philosophy lecture series; 27. New York: Cambridge University Press, 267-284. Sintonen, M. 1999. Why questions, and why just why-questions? Synthese 120, 125-135. Snirr, J., Smith, C. & Grosslight, L. 1995. Conceptually Enhanced Simulation: A Computer Tool for Science Teaching. In D.N. Perkins, J.L. Schwartz, M.M. West & M.S. Wiske (Eds.) Software Goes to School. Teaching for Understanding with New Technologies. New York. Oxford University Press, Inc. Thagard, Paul 1992. Conceptual Revolutions. Princeton, NJ: Princeton University Press. Tuominen, E., Peltola, K. and Kangassalo, M. 2003. Proactive Agents Supporting Children’s Exploratory Learning. JURE pre-conference, 25th – 26th of August, 2003, Padova, Italy. Tuominen, E. 2003. A Visually Impaired Child, Proactive Agents and Conceptual Learning. At 13th Annual Conference on Quality in early Childhood Education, EECERA Conference, Glasgow, Scotland, September 3-6, 2003.
256
M. Kangassalo and E. Tuominen / Inquiry Based Learning Environment for Children
Vosniadou, S. 1991. Conceptual Development in Astronomy. In Shawn M. Glynn, Russell H. Yeany and Bruce K. Britton (Eds.) The Psychology of Learning Science. Hillsdale, NJ: Lawrence Erlbaum, 149-177. Vosniadou 1999. Conceptual Change Research: State of Art and Future Directions. In Schnotz, W., Vosniadou, S. & Carretero, M. (Eds.) New Perspectives on Conceptual Change. Oxford, UK: Elsevier Science, 3-14. Vosniadou, S., Ioannides, C., Dimitrakopoulou, A. & Papademetriou, E. 2001. Designing learning environments to promote conceptual change in science. Learning and Instruction, 11, 381-419. Unpublished manuscript Kangassalo, Peltola, Tuominen 2003/2004. The manuscript of the Proagents learning environment. Department of Teacher Education. Early Childhood Education. University of Tampere, Finland. Tuominen, E. 2006. The Development of the Conceptual Models by Visually Impaired Children concerning the Selected Astronomical Phenomena in the Multimedia Learning Environment. The Manuscript of the Doctoral Dissertation, the Department of Teacher education, Early Childhood Education, University of Tampere.
Information Modelling and Knowledge Bases XIX H. Jaakkola et al. (Eds.) IOS Press, 2008 © 2008 The authors and IOS Press. All rights reserved.
257
A Perspective Ontology and IS Perspectives Mauri LEPPÄNEN Department of Computer Science and Information Systems P.O. Box 35 (Agora), FI-40014 University of Jyväskylä, Finland [email protected] Abstract. Information processing is often such a large and complex artifact that it goes beyond a human being’s capacity to conceive, model and develop it with all of its aspects at a time. For this reason, it is typical for one to focus on some aspects of it in one time and on other aspects in another time, depending on the problem at hand. For recurrent situations it is necessary to have a structured and well-defined set of perspectives which guide selections and shifts of focuses. This paper presents a light-weight perspective ontology which provides a set of well-defined perspectives, established on three dimensions, to conceive issues in information processing in an organized manner. The perspective ontology can be applied on different information processing layers, such as information systems (IS), information systems development (ISD) and method engineering (ME). To demonstrate the applicability of the ontology, it is used to derive a set of IS perspectives with basic IS concepts and constructs. The IS perspectives are then deployed as a framework in a comparative analysis of current perspectives in the IS literature.
Introduction Information systems (IS) are large and complex artifacts. That is why it has become a commonplace to decompose information systems development (ISD) work into activities, tasks and operations in such a way that in each of them it is possible to focus on certain features of an IS. To make an ISD process, in this sense, more structured and manageable, a number of perspectives, views and viewpoints have been proposed [1, 9, 13, 21, 22, 27, 45, 46, 51, 53, 55, 60, 62]. These are used, not only in structuring ISD processes, but also in specifying quality criteria for IS and contingency factors for ISD efforts. Perspectives are not advantageous merely in the IS context. They benefit considerations in other fields of information processing as well. ISD methods, for instance, are often quite complicated. To integrate, customize, configure and implement an ISD method for the use of an organization, or a project, it is necessary to focus on some specific features of the method at a time. Some of those features may relate, for instance, to the semantic contents of the method. Method engineering (ME), in turn, is commonly carried out with the support of some methodical artifacts, for example ME strategies (e.g. [49]), meta models (e.g. [25, 18]), ME techniques (e.g. [26, 50]) and ME steps (e.g. [30, 52, 56]). The development of these kinds of ME artifacts also benefit if there are well-defined perspectives which can direct one to pay attention to particular features of an ME artifact at a time. Sets of perspectives have been discussed since the 1970’s. However, most of them have no theoretical basis justifying how the perspectives in the sets are related to one another and how they should be used in a rigorous manner. In addition, perspectives are not clearly defined. They are also, with only a few exceptions, IS specific, meaning that they have been engineered merely for the development and evaluation of an IS. To our knowledge, there is only one presentation [16] which suggests a set of views for engineering ISD methods.
258
M. Leppänen / A Perspective Ontology and IS Perspectives
These views are not, however, grounded on any theory, and it is not made explicit how the views are related to one another. To our view, there is a need for a consistent set of well-defined perspectives that can be used as a common framework in the consideration of information processing on the IS, ISD and ME layers. The higher the concerned layer is, the more necessary it is that the same kinds of perspectives are available for considerations on the lower levels as well. For instance, in method engineering the perspectives are used to integrate and organize method components (i.e. models and techniques) taken from existing ISD methods. In this process it is necessary to know, among other things, which features of the IS are of relevance from which viewpoint of method components, and in which order the features should be discussed during this process. To have this kind of shared set of perspectives, we need a sound conceptualization of issues related to information processing. Ontologies are kinds of frameworks unifying different conceptions and serving as a basis of common understanding. More specifically, an ontology is an explicit specification of a shared conceptualization of some part of reality that is of interest [cf. 14]. The part of reality we are here interested in concerns information processing. The purpose of this paper is to present a light-weight perspective ontology which provides a set of well-defined perspectives to conceive, understand, structure and represent aspects of information processing in an organized manner. The perspective ontology is aimed to be general enough to support considerations on three layers, namely the IS, ISD and ME layers. The concepts and constructs in the perspective ontology have been defined in a deductive and an inductive manner. Following an iterative procedure based on [58] and [11], we first determined the purpose, domain and scope of the ontology. Second, we searched for theories that address the domain (i.e. information processing) and provide grounds for specifying perspectives. Third, we analyzed existing presentations for views, viewpoints and perspectives to find out whether some of them could be integrated, as such or adapted, into our ontology. Fourth, we defined the basic concepts and constructs of the ontology, including the criteria, dimensions and perspectives. Fifth, we evaluated the perspective ontology on the basis of quality criteria (e.g. [4, 15, 57]) in several stages. This included applying the ontology to derive the perspectives for IS, ISD and ME. In order to have a more detailed view of the perspectives on the IS layer, called the IS perspectives, we defined a comprehensive set of concepts and constructs referring to essential aspects of the IS from each of the IS perspectives. Furthermore, we made a comparative analysis of current IS perspectives to show how sets of IS perspectives presented in the IS literature compare to one another, and to our IS perspectives. The rest of the paper is organized into five sections. In Section 1 we define the basic concepts related to information processing and consider them in relation to contexts and processing layers. In Section 2 we define the notion of perspective ontology established on one or more dimensions, specify five perspectives and show how they are applied on the IS, ISD and ME layers. In Section 3 we define for each IS perspective an array of IS concepts and constructs to be used when applying the perspective. In Section 4 we deploy the IS perspectives to compare and analyze sets of perspectives presented in the literature. The paper ends with a summary and conclusions. 1. Information Processing Reality is anything that exists, has existed or will (possibly) exist. The subjective reality is the result from our mental processes [3, 39]. The physical reality is the source of sense data, which we obtain, and it is thus external to us. A thing means any phenomenon in reality whether subjective or physical. Based on semiotics, there are three kinds of things,
M. Leppänen / A Perspective Ontology and IS Perspectives
259
concepts, signs and referents. Concepts are mental things, words of mind [19]. A sign is any thing which can stand for something else. A referent is a thing to which a concept refers. Predicates are properties of things that are used to characterize things. They determine the applicability of a concept. The human mind produces a variety of conceptions about the same thing in the physical reality, depending on a point of view adopted. Using a point of view, some things and some properties of the things are selected because they are more relevant than the others. A universe of discourse (UoD) is a part of the subjective reality that becomes relevant from the point(s) of view adopted. To derive and relate the points of view, some framework is commonly deployed. A framework is a thing that guides a human being to select the points of view that are the most appropriate for the case or the problem at hand. A framework can be intuitive or formally established, vague or rigid. Human and social actions are based on expertise and its accumulation through thinking and communication processes. Expertise is knowledge which is a relative stable and sufficiently consistent set of information objects owned by single human beings (cf. [10]). Knowledge represented in a language is called data [10]. Information is a knowledge increment brought about by receiving data, by observing reality, or by inner thinking processes by which a human being organizes, compares and assesses her/his knowledge (cf. [10]). Information processing means actions by which information is created, collected, stored, processed, presented, disseminated and interpreted. Next, we elaborate the notion of information processing by considering it (1) as a context, (2) on a specific layer, and (3) in relation to other contexts (Figure 1). The discussion is based on the ontological framework, called OntoFrame [30], which contains the ontologies for the context (the context ontology), the layers (the layer ontology) and the perspectives (the perspective ontology). In the following we first give a brief introduction to the first two ontologies and then define the perspective ontology in Section 2. Context Ontology
Information Processing
Layer Ontology
P T
ME
Ar An
L
OS
F
IPS
US
O
ISD IS
Perspective Ontology S
P
D
I C
Figure 1. Framework to elaborate the notion of information processing
To have a better understanding of information processing it should be considered as something that comprises not merely actions and targets of actions but also actors, motivations, facilities and so on. Shortly, information processing should be seen as a context with all its contextual features. We have defined, based on case grammar [12], pragmatics [36], and activity theory [8], the contextual approach and the context ontology in [30]. The contextual approach has been earlier applied to enterprises [31], ISD [35], method integration [34] and method engineering [32]. Here, we apply it to information processing in general. According to the contextual approach any context can be conceived through concepts and constructs which belong to eight contextual domains: purpose, actor, action, object, facility, location and time. The domains are defined as follows:
260
M. Leppänen / A Perspective Ontology and IS Perspectives
x
Purpose domain consists of those concepts and constructs which refer, directly or indirectly, to goals, motives, or intentions of someone or something. x Actor domain encompasses those concepts and constructs, which refer to individuals, groups, positions, roles, or organizations. x Action domain is composed of those concepts and constructs which refer to functions, activities, tasks, or operations carried out in the context. x Object domain comprises those concepts and constructs which refer to something which an action is targeted to. The objects can be goods or services, material or informational. x Facility domain consists of those concepts and constructs which refer to means, whether a tool or a resource, by which something can be done or is done. x Location domain is composed of those concepts and constructs which refer to parts of space occupied by someone or something. The location is physical, like a room or a building, or logical, like a site in a communication network. x Time domain includes those concepts and constructs which refer to temporal aspects in the context. In addition to the domain-specific concepts and relationships presented above there are a number of inter-domain relationships. Second, we consider information processing on three layers that are information system, information systems development and method engineering. An information system (IS) is a context which provides information to its utilizing system. An information systems development (ISD) means a context which carries out ISD actions, ranging from requirements engineering to implementation and evaluation of an IS, in order to contribute to a renewed or a new IS. A method engineering (ME) means a context which performs ME actions to develop, customize, configure and implement a new or an improved ISD method. A context, here called by the general term information processing system (IPS), is always associated to two other contexts, namely (a) the one which information is about, and (b) the one which utilizes information provided by the IPS. These contexts are called the object system (OS) and the utilizing system (US), correspondingly. The OS and the US have specific meanings depending on the layer on which the IPS is. On the ME layer, the IPS produces prescriptions for the next lower layer (ISD) to facilitate it to produce, efficiently and effectively, prescriptions for the lowest layer (IS) so that it can satisfy the goals and needs of its utilizing system (USIS). Thus, the USME consists of those ISD’s, IS’s, and USIS’s that are related to the IPS. Correspondingly, the USISD is composed of the related IS’s and their utilizing systems. Information objects at the ME layer refer to the prior ISD contexts and the current ISD, as well as to their US’s and OS’s (i.e. USISD and OSISD). The prior ISD contexts mean those ISD contexts in which the method under construction has been earlier deployed. The object system of the ISD (OSISD), in turn, comprises the existing IS and a new IS, as well as their US’s and OS’s (i.e. USIS and OSIS). To conclude, the IPS means a context, which is highly related to the two other contexts (OS, US), is located on some of the three processing layers, and is conceptualized through the concepts and constructs of seven contextual domains. 2. Perspective Ontology In this section we first define the general notions of perspective and system of perspectives, then specify five perspectives, and lastly show how these perspectives are applied on three processing layers.
261
M. Leppänen / A Perspective Ontology and IS Perspectives
2.1
A System of Perspectives
Due to the complexity of reality, a human being tends to focus on some specific aspects, depending on a point of view adopted. In everyday life, a point of view can be situational and intuitive, established in an ad hoc fashion. However, for recurrent situation it is necessary to have structured and pre-defined viewpoints. Especially this holds for information processing systems such as ISD and ME in which abstract thinking is commonplace and which involve a large number of people in close cooperation. We define a perspective to mean such a strictly defined point of view and a system of perspectives to stand for perspectives with specified relationships. The perspective ontology provides a system of well-defined perspectives established on certain dimensions to conceive aspects of information processing in an organized manner (Figure 2).
Framework
System of perspectives 1..*
*
1..* UoD
1..*
1 conceivedFrom 1..*
Criterion 1 conceivedFrom 1..* 1..*
basedOn 1 1
1..*
1..* Point of view
Systelogical
Perspective
Infological
Conceptual
establishedOn 1..*
Datalogical
Dimension
Physical
Figure 2. Perspective ontology
The perspectives should be defined in a way which (a) enables decisions on which aspects are relevant and which aspects should be ignored from each of the perspectives, (b) relates the perspectives to one another in a rigorous manner, and (c) applies to a structured consideration of information processing on the IS, ISD and ME layers. Perspectives are commonly defined based on certain criteria or principles derived from theories relevant to the domain. The IS literature recognizes three theories to be such ones, namely semiotics (e.g. [23]), systems theory (e.g. [21, 43, 44]) and formal logic (e.g. [40]). Here, we first apply semiotics [47]. It distinguishes between linguistic expressions, on the one hand, and conceptual constructs signified by the expressions, on the other hand. This division results in a dichotomy-like dimension which has two ends, linguistic and conceptual. Second, complexity makes it difficult to perceive and understand information processing, if not decomposed and specialized into more perceivable parts [28, 43]. Decomposition and specialization are principles inverse to the first-order abstraction [33]. These two principles form our second dimension. The third dimension is based on the predicate abstraction with the criterion of realization independence [33]. It enables the partitioning of the predicates of information processing into predefined sets depending on how closely the predicates are related to realization. At the one end, information processing is viewed to be fully independent from realization, while at the other end one concentrates, in particular, on physical predicates, including those of individual persons and groups, detailed procedures, concrete data files and documents in certain spatiotemporal space.
262
M. Leppänen / A Perspective Ontology and IS Perspectives
To conclude, the system of perspectives in the perspective ontology is established on three dimensions: (a) the linguistic – conceptual dimension, (b) the first-order abstraction dimension, and (c) the predicate abstraction dimension. Figure 3 presents the perspectives along the three dimensions in relation to the IPS, the US and the OS. In the next section we define the perspectives and discuss them on the basis of this figure.
Figure 3. Dimensions and perspectives
2.2. Definitions of the Perspectives The system of perspectives is composed of five perspectives that are: systelogical, infological, conceptual, datalogical, and physical perspectives. The term ‘systelogical’ was introduced in [60], although in a slightly different meaning. The terms ‘infological’ and ‘datalogical’ were originally coined in [54] and [29] about in the same meanings as we use them here. Because we consider the perspectives as parts of the generic ontology, we define them in general in relation to the information processing system (IPS). According to the systelogical perspective the IPS is considered in relation to its utilizing system (US). The IPS has no value or purpose by itself. It becomes desired and necessary through the support it provides to its utilizing system. Hence, organizational, social, economic and informational impacts of the IPS on the utilizing system form the essence which the systelogical perspective is interested in. The generic question to be answered from this perspective is “Why”. To put it more precisely, applying the systelogical perspective means considering the following issues: x Why does the IPS exist? x What kind of utilizing system does it have? What are its objectives, actors, actions, events, rules and objects on a general level? x What information services does, or should, the IPS provide, for whom and for which actions in the US? According to the infological perspective the IPS is seen as a functional structure of information processing actions and information objects, independent from any representational and implementational features. The IPS is regarded as a context, in which the given mission is pursued by actions related to one another through information flows. The generic question to be answered is “What”. This means in more detail:
M. Leppänen / A Perspective Ontology and IS Perspectives
263
x What information is processed in the IPS and why? x What are the actions and rules of information processing? According to the conceptual perspective the IPS is considered through the semantic contents of information it processes. This means that whereas the infological perspective is based on linguistic terms, the conceptual perspective concentrates on the understanding of the meaning of those things in the OS which linguistic terms signify. The question to be answered is “What does it mean?” To put it more precisely, the conceptual perspective is interested in the following issues: x What is the meaning of the information processed in the IPS? x What does the information signify? x What kinds of structural and dynamic constraints are valid in the OS? From the datalogical perspective the IPS is viewed, through representation-specific concepts, as a context, in which actors work with facilities to process data. This implies that the perspective makes a difference between two parts: a human information processing system (HIPS) and a computerized information processing system (CIPS). Considerations cover all those contextual non-physical phenomena that are relevant to the execution of data processing actions within and between those parts (cf. user interface). The datalogical perspective is interested in “How” questions such as: x How is information represented in data in the IPS? x How are the rules of information processing derived from US rules and formulated into concrete work procedures and algorithms? x How do the users and the CIPS communicate with each other? The physical perspective ties the datalogical concepts and constructs to a particular organizational and technical environment, showing how the IPS looks like and behaves when it is implemented. It answers, for example, the following questions: x Who are those actors carrying out actions in the HIPS, how and when they act, and where are they located? x Where and how are the data stored? x How are the facilities used and by whom? x What hardware and software are used, and how are they related? Now we can discuss more closely the relationships between the perspectives and the dimensions (see Figure 3). The systelogical perspective provides the point of departure for considerations about the IPS. The main focus of this perspective is on the US, and the IPS is viewed as something which only provides services for its US. Changing the perspective from systelogical to infological means a shift along the first-order abstraction dimension: the IPS seen as a “black box” is now conceived as a context that is composed of purposes, actions and objects. Compared to the infological perspective, the application of the datalogical perspective and the physical perspective means moves along two dimensions, along the first-order dimension, on one hand, and along the predicate abstraction dimension, on the other hand. The purposes, the actions, and the objects are, in the first stage, decomposed and specialized into smaller “pieces”. In addition, actors and facilities are, on a general level, recognized. In the second stage, the process of decomposing and specializing continues and more and more realization-specific aspects of the IPS and its components are recognized. The three perspectives (i.e. the infological, datalogical, and physical perspectives) constitute a “hierarchical system of stratified levels” as defined by Mustonen [44]. The conceptual perspective is based on the use of the linguistic - conceptual dimension. While the other perspectives consider linguistic objects, the conceptual perspective focuses on their conceptual contents.
264
M. Leppänen / A Perspective Ontology and IS Perspectives
2.3. Perspectives on the Processing Layers The perspectives were above defined on a general level with the intention that they apply to any processing layer. Table 1 summarizes how they can be elaborated for each of the processing layers. Because it is not possible here to go into details, we only give some general comments on the table. We can see that regardless of which layer is concerned the systelogical perspective means looking from the viewpoint of the utilizing system, the infological perspective considers the intentions, actions and objects of the IPS, and the conceptual perspective is interested in the contents of the information objects of the IPS. Furthermore, the datalogical perspective elaborates the conceptions about the intentions, actions and objects, and extends to consider actors and facilities of the IPS as well. The physical perspective covers also the physical features of the IPS. In the next section we consider the perspectives on the IS layer, called the IS perspectives, more closely. The ISD perspectives and the ME perspectives are discussed more deeply in [30]. Table 1. Perspectives on three processing layers (L = Layer) L ME
ISD
IS
Systelogical Considers what services the ME provides to the USME (i.e. the ISD’s, the IS’s and the USIS). Considers what services the ISD provides to the USISD (i.e. the IS and the USIS). Considers services the IS provides to the USIS.
Infological Considers intentions, functional structures and information objects in the ME. Considers intentions, functional structures and information objects in the ISD. Considers intentions, functional structures and information objects in the IS.
Conceptual Considers the semantic contents of information objects in the ME (i.e. the OSME composed of the ISD’s, the IS’s and the OSIS’s). Considers the semantic contents of information objects in the ISD (i.e. the OSISD composed of the IS’s and the OSIS’s). Considers the semantic contents of information objects in the IS (i.e. the OSIS).
Datalogical Considers ME actors, ME actions, ME objects, ME facilities and their interplay on a general level. Considers ISD actors, ISD actions, ISD objects, ISD facilities and their interplay on a general level. Considers IS actors, IS actions, IS objects, IS facilities and their interplay on a general level.
Physical Considers the ME as a physical and technical construct in organizational and technical environment. Considers the ISD as a physical and technical construct in organizational and technical environment. Considers the IS as a physical and technical construct in organizational and technical environment.
3. IS Perspectives In this section we define concepts and constructs through which the IS can be perceived from the IS systelogical, IS infological, IS conceptual, IS datalogical, and IS physical perspectives. The emphasis of our discussion is on the first three perspectives. Defining the IS perspectives gives a concrete example of how to apply the perspective ontology. 3.1 IS Systelogical Perspective From the IS systelogical perspective the IS is seen to be in relation to its utilizing system (USIS). The utilizing system may be a business system such as a manufacturing department, or a public organization such as a library maintaining and lending copies of publications. There are several approaches to viewing the utilizing system (e.g. an enterprise modeling view [24, 38], a business process modeling view [41, 42, 48] with business rules [20], or an
265
M. Leppänen / A Perspective Ontology and IS Perspectives
organizational communication view (e.g. [6]). Each approach applies different concepts and constructs to conceive and structure things in the utilizing system. We integrate the IPO (Input-Process-Output) approach, commonly applied in enterprise modeling and business process modeling, with the main concepts of the purpose domain and the actor domain. Depending on the nature of the IS, we have two somewhat different viewpoints on the USIS. If the IS is a computerized information system (CIS), the IS is seen as a tool used in the USIS. If the IS contains a human information system (HIS) as well, the IS is seen as a context providing information services to the USIS. Here, we model the IS systelogical perspective from the former viewpoint (the tool viewpoint) (see Figure 4). US organization
1 1..*
US actor
US org. unit
1 1..* 1..* US human actor
US event
*
raisedBy
US position
1..*
1 supervision
1..*
US role
*
*
*
*
occupiedBy
US condition
US rule
1..* 1..* governs 1..* *
* 1..* * responsibleFor
strivesFor 1..*
US action
*
*
1..*
uses
* input output 1..* 1..* User
* supports
US purpose
performs
US object
*
US tool
*
US resource
*
Material
Informational
CIS
US facility
Figure 4. IS systelogical perspective (the tool viewpoint)
A US organization is an organization (i.e. an enterprise, a department or some other administrative arrangement), which utilizes, or is going to utilize, an IS. It consists of US organizational units, which in turn are composed of US positions. A US position is a post of employment occupied by one or more US human actors. US positions are composed of US roles with responsibilities and authorities to conduct certain US actions. A US action is an action, which strives for one or more utilization purposes. US actions are governed by US rules. US rules are composed of certain parts in accordance with the so-called ECAA structure [20]: US event, US condition, thenUSAction and elseUSAction. Conducting US actions may raise new US events that possibly trigger other US actions. The US purposes mean goals for business processes and/or reasons for setting up those goals. The US actions use US objects as their inputs and may produce US objects as their outputs. The US objects can be material (e.g. machines, components, bridges and china) or informational (e.g. insurance contract, payment and reorder). The US actions are partly performed by US tools (e.g. lathe, circular saw and nailer). Some of the US tools are computerized information systems (CIS) supporting US actions. A US actor conducting US actions with the support of a CIS is called a user. The US actions consume US resources, such as money, energy, goods, and manpower.
266
M. Leppänen / A Perspective Ontology and IS Perspectives
3.2. IS Infological Perspective In IS modeling there are two prevailing approaches, the structured approach [61] and the object-oriented approach [e.g. [2]). Here, we apply the structured approach. Based on that the IS infological perspective sees the IS to be a functional structure of information processing actions and information objects (see Figure 5). No attention is given to how the information objects are represented or implemented. This means that the “black box” conceived from the IS systelogical perspective is “opened” to reveal the aspects of the IS within three contextual domains: purpose, action, and object. The concepts in the purpose domain are used to specify why information is processed. The concepts in the action domain are used to conceive action structures needed to produce information objects. Correspondingly, the information objects are decomposed, classified, and structured with the concepts and relationships in the object domain. Sequence str
Selection str
Iteration str.
Control str.
Decomposition str.
influence *
*
IS goal
predAbstract 0..1
versionOf *
IS purpose
* 0..1
1..* *
IS object supports * *
input * output * copyOf * 0..1
IS action *
* refinement * dueTo * *
IS action str. * strivesFor 1..* * raisedBy
IS reason
* IS event
* governs * IS rule
Transient
IS condition
Permanent
Figure 5. IS infological perspective
IS purposes mean IS goals for information processing and/or reasons for setting up those goals. An IS goal is a desired state of affairs in the IS [38]. IS reasons can be functional or non-functional requirements for information processing, problems in prevailing information processing, strengths and weaknesses in, and/or opportunities for and threats against the existing or a planned IS. The IS goals are related to one another through complex influence and refinement relationships [24]. In striving for the IS purposes, IS actions use information objects, called IS objects, as inputs and produce IS objects as outputs. The range of various types of IS actions is large. An IS action can mean, for instance, collecting, storing, processing, transmitting, coding, encoding, arranging, locating, discovering, interpreting, integrating, reviewing, testing, approving, or editing information. The action structures relevant from the IS infological perspective are the decomposition structure and the control structures. The decomposition structure splits IS actions into IS functions, IS activities, IS tasks, and IS operations. The control structures enable to recognize sequence, selection and iteration relationships between the IS actions. The IS actions are governed by IS rules. An IS rule is composed of IS events, IS conditions, thenISActions and elseISActions. The IS rules can be classified in many ways.
M. Leppänen / A Perspective Ontology and IS Perspectives
267
First, there are dynamic and static rules. The dynamic IS rules restrict or guide IS actions and IS events. The static IS rules restrict IS objects. Examples of the IS rules are: back-ups of the files should be run once a week; a social security number of a person cannot be changed; a salary of an hourly paid employee is derived by the rule ‘Salary := number of hours x hourly fee’. The first example is a business rule, the second rule is an integrity constraint [7], and the last rule is an example of a derivation rule. An IS object is in the form that is free from any representational and implementational aspects. An IS object is transient or permanent. A transient IS object lasts only a short time (e.g. a reply to a routine request). A permanent IS object is valuable enough to “live” longer (e.g. personnel information, vehicle information). The IS objects are interrelated in many ways. They are composed of other IS objects. Producing them is supported by other IS objects (cf. derivation of the monthly salary from hourly fee and number of hours). An IS object can also be a version of, a copy of, or an (predicate) abstraction from, another IS object. 3.3. IS Conceptual Perspective The IS conceptual perspective considers the semantic contents of the IS objects, meaning that the structure and behavior of those things in the OSIS which are signified by the IS objects are revealed. Thus, the IS conceptual perspective addresses the so-called deep structure of the IS [59]. There are several approaches to OSIS modeling. Some of them are structural, such as the ER approach [5] and the ORM approach [17], some others (e.g. the object-oriented approach [2]) cover dynamics of the OSIS as well. We prefer the ER approach to the ORM approach and other attribute-free approaches, because we consider it important to separate between entities and attributes. We also want to make a clear distinction between the static features and the dynamic features of the OSIS, unlike the object-oriented approach. Hence, the IS conceptual perspective is based on the ER approach (the structural view) and the state machine (the dynamic view) (Figure 6). According to it, the OSIS is composed of related things that are either entities or relationships having states and affected by state transitions.
Figure 6. IS conceptual perspective
An entity means any perceivable thing in the object system with an independent existence (cf. [7]). Only those things that are relevant and “independent” enough to be signified by
268
M. Leppänen / A Perspective Ontology and IS Perspectives
the IS objects are regarded as entities. An OS relationship between two or more entities means any relevant connection, association or the like between entities. The OS relationships include the abstraction relationships (e.g. classification, generalization, composition, grouping) defined in the abstraction ontology [33]. An attribute is a relevant predicate used to characterize an entity or an OS relationship. A particular entity (and OS relationship) has zero, one or more attribute values for each of its attributes. An OSIS construct means a conceptual construct composed of specific entities related to one another through OS relationships and characterized by specific attribute values. An OSIS state means a state of the object system or its parts, composed of OSIS constructs. An OSIS transition means a transition from one OSIS state, called the pre-state, to another OSIS state, called the post-state [10]. An OSIS transition can involve entities (e.g. the birth of a child), OS relationships (e.g. the divorce) and/or attribute (e.g. the quantity available). The transitions constitute the potential OSIS behavior. OSIS transitions can be composed to establish OSIS transition structures. An OSIS event means an event which may trigger an OSIS transition from the pre-state to the post-state and which may be caused by another OSIS state transition. 3.4. IS Datalogical Perspective and IS Physical Perspective From the IS datalogical perspective the IS is viewed, through representation-specific concepts, as a context, where IS actors work with IS facilities to process IS data. Thus, the IS objects, seen as information objects from the IS infological perspective, are here considered to be data objects represented in some non-formal, semi-formal or formal language(s). There are also special IS actions which transform data objects from one form to another. Although no reference is made to data carriers or other physical features of the IS, the IS datalogical perspective enables to make a difference between a human information system (HIS) and a computerized information system (CIS). To conceive the interaction between and cooperation among these two parts, we also distinguish user interface (UI). Each of these parts is conceptually quite large. Due to the scarcity of space, we only present the model of the IS datalogical perspective in Appendix (Figure A.1). The IS physical perspective considers the IS with all its physical aspects. It ties the IS datalogical concepts and constructs to a particular organizational and technical environment, showing how the IS looks like and behaves when it is implemented. The IS contains the HIS, and possibly the CIS and the UI. For all these parts, a highly detailed and realization-dependent view is provided by this perspective. Figure A.2 in Appendix presents concepts and constructs referring to a part of the CIS from the physical perspective. 3.5. Relationships between the IS Perspectives In the previous sections we have considered the contextual concepts and relationships within each of the IS perspectives. Here, we relate the IS perspectives to one another through main inter-perspective relationships. The perspectives have been established along three dimensions. Based on the discussion in Section 2.1, we can describe the relationships between the IS perspectives as shown in Figure 7. The small rectangles inside the IS systelogical, IS infological, IS datalogical and IS physical perspectives stand for information objects which signify conceptual constructs in the object system (cf. the IS conceptual perspective). The common denominator between the IS systelogical perspective and the IS infological perspective is the IS, implying that moving from the former perspective to the latter means that the IS, first seen as a black box,
M. Leppänen / A Perspective Ontology and IS Perspectives
269
is “opened” in order to expose IS purposes, IS actions, IS objects and relationships between them. In this process, the principles of decomposition and specialization are mainly applied.
Figure 7. Relationships between the IS perspectives
The IS infological, IS datalogical and IS physical perspectives are parts of a hierarchical system of perspectives within which the relationships are based on the same criterion of realization dependence. This means that in moving downwards in the hierarchy the conceptions about the IS first become representation-specific (cf. the IS datalogical perspective) and then implementation-specific (cf. the IS physical perspective). In parallel to this, the conceptions of the IS are concretized by decomposition and specialization. Each of the aforementioned IS perspectives recognizes information objects. In the IS systelogical perspective they are called informational US objects. The IS infological perspective identifies the IS objects, and the IS datalogical perspective views them as digital or non-digital data objects. Data files, data records and data fields represent the conceptions of IS objects from the IS physical perspective. In all those cases, there are ‘signifies’ relationships between the information objects and the things conceived as OSIS constructs from the IS conceptual perspective. Through these relationships it is possible to make sense of the meanings of the information objects. Still one type of a generic relationship can be found between the IS perspectives. If the OSIS overlaps with the USIS or the IS, there is ‘abstractedFrom’ relationships between the OSIS and the USIS, in the first case, and between the OSIS and the ISIS, in the second case. By this abstraction, most of the contextual aspects of the USIS (ISIS) are ignored in order to establish OSIS constructs composed of entities, OS relationships and attribute values. For example, US actions such as hiring and firing an employee are abstracted to OS events affecting on the OS state of a particular employee.
270
M. Leppänen / A Perspective Ontology and IS Perspectives
4. Comparative Analysis of Current IS Perspectives The IS literature provides a large variety of IS architectures (e.g. [53, 62]), IS frameworks (e.g. [21, 45, 46, 55]), reference models (e.g. [22]) and IS meta models (e.g. [13]) that are based on views, perspectives and viewpoints on the IS. For simplicity, we call these presentations the frameworks, and the views and viewpoints the perspectives. The purpose of this section is to make a comparative analysis of frameworks to find out which kinds of perspectives, underlying criteria and dimensions they propose and how they relate to our IS perspectives. For the analysis, we selected those frameworks (a) which have been developed for a comprehensive analysis and/or comparison of the concepts in the fields of IS and/or ISD, and (b) in which systems of perspectives have been clearly specified. These frameworks are (in temporal order): Welke [60], Olive [45], Essink [9], Olle et al. [46], Iivari [21], Sol [51], Sowa et al. [53], van Swede et al. [55], Freeman et al. [13], Avison et al. [1] and ISO [22]. Table 2 summarizes the results of the analysis. The dimensions, along which the perspectives have been established in the frameworks, are called “levels of abstraction” [45, 9, 21, 13], “design levels” [53], “aspects” [46], ”views” [1], “viewpoints” [22], “subproblems” [51]) or “perspectives” [55, 60]. The frameworks apply different criteria to distinguish between and relating the perspectives. Sowa et al. [53] and van Swede et al. [55] have based their perspectives on views of stakeholders. Iivari [21] has derived his levels of abstraction from abstractions of the host organization, the universe of discourse, and technology. Welke [60] has built the perspectives on consequences that changes in the existing IS result in the object system, the use of information, and the data processing sequences. Avison et al. [1] argue that their five views are needed to answer the vital questions of users. Freeman et al. [13] compare their levels of abstraction with the phases of software development. Sol [51] has made his division on the basis of the kinds of problems that must be solved during the ISD. Some frameworks (i.e. [45, 22]) give neither explanation for the perspectives nor apply any explicit underlying criteria. The correspondences of the perspectives in the frameworks to our perspectives are indicated by the markings ‘X’ (strong) and ‘x’ (weak) in Table 2. The following remarks can be made on them. In all the frameworks the upper perspectives relate to the US and the lower perspectives are more technology-specific. The perspectives between the extreme ends are defined through “independence” from something (e.g. from “the object system” in [45]), and from the technology in [21]). The frameworks differ from one another in the emphasis they give on the perspectives. Welke [60] and van Swede et al. [55], for instance, focus on the upper perspectives in their frameworks, and Avison et al. [1] and ISO [22] suggest more perspectives for the consideration of technological issues. Conceptual issues are included in the topmost perspectives ([9], and partly in [46]), or as it is the case more commonly, in the next lower perspectives. The framework of [45] is the only one providing a special perspective for perceiving the conceptual aspects of the IS. The framework of van Swede et al. [55] does not to consider IS conceptual issues at all. Iivari [21], Avison et al. [1], Freeman et al. [13], Sowa et al. [53] and Olle et al. [46] pay special attention to user interface. Essink [9] mentions it only incidentally, and the frameworks of [60] and [51] are too general to recognize it. In our view, it is important to have separate perspectives for each set of different aspects of the IS. Therefore, the IS systelogical perspective is needed to consider the IS in relation to the US. The IS conceptual perspective is necessary to address the conceptual contents of the IS objects. Unlike Avison et al. [1], Sowa et al. [53], Sol [51], Freeman et al. [13] ISO [22], we see it vital to clearly differentiate between the infological perspective that represents the “linguistic world”, and the IS conceptual perspective which stands for the “conceptual world”. On the predicate abstraction dimension, at least three related
271
M. Leppänen / A Perspective Ontology and IS Perspectives Table 2: A summary of the comparative analysis of the perspectives (S = systelogical, I = infological, C = Conceptual, D = Datalogical, P = physical) Frameworks Welke [60] - Systelogical perspective - Infological perspective - Datalogical perspective Olive [45] - External level - Conceptual level - Logical level - Architectural level - Physical level Essink [9] - Object system modelling - Conceptual IS modelling - Data system modelling - Implementation modelling Olle et al. [46] - Business analysis stage - System design stage - Construction design stage Iivari [21] - Organizational level - Conceptual/infological level - Datalogical/technical level Sol [51] - Systelogical problems - Infological problems - Datalogical problems - Technological problems Sowa et al. [53] - Scope level - Enterprise model level - System model level - Technology model level - Components level van Swede et al. [55] - Business perspective - Information perspective - Functionality perspective - Implementation perspective Freeman et al. [13] - World level - Conceptual level - Design level - Implementation level ISO [22] - Enterprise viewpoint - Information viewpoint - Computational viewpoint - Engineering viewpoint - Technology viewpoint Avison et al. [1] -Human-activity view -Information view -Socio-technical view -HCI-view -Technical view
Criteria Changes in IS and/or its usersubsystem should be addressed from several perspectives ([60] p. 150) “The model at the highest level is the most general and those at the lower level are more detailed” ([45] p. 63)
S
I
C
X
x X
x
X
D
P
x X
x x
x X X X X
“..are classes of problems that are relevant from a specific view on IS’s” ([9] p. 356)
x
X X
x X X
Not clearly specified X
X X X
Derived from abstractions of the host system, the UoD and technology
X X
X
x X
X
Not clearly specified X X
x X X
Levels correspond to views of specific stakeholders
X x X
x x
X X X
Perspectives correspond to views of specific groups of people
“..loosely corresponds to the phases of software development” ([13] p. 287)
X X
x x
x X
X
X X
X X X
“..to focus on particular concerns within a system” ([22] Section 3.2.7)
“..necessary to form a system which is complete in both technical and human terms”
X X
X X
X X X
X X
X X X
X X
X
272
M. Leppänen / A Perspective Ontology and IS Perspectives
perspectives can be clearly distinguished. The first one (infological) is independent from representational and realization-dependent aspects. The second one (datalogical) is independent from realization-dependent aspects. The third one (physical) recognizes all the concrete issues related to a specific realization. 5. Summary and Conclusions In this paper we have presented a light-weight perspective ontology, which defines and organizes concepts and constructs by which diverse aspects of information processing can be categorized and examined according to the perspective(s) depending on the problem at hand. The ontology defines five perspectives established on the well-defined criteria and dimensions. The perspective ontology has been anchored on relevant theories (i.e. case grammar [12], pragmatics [36], activity theory [8], semiotics [47] and systems theory [43]) and engineered in accordance with the guidelines of ontology engineering [11, 58]. The paper has demonstrated how the perspectives can be applied in the contexts of IS, ISD and ME. To further concretize the conceptions of the IS perspectives, the basic IS concepts and IS constructs have been modelled and defined for each of the IS perspectives. The IS perspectives have also been used as a framework in a comparative analysis of relevant works in the IS literature. To our best knowledge there are no earlier suggestions for a perspective ontology with the same kind of purpose. The term “perspective ontology” does exist but in different meanings. For instance, in the philosophy there is the well-known Zhuang Zi’s perspective ontology according to which the being and identity of an entity is contextually situated and perspective-dependent [37]. On a general level, perspectives are discussed as levels of abstraction by several researchers (e.g. [23, 43, 44]) but they do not go into detail as we do in this paper. The perspective ontology can be deployed as groundwork to elaborate current frameworks and frames of reference, especially when it comes to the underlying criteria and dimensions. The perspectives can also be used as a foundational structure in analyzing the conceptual contents of ISD methods, and in engineering new methods by integration, adaptation and customization. Furthermore, they can be applied to recognize and categorize diverse contingency factors related to the IS, ISD and ME. In future research we concentrate on the refinement of a so-called perspective-based approach to method engineering [32] in which the perspective ontology is used as a cornerstone for a conceptual framework of ME. References [1] [2] [3] [4] [5] [6] [7]
Avison, D., Wood-Harper, A., Vidgen, R. & Wood, R. 1996. Multiview: a further exploration in information systems development. Maidenhead: McGraw-Hill. Booch, G., Rumbaugh, J. & Jacobson I. 1999. The Unified Modeling Language – user guide. Reading: Addison-Wesley. Bunge, M. 1977. Treatise on basic philosophy, Vol. 3: Ontology I: The furniture of the world. Dortrecht: D. Reidel Publishing Company. Burton-Jones, A., Storey, V., Sugumaran, V. & Ahluwalia, P. 2005. A semiotic metric suite for assessing the quality of ontologies. Data & Knowledge Engineering 55(1): 84-102. Chen, P. 1976. The entity-relationship model – toward a unified view of data. ACM Trans. On Database Systems 1(3): 9-36. Dietz, J. 2003. The atoms, molecule and fibers of organizations. Data & Knowledge Engineering 47(3): 301-325. Elmasri, R. & Navathe, S. 2000. Fundamentals of database systems. 3rd edition, Reading: AddisonWesley.
M. Leppänen / A Perspective Ontology and IS Perspectives [8] [9] [10] [11] [12] [13] [14] [15] [16] [17] [18] [19] [20] [21] [22] [23] [24] [25] [26] [27] [28] [29] [30] [31] [32] [33] [34] [35]
273
Engeström, Y. 1987. Learning by expanding: an activity theoretical approach to developmental research. Helsinki: Orienta-Konsultit. Essink, L. 1988. A conceptual framework for information systems development methodologies. In H. J. Bullinger et al. (Eds.) Information Technology for Organizational Systems. Amsterdam: Elsevier Science Pub., 354-362. Falkenberg, E:, Hesse, W., Lindgreen, P., Nilsson, B., Oei, J. L. H., Rolland, C., Stamper, R., van Asche, F., Verrijn-Stuart, A. & Voss, K. 1998. A framework of information system concepts, The FRISCO Report (Web edition), IFIP. Fernandez-Lopez, M., Gomez-Perez, A., Pazos-Sierra, A. & Pazos-Sierra, J. 1999. Building a chemical ontology using METONTOLOGY and the ontology design environment. IEEE Intelligent Systems & Theory Applications 4(1): 37-46 Fillmore, C. 1968. The case for case. In E. Bach & R. T. Harms (Eds.) Universals in Linguistic Theory. New York: Holt, Rinehart and Winston, 1-88. Freeman, M. & Layzell, P. 1994. A meta-model of information systems to support reverse engineering. Information and Software Technology 36(5): 283-294. Gruber, T. 1993. A translation approach to portable ontology specification, Knowledge Acquisition 5(2): 119-220. Gruber, T. 1995. Towards principles for the design of ontologies used for knowledge sharing. International Journal of Human-Computer Studies 43(5/6): 907-928. Gupta, D. & Prakash, N. 2001. Engineering methods from method requirements specifications. Requirements Engineering 6(3): 135-160. Halpin, T. 1998. ORM/NIAM Object-Role Modelling. In P. Bernus, K. Mertins & G. Schmidt (Eds.) Handbook on Information Systems Architecture. Berlin: Springer-Verlag, 81-101. Harmsen, F. 1997. Situational method engineering. University of Twente, Moret Ernst & Young Management Consultants, The Netherlands, Dissertation Thesis. Hautamäki, A. 1986. Points of views and their logical analysis. Helsinki: Acta Philosophica Fennica, Vol. 41. Herbst, H. 1995. A meta-model for business rules in systems analysis. In J. Iivari, K. Lyytinen & M. Rossi (Eds.) Advanced Information Systems Engineering. LNCS 932, Berlin: Springer, 186-199. Iivari, J. 1989. Levels of abstraction as a conceptual framework for an information system. In E. Falkenberg & P. Lindgren (Eds.) Information System Concepts: An In-Depth Analysis. Amsterdam: Elsevier Science Pub., 323-352. ISO 1996. Information Technology – Open Distributed Processing - Reference Model: Overview, 10746-1. Kangassalo, H. 1980. Structuring principles of conceptual schemas and conceptual models. Report A85, Department of Mathematical Sciences, University of Tampere, Finland. Kavakli, V. & Loucopoulos, P. 1999. Goal-driven business process analysis application in electricity deregulation. Information Systems 24(3): 187-207. Kelly, S., Lyytinen, K. & Rossi, M. 1996. MetaEdit+: a fully configurable multi-user and multi-tool CASE and CAME environment. In Y. Vassiliou & J. Mylopoulos (Eds.) Proc. of the 8th Conf. on Advanced Information Systems Engineering (CAiSE’96). Berlin: Springer, 1-21. Kinnunen, K. & Leppänen, M., 1996. O/A matrix and a technique for methodology engineering. Journal of Systems and Software 33(2): 141-152. Kruchten, P. 1995. Architectural blueprints – The “4+1” view model of software architecture. IEEE Software 12(6): 42-50. Langefors, B. 1971. Theoretical analysis of information systems. Lund, Sweden: Studentlitterature. Langefors, B. & Sundgren, B. 1975. Information systems architecture. New York: Petrocelli. Leppänen, M. 2005. An ontological framework and a methodical skeleton for method engineering, Dissertation thesis, Jyväskylä Studies in Computing 52, University of Jyväskylä, Finland. Leppänen, M. 2005. A context-based enterprise ontology. In G. Guizzardi & G. Wagner (Eds.) Proc. of the EDOC International Workshop on Vocabularies, Ontologies and Rules for the Enterprise (VORTE’05), Enschede, The Netherlands, CTIT Workshop Proceedings, 17-24. Leppänen, M. 2006. Conceptual evaluation of methods for engineering situational ISD methods. Software Process: Improvement and Practice 11(5): 539-555. Leppänen, M. 2007. Towards an abstraction ontology. In Y. Kiyoki, H. Kangassalo & M. Duži (Eds.) The 16th European – Japanese Conference on Information Modelling and Knowledge Bases (EJC 2006), to be printed by IOS in the Series of the “Frontiers on Artificial Intelligence”. Leppänen, M. 2007. A contextual method integration. In Proc. of the 15th Int. Conf. on Information Systems Development (ISD 2006). Springer-Verlag, Berlin (in print) Leppänen, M. 2007. Towards an ontology for information systems development – A contextual approach. In K. Siau (Ed.) Contemporary Issues in Database Design and Information Systems Development. Idea Group Inc. (in print)
274 [36] [37] [38] [39] [40] [41] [42] [43] [44] [45] [46] [47] [48] [49] [50] [51] [52] [53] [54] [55] [56] [57] [58] [59] [60] [61] [62]
M. Leppänen / A Perspective Ontology and IS Perspectives Levinson, S. 1983. Pragmatics, London: Cambridge University Press. Li, C. 1999. The Tao encounters the West. Explorations in comparative philosophy. Albany: State University of New York Press. Loucopoulos, P., Kavakli, V., Prekas, N., Rolland, C., Grosz, G. & Nurcan, S. 1998. Using the EKD approach: the modeling component. ELEKTRA – Project No. 22927, ESPRIT Programme 7.1. Lyons, J. 1977. Semantics. Volume I-II, Cambridge: Cambridge University Press. Maibaum, T.S. 1986. Role of abstraction in program development. In H. Kugler (Ed.) Information Processing 86, Amsterdam: Elsevier Science Pub, 135-142. Melao, N. & Pidd, M. 2000. A conceptual framework for understanding business processes and business process modeling. Information Systems Journal 10(2): 105-129. Mentzas, G., Halaris, C. & Kavadias, S. 2001. Modeling business processes with workflow systems: an evaluation of alternative approaches. International Journal of Information Management 21(2): 123-135. Mesarovic M., Macko D. & Takahara Y. 1970. Theory of hierarchical, multilevel, systems. New York: Academic Press. Mustonen, S. 1978. Tavoitteisen järjestelmän kyberteettinen analyysi päätäntäteorioiden ja systemoinnin metatutkimuksessa (in Finnish). Report A7, Institute of Data Processing Science, University of Oulu, Oulu. Olive, A. 1983. Analysis of conceptual and logical models in information systems development methodologies. In T. Olle, H. Sol & C. Tully (Eds.) Information Systems Design Methodologies: A Feature Analysis. Amsterdam: Elsevier Science Pub., 63-85. Olle, T., Hagelstein, J., MacDonald, I., Rolland, C., Sol, H., van Assche, F. & Verrijn-Stuart, A. 1988. Information Systems Methodologies – A Framework for Understanding. 2nd edition. Reading: Addison-Wesley. Peirce, C. 1991. The essential Peirce, Vol. 1, edited by N. Houser & c. Kloesel. Blooomington: Indiana University Press. Phalp, K. 1998. The CAP framework for business process modeling. Information and Software Technology 40(13): 731-744. Ralyté, J., Deneckere, R. & Rolland, C. 2003. Towards a generic model for situational method engineering. In J. Eder & M. Missikoff (Eds.) Proc. of the 15th Int. Conf. on Advanced Information Ssystems Enginering (CAiSE’03). LNCS 2681, Berlin: Springer-Verlag, 95-110. Slooten van, K. & Hodges, B. 1996. Characterizing IS development projects. In S. Brinkkemper, K. Lyytinen & R. Welke (Eds.) Proc. of the IFIP TC8 WG8.1/WG8.2 Working Conf. on Method Engineering: Principles of Method Construction and Tool Support. London: Chapman & Hall, 29-44. Sol, H. 1992. Information systems development: a problem solving approach. In W. Cotterman & J. Senn (Eds.). Challenges and Strategies for Research in Systems Development. New York: John Wiley & Sons, 151-161. Song, X. 1997. Systematic integration of design methods. IEEE Software 14(2): 107-117 Sowa, J. & Zachman, J. 1992. Extending and formalizing the framework for information system architecture. IBM Systems Journal 31(3): 590-616. Sundgren, B. 1975. Theory of data bases. New York: Petrocelli/Charter. Swede van, V. & van Vliet, J. 1993. A flexible framework for contingent information systems modeling. Information and Software Technology 35(9): 530-548. Tolvanen, J.-P. 1998. Incremental method engineering with modeling tools – Theoretical principles and empirical evidence. Dissertation Thesis, Jyväskylä Studies in Computer Science, Economics and Statistics, No. 47, University of Jyväskylä, Finland. Uschold, M. 1996. Building ontologies: towards a unified methodology. In Proc. of 16th Annual Conf. of the British Computer Society Specialist Group on Expert Systems. Cambridge, UK Uschold, M. & King, M. 1995. Towards a methodology for building ontologies. In Workshop on Basic Ontological issues in Knowledge Sharing, held in conjunction with IJCAI’95, Montreal, Canada. Wand, Y. & Weber, R. 1995. On the deep structure of information systems. Information Systems Journal 5(3): 203-223. Welke, R. 1977. Current information system analysis and design approaches: framework, overview, comments and conclusions for large – complex information system education. In R. Buckingham (Ed.) Education and Large Information Systems. Amsterdam: Elsevier Science Pub., 149-166. Yourdon, E. 1989. Modern structured analysis. Englewood Cliffs: Prentice-Hall. Zachman, J. 1987. A framework for information systems architecture. IBM Systems Journal 26(3): 276-292.
275
M. Leppänen / A Perspective Ontology and IS Perspectives
Appendix IS action
1..* HIS purpose
strivesFor
UI
Dialog *
1..* HIS rule
1..*
* governs
* *
1..* 1..* responsibleFor
*
1..*
HIS action
*
Window
navigation
* operatesWith
1..*
input output * *
1..* *
* UI component
IS role Data object 1..* 1..* contains 1..* *
1 UI data
* IS position
UI data comp.
UI action comp.
supervision *
*
*
*
CIS rule
1..* nonDigital
UI state
1 IS org.unit
precedes
* *
1..*
*
*
CIS action
Transaction 1..*
*
1..*
*
*
UI transition
* causedBy *
1
governs
*
1
resultsIn
Digital
*
governs
1..*
*
IS organization
1..*
implements
presents 1..*
* triggers *
Algorithm
UI event
HIS
CIS
output input
Figure A.1. IS datalogical perspective (HIS = Human Information System, UI = User Interface, CIS = Computerized Information System)
1..*
1..* Data storage
allocated
Application SW
Memory device
1..* 1..*
Data file
Data base
SW component
1..*
*
Processor
1..* 1..*
1
1..* allocated
Record
Layer
1
1
Data message 1..*
1..*
1..* 1..* transmittedThrough 1..*
Data field
1..*
Node 1..*
1
connects 1..* 1..*
1..*
Communication line
SW architecture situated
1..* applies 1..*
1..* 1
1..*
1 1..*
Protocol
PhysicalLocation
HW architecture
CIS 1..*
Figure A.2. IS physical perspective covering a part of the CIS
276
Information Modelling and Knowledge Bases XIX H. Jaakkola et al. (Eds.) IOS Press, 2008 © 2008 The authors and IOS Press. All rights reserved.
The improvement of data quality – a conceptual model Tatjana WELZER, Izidor GOLOB, Boštjan BRUMEN, Marjan DRUŽOVEC, Ivan ROZMAN Faculty of Electrical Engineering and Computer Science, University of Maribor Smetanova 17, Maribor, Slovenia {welzer, izidor.golob, bostjan.brumen, marjan.druzovec, ivan.rozman}@uni-mb.si Hannu JAAKKOLA Tampere University of Technology Pori, Finland [email protected] Abstract. Usage of data in various areas and its electronic availability has upgraded the importance of data quality to the highest level. In general, data quality has at least a syntactic and a semantic component. The syntactic component is relatively easily reached, mostly supported by tools, while the semantic component requires further research. In many cases, data is taken from different sources which are distributed among enterprises and vary in levels of quality. Special attention needs to be paid to data upon which critical decisions are met. In the paper we will focus on data quality in connection with conceptual modeling, including reuse of models and/or parts of them and data policy for increasing the quality of data.
Introduction Database design is concerned with arranging data required by one or more applications in an organised structure. We are facing an increasing demand for more and more complex applications on databases. This rapid growth has stimulated the need for higher level concepts, tools and techniques for database design and development. At the beginning, when databases first entered the information system market, database designers needed to invest a lot of hard work to the development of databases which was then only supported by very rough tools. But nowadays, however, designing databases has become a popular activity, performed not only by database designers, but also by nonspecialists, raising the issue of a possible inflation of quality, either as far as the database itself is concerned or concerning the saved data in the database. The starting point for the design of a database is mostly some abstract and general description of the reality, namely a conceptual model which is developed in the first phase of the database design. In the context of the database design, the conceptual model has various usages [cf. Frost 1986]: x at the start of the database design, it should integrate various interests and views of the end user; x it is a useful description for communication with users as well as for communication with non-specialists;
T. Welzer et al. / The Improvement of Data Quality – A Conceptual Model
277
it helps the database designer to build a more durable database system; it enables efficient introduction of the already designed database. To achieve the above mentioned usages of the conceptual model as well as in order to design an effective and high quality model, we must take into consideration that the conceptual database design is extremely complex and iterative [Ramamoorty 1989]. It can be greatly simplified by using different database design aids (methodologies and tools). Design methodologies for conceptual design should be rigorous as well as flexible. The methodologies should be based on a formal approach and also be applicable to a variety of situations and environments. Additonally, they should take care also about the data data quality and consecutively on information quality. Many approaches to improve information quality have been developed in the last years and have been implemented in various situations. Most of these approaches are only vaguely familiar with the fact that data quality is a prerequisite for the final quality of information. Data are of high quality if they are suitable for being used in a database. Data appear to be suitable for usage if they are free of damages and possess desired features [7],[1]. In information technology, aggressive steps to improve the data quality are being done. In our contribution, we will concentrate on the problem of improving data quality on basis of developing a conceptual model. In the following chapters we will be concentrating on data quality in general, as well as on possible data policy. Further research and final remarks will be presented in the conclusion. x x
1. Quality of Data The quality of data limits the ability of the final user to make correct decisions, which can have fatal consequences. There are a number of indicators which relate to the quality of data: accuracy, integrity, consistency, accessibility, comprehensives, timeliness and completeness, among others. The data must follow business rules and be free from anomalies. Although being a subjective measurement, the user's satisfaction with the quality of the data and the information derived from it, is arguably the most important indicator of them all [12],[2]. There are many reasons why it is difficult to capture and maintain quality data. Some of the difficulties are process-related, some are human-related problems, and, furthermore, other obstacles have their source in technology itself. All of the problems result in bad data, either from the semantic or syntactic point of view. Usually, these two components are interrelated and inconsistent. Process-related problems are frequently caused by the user by entering the data into an operational system at the wrong point of the business process or by lacking the understanding for the meaning of the data [13]. Difficulties with employees entering incorrect data into systems can be decreased by changing the emphasis on pure speed of processing to the quality of processing, where quality is composed of both speed and accuracy. Regardless of the source of the problems, it is important to identify the source of the problem, analyze its impacts and, where possible, propose a solution. We have to be aware of the fact that usually the customers vary in their needs and this might lead to a conflict. Additionally, the customers´ needs change all the time and what was good enough one day (talking about suitable data quality), is simply not good enough the next day (does not meet the data quality) [8]. The issue of data quality is an issue particularly important in data warehouses and data mining, especially in combination with sensitive domains (e.g. medicine, energy, flights).
278
T. Welzer et al. / The Improvement of Data Quality – A Conceptual Model
The introduction of sensitive domains increased the priority of data quality, as the risks and costs of inadequate quality become more visible and more real, and after all, more expensive [5]. The problem of poor data quality is one of the most difficult problems to be solved while constructing any conceptual model [8]. Because of bad data quality, more time and money is spent than anyone assumed initially. Pyle [6] estimates that the data preparation sub-process can take up to 90% percent of the time and money available for the whole system development. Data quality is also connected to the conceptual model while during its development we are checking data through defining entities, relationships and especially attributes describing both of them. Mostly in this phase the emphasis is on the syntactic and semantic quality, but some authors define additional types of data quality (e.g. physical quality, perceived semantic quality, pragmatic quality, social quality, language quality and knowledge quality) (Krogstie, 1998).
2. Conceptual modeling and data quality As mentioned before, we are connecting conceptual modeling with data quality which is a very likely and natural connection. In our work we pointed out mainly the syntactic quality (correspondence between the model and the tool in which the model is written (Krogstie, 1998)) and semantic quality which is presenting the correspondence between the model and the domain. The rest of possible data quality parts are not the topic of this paper, but we would like to point out that in one way or another they are involved or connected mostly with the semantic quality. According to the before mentioned definition for semantic quality, we have to bring up two semantic goals: validity – all statements made in the model are correct and relevant to the domain (no invalid statements) and - completeness (also a component of a data policy structure): the model contains all statements that would be correct and relevant about the domain. So it becomes apparent that the role of the domain is very important with a strong focus on data quality. We have to be very careful with those domains which are related to very sensitive data like that in the medical environment. Medical environments require special taking care for data because of their dual nature. First, the business part is required, which is just as complex as the business part in any other enterprise. Second, the medical part needs to be interwoven with the business part. In a general business system, we usually do not have such a duality. The medical part has its own specifics and requirements and if it is, in general, possible to reuse business objects (data models, applications) from different enterprises for the business part of a medical system [11], [4] we cannot apply the same for the medical part alone. As mentioned, it is essential that the medical part is tightly connected with the business part. Decisions made in a medical environment are very sensitive because they affect the “business object” (the patient) directly. Any decision, being it managerial or medical, can have fatal and devastating consequences. Moreover, the data should provide a firm foundation for information retrieval and, furthermore, for knowledge discovery. From the data quality point of view, we have to assure quality in each of the sub-systems. Additionally, the final (integrated) medical system, composed of both the business and medical part, needs to be validated again concerning data quality. The integrated high quality systems are the desired goal of system designers. To reach the goal, the following steps have to be followed: x Separate data issues from more traditional technical issues and assign lead responsibility for data to someone within the medical community.
T. Welzer et al. / The Improvement of Data Quality – A Conceptual Model
279
x
As with all quality efforts, the needs have to be understood correctly. Additional explanations and comments are welcome. x Already existing data have to be checked again. The importance of additional checking is growing, as the sensitivity of the domain increases (medical environment). We want also to emphasize that an organization sometimes may be doing better not having certain data (responsibility of community, request for additional check) than having inaccurate data, especially if those relying on the data are not aware of its inaccuracy. For example, a hospital would be put to a better position not knowing a patient's blood type than wrongly believing it to be O+. A problem of semantic data quality is evident. How can all these problems be solved? One of the possible solutions is to introduce data policy.
3. Data Policy A policy is a plan, course of actions or set of rules intended to influence and determine decisions, actions and other matters [9]. The origin of defining data policy is the assumption that the responsibility for the quality of data has to be assigned to those who create the data or to those who are as close to data creation as possible. That means that the data policy supports the work of these people through suggestions and rules that they have to follow or at least take into consideration. With the aim to define easily used, but nevertheless, powerful policy, we are suggesting the following structure: x Introduction (purpose, audience, definitions, related work, basic approach, responsibility, comments) x Data policy (data policy and its benefits, components of a good data policy, needs/reasons for a successful data policy, comments) x Structure (objectives, categories of data policy) x Domain rules (specific environment rules) x Syntactic and/or other data quality types x Dimensions of data quality (relevance, availability, clarity of definition, comprehensiveness, accuracy, integrity, homogeneity, structural consistency, consistency (semantic consistency), accessibility, security, timeliness completeness, portability) x Others (definitions, homogeneity, naming, redundancy, comments…) x Policy management (responsible for reviews, schedule of reviews, recommendation for reviews, policy issuance and revision date, comments) We have already mentioned that the focus in this paper is on the dimension of data quality (availability, security, comprehensiveness, flexibility, appropriate use, semantic consistency, simplicity, relevancy, completeness, consistency, portability, naming, relevancy, concurrency, definitions, robustness, homogeneity, redundancy). Furthermore, we set out especially those dimensions (cf. the policy structure) which assure data quality by considering conceptual models and/or parts of them also concerning the system of reusability – reusable components are already existing models and/or just confirmed parts of them (Welzer, 2004): x Relevance - objects needed by the applications are included in conceptual models. x Clarity of definition - all terms used in the conceptual model are clearly defined. x Comprehensiveness - each needed attribute should be included.
280
T. Welzer et al. / The Improvement of Data Quality – A Conceptual Model
x x
Occurrence identifiably - identification of the individual objects is made easy Homogeneity, structural consistency - object level enables the uniformity of stored concepts. x Minimum redundancy - only checked conceptual models are included. x Semantic consistency - conceptual models are clear and organized according to the application domains. x Robustness, flexibility - through the reuse both characteristics are fulfilled. x Security – reusability is increasing security, even that each specific domain could have especially needs and demands. x Appropriate use – in general, this varies from domain to domain, but the reusability raises the trustworthiness x Availability – specific domains could have especially needs or demands for it. It becomes obvious that those dimensions are connected both to the data policy structure (categories) and to different quality types, especially the semantic one.
4. Conclusion Medicine is a specific environment and thus creates a form of knowledge discovery for data and about data. Not only that the data need to be accurate, valid and timely (the semantic component of the data quality), but also the structures, from which we obtain the required data, have to be valid and syntactically correct. Too often the structures are neglected or taken for granted. For instance, if the information or knowledge obtained from the data does not satisfy users’ wishes, needs and demands, the process needs to be reverted to the previous steps – obtaining the data. But this does not lead to the desired outcome, since the structures are not suitable for the task. We argue that the data (information, knowledge) quality heavily relies on the data policy whose structure is introduced in the paper. For future research more practical results of introducing data policy in specific environments (medicine) are expected to confirm the data policy structure or to introduce some changes in dimensions (additional dimensions, changes for existing dimensions, withdrawing of existing dimensions) and/or data quality types. The result will not so easily be reached because medical data are very sensitive data, as we have mentioned, and most of them are secure and protected in various ways. So we will get quite easily access to statistical data (public databases are available), but much more difficult to handle will be the situation with conceptual models on which we have to check our conclusions on data quality types, data policy structures and influence of the conceptual model on the data quality.
References [1] Welzer, T. and Rozman, I. (1998): Information Quality by MetaModel. In Proceedings of Software Quality Management VI. Quality improvement issues. 81-88. HAWKINS. C (eds). Springer. London. [2] Welzer, T., Brumen, B., Golob, I. and Družovec, M. (2002): Medical diagnostic and data quality. In Proceedings of 15th IEEE Symposium on computer-based medical systems. 97-101. KOKOL P., STIGLIC B., ZORMAN M. and ZAZULA, D. (eds) IEEE Computer society. Los Alamitos. [3] Welzer, T. and Družovec, M. (2000): Similarity search in Database Reusability – a Support for efficient design of conceptual models. In Contemporary Applications and Research Issues in Industrial Product Moddeling. 23-34. HELLO. P. and WELZER. T. (eds). University of Vaasa.
T. Welzer et al. / The Improvement of Data Quality – A Conceptual Model
281
[4] Freeman, P. (1987): Reusable Software Engineering Concepts and Research Directions. In IEEE Tutorial Software Reusability. FREEMAN. P. (ed.). IEEE. [5] English, L.P. (1999): Improving data warehouse and Business Information Quality. John Wiley & Sons. [6] Pyle, D. (1999): Data Preparation for Data Mining. Morgan Kaufmann Publishers, Inc., San Francisco, California, USA. [7] Redman, T.C. (1996): Data Quality for the Information Age. Artech House. [8] Redman, T.C. (2001): Data Quality, The field Guide. Digital Press, Boston, USA. [9] Whitham, M.E. and Mattord (2003): Principles of information Security. Thomson. Canada. [10] Reiter, R. (1987): A Theory of Diagnosis from First Principles. Artificial Intelligence 3(2):57-95. [11] Rine, D.C. (1997): Supporting Reuse with Object Technology. IEEE Computer, 30(10): 43-45. [12] Tayi, G.K. and Ballou, W (1998): Examining Data Quality. Communications of the ACM, 4(12):54-57. [13] Welzer T., Brumen B., Golob I., and Sanchez J.L., Družovec M. (2004): Diagnostic process from the data quality point of view, Journal of Medical Systems, Kluwer.
282
Information Modelling and Knowledge Bases XIX H. Jaakkola et al. (Eds.) IOS Press, 2008 © 2008 The authors and IOS Press. All rights reserved.
Knowledge Cluster Systems for Knowledge Sharing, Analysis and Delivery among Remote Sites Koji ZETTSU a , Takafumi NAKANISHI a , Michiaki IWAZUME a , Yutaka KIDAWARA a , and Yasushi KIYOKI a,b a National Institute of Information and Communications Technology, Japan b Keio University, Japan Abstract. We, NICT, recently started a new project for research and development of “knowledge cluster systems” for knowledge sharing, analysis, and delivery among remote knowledge sites. We introduce several key concepts of the knowledge cluster systems. The “Three-site model” for knowledge system architecture defines three roles of remote sites: knowledge capture, knowledge transfer, and knowledge provision, with respect to the lifecycle of knowledge communication. The “global knowledge grid” is as an infrastructure that is suitable for implementing knowledge cluster systems on the basis of the three-site model. The knowledge cluster systems build an evolving network of community knowledge by connecting heterogeneous knowledge bases. The “global risk management system” is being developed as an application of the knowledge cluster systems. Keywords. Knowledge cluster systems, three-site model architecture, global knowledge grid, connection of heterogeneous knowledge bases
Introduction In today’s networked society, knowledge-intensive work involves a significant amount of communication, coordination, and cooperation practices that cross the boundaries of organizations, counties, cultures, and/or disciplines. As a motivating example, let us consider managing risks against natural disasters like Tsunami, volcanic eruptions, or avian flu. Natural disasters cause damages in various fields like health, economy, natural ecosystems, and so on. Moreover, the damage may spread beyond national boundaries. Therefore, experts from various fields need to collect and analyze information related to disasters. In addition, disaster victims and relatives need to be adequately informed. We believe that it is important for next-generation knowledge systems to place a particular emphasis on knowledge communication in order to manage knowledge in a world of networks [1]. While knowledge systems will be characterized by weakly structured and less predictable processes, in order to make communication stronger, we need to share, analyze and deliver knowledge among
K. Zettsu et al. / Knowledge Cluster Systems for Knowledge Sharing, Analysis and Delivery
283
Figure 1. Basic concept of global risk management system (for hot mud flood disaster in Indonesia).
remote knowledge sites. We introduce a concept of “knowledge cluster systems”, and explain our efforts toward their realization, especially in the context of global risk management against natural disasters.
1. Background The knowledge cluster system project started from April 2006 as a five-year research project of the National Institute of Information and Communications Technology (NICT), Japan. The main objective is to research and develop nextgeneration knowledge infrastructure in the networked age. The knowledge infrastructure consists of three functional layers: distributed knowledge access, knowledge computing and analysis, and knowledge presentation media. The “Global risk management system” is proposed as the first application, which aims to evaluate the impacts of various risks caused by natural disasters in a global context. It is a good example of knowledge-communication-intensive work, as described above. From February 2007, NICT and Electronic Engineering Polytechnic Institute of Surabaya, Institut Teknologi Sepuluh Nopember (EEPISITS) in Indonesia started a joint project on the research and development of global risk management system for natural disasters, especially for hot mud flow disaster. The concept of the global risk management system is illustrated in Figure 1. The meta-level architecture of the global risk management is shown in Figure 2. The local risk management subsystem focuses on shallow but real-time analysis of local risks, while the meta-level risk management subsystem focuses on deep but non-real-time analysis of global risks. To design the technology that is used in knowledge cluster systems, NICT held the First International Workshop on Knowledge Cluster Systems in March 2007 at Kyoto, Japan. The participants were from EEPIS-ITS (Indonesia), Tampere University of Technology (Finland), Christian Albrechts University at Kiel
284
K. Zettsu et al. / Knowledge Cluster Systems for Knowledge Sharing, Analysis and Delivery
Meta-level Risk Management Subsystem NICT
Global Risk Management Knowledge Base
Globalize local problems
EEPIS-ITS
Monitor local disaster events
Exploit global solutions Local Risk Management Subsystem for Indonesia
Local Risk Management Subsystem for Japan
Indonesian Risk Management Knowledge Base
Japanese Risk Management Knowledge Base
(to be added)
Localize global solutions in dependence upon local situations Inform the general public about local situations
Figure 2. Meta-level architecture of global risk management system.
(Germany), VSB-Technical University of Ostrava (Czech Republic), Saga University, Keio University, and Kanagawa Institute of Technology (Japan).
2. Three-site Model Knowledge System Architecture In the networked society, various communities organize their own knowledge repositories, each of which aggregates perception, skills, training, common sense, and experience of a community of people. Knowledge cluster systems are used to facilitate knowledge communication across the boundaries of these communities. The lifecycle of knowledge communication consists of the following three phases: Knowledge capture: deriving knowledge from information as an understanding of the information depending on the discipline or context where it is used. Knowledge transfer: conveying (or projecting) the knowledge of one community to another community. It can be considered to be a process of transmission and absorption. Knowledge provision: providing actionable information available in the right format, at the right time, and at the right place. In knowledge cluster systems, the above three phases are defined as three different roles of remote sites. We propose the “three-site model” for the knowledge cluster systems architecture, in which remote sites play the above three roles and thus realize the knowledge communication. The three-site model in the global risk management system is illustrated in Figure 3. The sites located in/around the disaster area will play the “knowledge capture” role (site-1) in order to capture the knowledge about the disaster by col-
K. Zettsu et al. / Knowledge Cluster Systems for Knowledge Sharing, Analysis and Delivery
285
Site-2: Knowledge Transfer Knowledge Base
Site-1:Knowledge Capture Correlation
Site-3:Knowledge Provision
Disaster information Information analysis
Mobile Phones
sensor Sensor Manager
DB1
DB2
Web
GIS Visualization
Sensor Manager
3D Visualization
Sensor Manager
Observe social situations and natural environments
• Analyze disaster information from Site-1 • Evaluate risks by employing various knowledge bases • Send risk information to Site-3
Warning Center
Aggregate, select and convert risk information from Site-2 for intended recipients
Figure 3. Three-site model knowledge system architecture in global risk management system.
lecting and organizing the disaster information on site. The remote sites playing the “knowledge transfer” role (site-2) evaluates the impacts on various risks by analyzing the disaster knowledge from the site-1. In the course of the risk analysis, various communities (or domains) of knowledge are employed (e.g., healthcare knowledge, economic knowledge, and ecosystem knowledge). The risk knowledge discovered by the site-2 is sent to the remote sites playing the “knowledge provision” role (site-3). At the site-3, various risk knowledge are aggregated, selected, and converted into the right format for the intended recipients. For example, local disaster victims of a local disaster, will be informed of the first-aid actions in real time. On the other hand, policy decision makers will be provided with , comprehensive risk information on demand. In the course of the knowledge provision, the risk knowledge may be localized, personalized, and/or adapted on the basis of the situations of the intended recipients. Note that the application requirements and capability of the remote sites will significantly affect how the three roles are assigned to the remote sites. For example, to manage local risks in real time in the early phase of a disaster, all of the three roles may be assigned to the local site in a disaster area. As the damage spreads to other regions and/or various fields, the knowledge transfer role will be assigned to those remote sites, which have enough knowledge to evaluate the risks, and the knowledge provision role will be assigned to the remote sites influenced by the disaster. 3. Global Knowledge Grid: An Infrastructure for Knowledge Cluster Systems The “global knowledge grid” is an infrastructure for implementing knowledge cluster systems based on the three-site model. The concept of the knowledge grid
286
K. Zettsu et al. / Knowledge Cluster Systems for Knowledge Sharing, Analysis and Delivery
Figure 4. Global knowledge grid as an infrastructure for knowledge cluster systems.
has recently emerged as an integrated infrastructure for coordinating knowledge sharing and problem solving in distributed environments. The knowledge grid was originally presented as the implementation of parallel and distributed knowledge discovery (PDKD) on top of a computational grid[2]. The knowledge grid uses the basic functions of a grid and defines a set of additional layers to implement the functions of distributed knowledge discovery. The knowledge grid enables collaboration between knowledge providers who must mine data stored in different information sources, and knowledge users who must use a knowledge management system operating on several knowledge bases. Our intentions are (1) to make the knowledge grid accessible to anyone via a global network (i.e., the Internet) and (2) to implement an additional layer on top of the knowledge grid in order to provide the capability of knowledge communication. The additional layer, known as “knowledge cluster service layer”, comprises of software modules that provide the services defined in the three-site model (i.e., knowledge capture, knowledge transfer, and knowledge provision). The knowledge cluster services are developed by the knowledge grid nodes in parallel. The knowledge cluster service layer is organized on the basis of a serviceoriented architecture (SOA)[3]. The application of the knowledge cluster system is developed by composing the knowledge cluster services (e.g., service mash-up), as illustrated in Figure 4. Our challenge is to develop a mechanism for discovering optimal composition of the knowledge cluster services with respect to various requirements of knowledge communication. In contrast with traditional SOA-based applications like business transactions, in which service composition can be defined previously based on standardized business protocol (e.g., BPEL[4]), the knowledge cluster application requires dynamic assignment of knowledge cluster services. For example, in the global risk management system, the knowledge capture services will be assigned to the grid nodes in disaster area, while the knowledge transfer services will be on the grid nodes which have the knowledge for evaluating the risks, and
K. Zettsu et al. / Knowledge Cluster Systems for Knowledge Sharing, Analysis and Delivery
287
the knowledge provision services will be on the grid nodes near to the disaster victims. The knowledge cluster applications need to define the conditions under which the knowledge communications are activated or deactivated. At runtime, the knowledge cluster services that satisfy the conditions are bound dynamically. The dynamic binding of knowledge cluster services will be discussed in our future work.
4. The Web of Knowledge: Connecting Heterogeneous Knowledge Bases Cross-boundary knowledge communications builds an evolving network of community knowledge. In the same way as the World Wide Web, the knowledge cluster systems provide a framework of infinitely-evolving knowledge repository by connecting heterogeneous knowledge bases owned by different communities. A typical example of the connection is based on causal relation from one knowledge base to another knowledge base. For instance, a disaster knowledge base can be connected with healthcare knowledge base by establishing a causal relation in order to find diseases caused by specific disasters. In that way, “a web of knowledge” will be formed. Connecting two different knowledge bases requires a “bridge concept”. In conventional approaches, schema mappings and bridge ontology are typically used as the bridge concept. They try to pre-define universal relations between two different communities of knowledge, while it is quite difficult in most cases. As a result, conventional approaches can only work on a small scale. As a workaround, most recent approaches try to employ word-level universal relations, like those found in a thesaurus or on the WordNet lexicon. However, in these approaches, knowledge is ideally broken into a bag of words with losing its contextual information. In order to enhance the scalability of the knowledge base connection, we have put defining the universal relations to one side, and focused on finding contextdependent correlations between different communities of knowledge. For example, in the context of health risk against hot mud flood disaster, knowledge about volcanic gases in the disaster knowledge base may have correlations with knowledge about respiratory organs illness in the healthcare knowledge base. Especially, the correlation between hydrogen sulfide (H2 S) and pulmonary edema may be stronger than any other correlations. In this way, we intend to develop a mechanism for first managing the contexts for evaluating the correlations between different communities of knowledge, and second, measuring the strength of correlations in each context, simultaneously. We are now developing the above mechanism based on the semantic space model[5], and. how it works is shown in Figure 5 . In the semantic space model, knowledge is represented by a vector. The basic idea is to project the vector from one semantic space to another semantic space, then search the target space for the vectors highly similar to the projected vector. The strength of correlation is measured by the similarity value. The context is given by the vector projection function (e.g., causal relation matrix in Figure 5). The correlation measurement approach allows us to discover on demand the knowledge related to the given knowledge in the given context. It also allows ambiguity or uncertainty of the
288
K. Zettsu et al. / Knowledge Cluster Systems for Knowledge Sharing, Analysis and Delivery
Figure 5. Connecting heterogeneous knowledge bases by causal relations based on semantic (vector) space model.
bridge concept. We plan to include time scale and geographical scale in the bridge concept, because they have a high affinity with the correlation measurement approach.
5. Future Work In response to the results of the First International Workshop on Knowledge Cluster Systems, we have expanded the scope of our research and development to include the following topics under the collaborations with international partners. Software engineering for knowledge cluster systems: (1) software development framework for knowledge cluster services, and (2) service mediation for dynamic binding of knowledge cluster services. Towards “Web 3.0”: (1) collaboration architectures on demand with the collective intelligence, and (2) treatment of collaborative data with social interaction and community management. Quality-driven content and information mining: (1) knowledge discovery depending on the source characteristics, portfolio and tasks, intentions of the mining, and (2) development of content and information mining workbench.
K. Zettsu et al. / Knowledge Cluster Systems for Knowledge Sharing, Analysis and Delivery
289
Integration of sensor data analysis and knowledge analysis: (1) shallow analysis on real-time basis for warning systems, evacuation information systems, and (2) deep analysis on non-real-time basis for social/global impact analysis. Adequate and comprehensive information provision: (1) navigation using geographical data and navigation facilities based on mobile/ubiquitous technology, and (2) involving humanitarian organizations
6. Conclusions We introduced the knowledge cluster system project conducted by NICT, Japan. The main objective of the project is to create an infrastructure for knowledge communication and distribution in a world of networks. We proposed several key concepts of the knowledge cluster systems. The‘ ‘Three-site model” for knowledge system architecture defines three roles of remote sites: knowledge capture, knowledge transfer, and knowledge provision, with respect to the lifecycle of knowledge communication. The “global knowledge grid” is being developed as an infrastructure that is suitable for implementing knowledge cluster systems on the basis of this model. An application of the knowledge cluster systems, “the global risk management system,” is being developed as a joint research project between NICT and EEPIS-ITS. We also discussed building an evolving network of community knowledge by connecting heterogeneous knowledge bases. The knowledge cluster system project will continue until 2011. To ensure the success of the project, we are looking for international collaborations including but not limited to the following: • Research and development of: (1) the global knowledge grid, (2) knowledge discovery and data mining, (3) knowledge bases, (4) knowledge presentation media, and (5) applications of knowledge cluster systems including global risk management system. • Field experiments and technology demonstrations. • Exchange program for researchers and/or students. References [1]
[2]
[3] [4] [5]
Zettsu, K. and Kiyoki, Y.: Towards Knowledge Management based on Harnessing Collective Intelligence on the Web, Proceedings of the 15th International Conference of Knowledge Engineering and Knowledge Management – Managing Knowledge in a World of Networks – (EKAW2006), Lecture Notes in Computer Science Vol. 4248 pp.350–57 (2006). Cannataro, M. and Talia, D.: The Knowledge Grid: Designing, Building, and Implementing an Architecture for Distributed Knowledge Discovery, Communications of the ACM, Vol. 46, No. 1, pp.89–93 (2003). Papazoglou, M. P. and Georgakopoulos, D.: Service-Oriented Computing, Communications of the ACM, Vol. 46, No. 10, pp.24–28 (2003). Fu, X., Bultan, T. and Su, J.: Analysis of Interacting BPEL Web Services, Proceedings of the 13th international conference on World Wide Web, pp. 621 - 630 (2004). Kiyoki, Y. Kitagawa, T. and Hayama, T.: A Metadatabase System for Semantic Image Search by A Mathematical Model of Meaning, ACM SIGMOD Record Vol.23 No.4 pp.34– 41 (1994).
290
Information Modelling and Knowledge Bases XIX H. Jaakkola et al. (Eds.) IOS Press, 2008 © 2008 The authors and IOS Press. All rights reserved.
A Formal Ontology for Business Process Model TAP: Tasks-Agents-Products Souhei ITO, Shigeki HAGIHARA and Naoki YONEZAKI Department of Computer Science, Graduate School of Information Science and Engineering, Tokyo Institute of Technology Abstract. TAP is a general process modeling framework with which we can describe process specifications in terms of tasks, agents and products that are relevant to the concept of processes. However, its formal semantics and pragmatics have not been rigorously studied, i.e. the usage of vocabulary words provided by TAP framework and the pieces of reality which TAP framework captures has not been formally considered. In this paper, we clarify the pragmatics of TAP and the world structures of TAP process models by formal approach. We present the semantic structure for TAP process models and introduce a logical language to restrict world structures of TAP. A formal ontology for business process model TAP is the characterization described as a set of axioms in our logical language.
1 Introduction In the fields of business process modeling, there are several ontologies, e.g. the AIAI Enterprise Ontology [6], the Toronto Virtual Enterprise Ontology (TOVE) [1], the Resource Event Agent (REA) Enterprise Ontology [5, 2] and e3 -valueTM [3]. Generally, the term “ontology” is thought of as specifications of concepts and relationships between concepts. This definition of ontology permits several forms of ontologies. In fact, the above ontologies are not defined in the same way. TAP (Tasks-Agents-Products) [7] is also a business process modeling framework. Things which are relevant to processes are modeled as modeling objects in TAP. First class modeling objects in TAP are tasks, agents and products. Processes are modeled by describing relations between objects and behavioral specifications. Therefore, TAP can also be viewed as an business ontology. To precisely understand the business processes which each ontology captures and to compare these ontologies rationally (e.g. what is the common concept), we think the formal approach is essential. Therefore, we introduce the formal semantic structures corresponding to pieces of reality and a logical language to describe the ontologies. This type of ontology is sometimes referred to as formal ontology. In this paper, we use TAP as a general business process model to apply formal approach, since TAP has specific concepts such as an enaction of agents or tools, meta-tasks, abstractinstance level objects etc. which other business ontologies lack. To formalize these characteristics is a very challenging issue. First, we give the semantic structure of TAP. Then, the formal vocabulary to describe TAP specifications and the logical language to describe axioms are introduced. The semantics of this logical language is defined according to the semantic structure. Finally we present axioms to specify the characteristics of the intended world structures (concepts and relationships between them) of TAP in our logical language. It also constrains the consistent usage
S. Ito et al. / A Formal Ontology for Business Process Model TAP: Tasks-Agents-Products
291
of the vocabulary words in TAP and can be viewed as the pragmatics of TAP specifications. This set of axioms is a formal ontology for business process model TAP. This paper is organized as follows. Section 2 summarizes the business process model TAP. Section 3 introduces the semantic structure for TAP. Section 4 introduces the logical language LTAP for describing axioms which specify the characteristic of the world structures of TAP. In Section 5 we present the set of axioms. The final section is the conclusion.
2 Business process model TAP A process is modeled in terms of modeling objects, that are related to the important concepts in business processes. Modeling objects are either abstract level modeling objects or instance level modeling objects. The former are abstract description of something that may exist during an actual process performance and are used to describe a generic process model that covers various possible situations. The latter denote the actual observable things or phenomena that appear in the process and can be viewed as descriptions of their real counterparts. Figure 1 illustrates the modeling objects in TAP approach and the round boxes stand for them. A pair of modeling objects can be connected with a directed arc, which represents relationship between them.
Figure 1: Modeling objects and relationships in TAP approach
Each modeling object encapsulates attributes which are data elements, and has its own value for each attribute. For example, it can be considered that the modeling objects “task” has eight attributes - “name”, “definition”, “task category”, “duration”, “complexity”, “application domain”, “business domain”, and “size”. Moreover, TAP has five dimensions of process modeling - generalization, classification, aggregation, control, and behavior. We cannot give a detailed explanation for all of concepts and dimensions for lack of space so we only explain some of them. A task, one of the central concepts, is a general description of work and is independent of any specific situation in an actual process. Therefore the task definition does not contain information about when or by whom it is performed. Instead, there is a modeling object that denotes task instance, called task performance, to which such attributes belong. A task can have a substructure (subtasks). The temporal and causal relationships with each other are
292
S. Ito et al. / A Formal Ontology for Business Process Model TAP: Tasks-Agents-Products
Figure 2: Top level description of an example process Figure 3: Behavioral description of Tasks
described in conditional Petri net. It is the behavior description of tasks. We give an example of process model (Figure 2, 3). Figure 3 is a behavioral description of task “Develop Change & Test Unit” in Figure 2. Some task might produce or consume products which are also TAP specifications. These kinds of tasks are scheduling tasks, monitoring tasks and are referred to as meta-task or controlling task. A task can contain a meta-task as its subtask. In Figure 3, meta-tasks are represented as bold boxes.
3 The semantic structure for TAP In this section, we define the semantic structure for TAP. As mentioned in Section 2, TAP has several conceptual components of process modeling. The semantic structure for TAP contains such components. Definition 1 (Frame) A frame is a 7-tuple F = DO , DT , S, A, R, F, , where DO is a semantic domain of modeling objects tasks, agents and products etc. and DT is a semantic domain of time attribute values, S is a set of states, A ⊆ S × S is an accessibility relation, R is a set of intensional relations and F is a set of intensional functions. DT has the special element ε. DT − {ε}, is a totally ordered set and neither ε t nor t ε hold for all t ∈ DT . Let σi ∈ {O, T } for all i. An intensional relation of arity σ1 × · · · × σn on DO , DT , S is a total function p : S −→ P(Dσ1 × · · · × Dσn ), where P(D) is the power set of D. An intensional function of arity σ1 × · · · × σn −→ σn+1 on DO , DT , S is a total function f : S −→ (Dσ1 × · · · × Dσn −→ Dσn+1 ). In this paper, we only consider time attribute among many data elements because time attributes such as “start time” and “end time” are important to represent the enaction order of task performances. It is easy to extend the frame definition to the case where we consider other attributes by incorporating domains for them into the frame. S represents state of affairs of modeling objects. A is the state transition relation on S. s, t ∈ A means that a state s can become a state t. An intensional relation is a function from states to a mathematical (ordinary) relation. Therefore the extension of it varies by states. For example, the situation may happen that an extensional relation in a of a binary intensional relation p in a state s
S. Ito et al. / A Formal Ontology for Business Process Model TAP: Tasks-Agents-Products
293
holds between a and b but that of in a different state t does not. Intensional relations are semantic objects for relationships in TAP. Therefore relationships in TAP can be captured by intensional relations. Similarly, an intensional function is a function from states to a mathematical (ordinary) function. Intensional functions are semantic objects for attributes in TAP. Some attribute values may change along the proceeding of the enaction. Therefore an intensional function is used for a semantic object for attributes. ε is the value which indicates the value of a function ranging over DT undefined. For example, the situation that some task performance has not started yet can be represented in our structure as that the value mapped by the function corresponding to the attribute “start time” for the object corresponding to the task performance is ε.
4 The logical language LTAP In this section, we introduce the logical language LTAP with a formal vocabulary. The formal vocabulary consists of a set of predicates and attributes used in describing TAP specifications. The formal ontology is described in LTAP as a set of axioms.
4.1 Syntax Our logical language LTAP is many-sorted logic. Definition 2 (Sort) The set of sorts is Sort = {O, T }. Sort O represents modeling objects and T represents times. If we consider attributes other than time, we add sorts for them. Now, we introduce symbols of LTAP . Definition 3 (Symbol) The symbols of LTAP consists of the following: 1. Vocabulary V consisting of the following predicate symbols and function symbols. • Predicate symbols of arity O. Task , Product, AgentType, AgentRole, ToolType, ToolRole, TaskPerformance, ProductInstance, AgentInstance, AgentEnaction, ToolInstance, ToolEnaction, Edge, Bar , MetaTask . • Predicate symbols of arity O × O. is instance of , produces, is consumed by, performed by, supported by, plays role of , activates, of , constitutes. • Function symbols of arity O −→ O. source, destination. • Function symbols of arity O −→ T . start time, end time. 2. The set of constant symbols Con = Con O ∪Con T . Con O is the set of constant symbols of sort O and has the special symbols ⊥O and TAPspec. Con T is the set of constant symbols of sort T and has the special symbol ⊥T . 3. The set of variables Var = Var O ∪ Var T . Var O is the set of variables of sort O and Var T is the set of variables of sort T . 4. The set of connectives {∧, ∨, ¬, →, ↔, ∀O , ∀T , ∃O , ∃T , 2, 3}. 5. Equality symbol = of arity O × O and T × T .
294
S. Ito et al. / A Formal Ontology for Business Process Model TAP: Tasks-Agents-Products
6. A binary symbol ≤ of arity T × T . The vocabulary words for static specifications of process models are modeling object type names and relationship names appearing in Figure 1. constitutes(x, y) represents that x is a substructure of y. Edge, Bar , source and destination are used to describe conditional Petri net description. For example, a fragment of Figure 3 can be described as {Edge(approved), source(approved) = Modify Design, destination(approved) = b, Bar (b)}. start time and end time are attributes. If we consider other attributes, we add function symbols for them. Constant symbols ⊥O and ⊥T are used to indicate that the value of some function on some object is undefined. Definition 4 (Term, Atom) The sets Term O of terms of sort O, Term T of terms of sort T and Atom of atoms are the smallest X, Y and Z respectively satisfying the following properties: 1. 2. 3. 4. 5. 6. 7.
Con O ∪ Var O ⊆ X, Con T ∪ Var T ⊆ Y , If the arity of f ∈ V is O −→ O and t ∈ X then f (t) ∈ X, If the arity of f ∈ V is O −→ T and t ∈ X then f (t) ∈ Y , A1 , . . . , An ∈ Z ⇒ spec(A1 , . . . , An ) ∈ X, If t ∈ X and the arity of p ∈ V is O then p(t) ∈ Z, If t1 , t2 ∈ X and the arity of p ∈ V is O × O then p(t1 , t2 ) ∈ Z.
The term spec(A1 , . . . , An ) represents the TAP specification {A1 , . . . , An }. This construct is used to describe objects which are produced or consumed by meta-tasks. Definition 5 Let σ ∈ Sort. Formulas in LTAP are defined inductively as follows: 1. Atoms are formulas. 2. If ϕ and ψ are formulas then ϕ ∧ ψ, ϕ ∨ ψ, ¬ϕ, ϕ → ψ, ϕ ↔ ψ, 2ϕ and 3ϕ are also formulas. 3. If ϕ is a formula and x ∈ Var σ then ∀σ xϕ and ∃σ xϕ are also formulas. 4. If s, t ∈ Term σ then s = t is a formula. 5. If s, t ∈ Term T then s ≤ t is a formula. Notational Convention. The order of strength of connection between connectives is ¬, ∀, ∃, 2, 3, ∧, ∨, →, ↔. We write s = t instead of ¬(s = t). We simply write ∀xϕ and ∃xϕ instead of ∀O xϕ and ∃O xϕ respectively.
4.2 Semantics We define the semantics of LTAP with respect to frames (Definition 1) which is our semantic structures of TAP process models. Definition 6 (Model) Let F = DO , DT , S, A, R, F, be a frame. An interpretation I of symbols with respect to F is a function satisfying the following: 1. 2. 3. 4. 5. 6.
I(c) ∈ DO if c ∈ Con O , I(c) ∈ DT − {ε} if c ∈ Con T − {⊥T }, I(⊥T ) = ε, I(x) ∈ DO if x ∈ Var O , I(x) ∈ DT if x ∈ Var T , I(spec(A1 , . . . , An )) ∈ DO if A1 , . . . , An ∈ Atom,
S. Ito et al. / A Formal Ontology for Business Process Model TAP: Tasks-Agents-Products
295
7. I(p) ∈ R if p is a predicate symbol in V where the arity of p and I(p) are the same. 8. I(f ) ∈ F if f is a function symbol in V where the arity of f and I(f ) are the same. Let F be a frame and I be an interpretation with respect to F. A model is a pair M = F, I. Definition 7 (Interpretation of terms) Let M = DO , DT , S, A, R, F, , I be a model and s ∈ S. An interpretation (M, s) of terms in Term is a function satisfying the following: 1. 2. 3. 4.
(M, s)(c) = I(c) if c ∈ Con, (M, s)(x) = I(x) if x ∈ Var , (M, s)(spec(A1 , . . . , An )) = I(spec(A1 , . . . , An )), (M, s)(f (t1 , . . . , tn )) = I(f )(s)((M, s)(t1 ), . . . , (M, s)(tn )).
Definition 8 Let M = DO , DT , S, A, R, F, , I be a model and s ∈ S. The satisfaction relation |= is defined inductively as follows: M, s |= p(t1 , . . . , tn ) M, s |= t1 = t2 M, s |= ¬ϕ M, s |= ϕ ∧ ψ M, s |= ϕ ∨ ψ M, s |= ϕ → ψ M, s |= ϕ ↔ ψ M, s |= ∀σ xϕ M, s |= ∃σ xϕ M, s |= 2ϕ M, s |= 3ϕ
iff iff iff iff iff iff iff iff iff iff iff
(M, s)(t1 ), . . . , (M, s)(tn ) ∈ I(p)(s) (M, s)(t1 ) = (M, s)(t2 ) M, s |= ϕ M, s |= ϕ and M, s |= ψ M, s |= ϕ or M, s |= ψ M, s |= ϕ or M, s |= ψ M, s |= ϕ iff M, s |= ψ M[x → d], s |= ϕ for all d ∈ Dσ M[x → d], s |= ϕ for some d ∈ Dσ M, s |= ϕ for all s such that s, s ∈ A+ M, s |= ϕ for some s such that s, s ∈ A+
M[x → d] is the same as M except that it has I[x → d] as the interpretation of symbols. I[x → d] is the same as I except that it maps x to d. A+ is the transitive closure of A. Our modal logic holds transitivity so the system of our modal logic is K4.
5 Axioms In this section we describe axioms which characterize the intended world structures of TAP. Models of the axioms approximate the worlds which TAP approach intends to model. We pick several axioms from our formal ontology since we cannot include all of them for lack of space. In TAP approach, once objects or relations appeared at some state then they remain hereafter. In other words, histories must be hold in TAP approach. Axiom 1 and 2 state this. 1. ∀x(P (x) → 2P (x)), where P is a metavariable represent a unary predicate symbol in V . 2. ∀x∀y(P (x, y) → 2P (x, y)), where P is a metavariable represent a binary predicate symbol in V . Axiom 3 says that once the start time or the end time of some object are set at some state then they do not change hereafter. The start time or the end time are for task performances and are peculiar to them. 3. ∀x∀T y(f (x) = y ∧ y = ⊥T → 2f (x) = y), where f is a metavariable represents either start time or end time. We stipulate object types of binary relations, e.g. 4. ∀x∀y(produces(x, y) → (Task (x) ∧ Product(y)) ∨ (TaskPerformance(x) ∧ ProductInstance(y))).
296
S. Ito et al. / A Formal Ontology for Business Process Model TAP: Tasks-Agents-Products
Our ontology contains this kind of axioms for each binary relation in Figure 1. This type of axioms are for pragmatics of TAP. We have axioms for inter-relationships between abstrat level and istance level such as: 5. ∀x∀y∀z(activates(x, y) ∧ of (y, z) ∧ AgentEnaction(y) → ∃u∃v(is instance of (x, u) ∧ is instance of (z, v) ∧ performed by(u, v))). Agent enaction connects an agent instance to a task performance, describing e.g. when an agent is assigned to a task performance. In other words, agent enaction is a collection of information of an agent relevant to the task performance. Therefore, all agent enactions have a connection to some agent instance. We have the similar axiom for ToolEnaction. 6. ∀x(AgentEnaction(x) → ∃yof (x, y)). Moreover, each agent enaction or tool enaction is peculiar to some agent instance or tool instance. 7. ∀x∀y∀z(of (x, y) ∧ of (x, z) → y = z). Our ontology contains axioms of inter-relationship between is instance of relation and constitutes relation. This relationship is similar to the relationship between is-a relation and has-a relation [4], but there are some differences. We show such an axiom in the following: 8. ∀x∀y∀u∀v(constitutes(x, y) ∧ is instance of (x, u) ∧ is instance of (y, v) → constitutes(u, v)). In this axiom, u and v are tasks and x and y are task performances (this fact is derived from other axioms omitted in this paper). This do not hold for relationships between is-a and has-a. For example, let consider v as “car” and u as “air-conditioner”. y is an instance of car and x is an instance of air-conditioner. In this case the fact that y has x do not imply the fact that v has u, because some types of car (e.g. formula cars) do not have air-conditioners. However, TAP has this property since this axiom is a relationship of tasks and task performances. constitutes relations between task performances should conform to constitutes relations between tasks. Otherwise, the performance of some task may not be executed according to its substructure. def We define the macro formula time(x, y) = start time(x) = y ∨ end time(x) = y and def set time(x, y) = (start time(x, ⊥T ) ∧ 3start time(x, y)) ∨ (end time(x, ⊥T ) ∧ 3end time(x, y)). time(x, y) means that a time y is set to the start time or the end time of some object x. set time(x, y) means that a time y will be set to the start time or end time of some object x. 9. ∀x∀T y(set time(x, y) → ∀u∀T v(time(u, v) ∧ v = ⊥T → v ≤ y)). This axiom says that if a time y is set to some x in a state and a time v is set to some u in a future state reachable from the state then y ≤ v. One of conspicuous features of TAP is the notion of meta-task. Meta-tasks are tasks of planning, monitoring, and executing processes and are defined as tasks which produce or consume TAP specifications. The following are axioms for meta-tasks. 10. ∀x(MetaTask (x) → Task (x)). 11. Product(TAPspec). 12. is instance of (spec(A1 , . . . , An ), TAPspec). 13. ∀x(MetaTask (x) → ∃yconstitutes(x, y)). 14. ∀x∀y(MetaTask (x) ∧ (produces(x, y) ∨ is consumed by(y, x) → y = TAPspec). 15. ∀x∀y∀z((produces(x, y)∨is consumed by(y, x))∧is instance of (x, z)∧MetaTask (z) → is instance of (y, TAPspec)). In the set of axioms, meta-tasks are accounted as a task which produces and consumes TAP specifications. However, what is the semantics of producing or consuming TAP specifications was not specified. Such semantics is dependent on the actual work of a meta-task, but the kinds of such tasks are not many. For example, planning, evaluating, monitoring and executing are brief classification of meta-tasks. The semantics of producing and consuming TAP specifications are also classified according to these classes. We are now interested in
S. Ito et al. / A Formal Ontology for Business Process Model TAP: Tasks-Agents-Products
297
classifying meta-tasks and characterize them as formulas in our logical language. For this, we may have to augment the expressivity of our language. 6 Conclusion We introduced the semantic structure for TAP and the logical language to describe TAP specifications. Then we presented some axioms from our ontology which specifies the characteristic of the world structures of TAP in our logical language. There are several significance in this formal approach. From an ontological viewpoint, the axioms characterize the intended world structures which TAP can conceptualize and account for the intensional meanings of entities and relationships. Thus we can understand what is the world structure of business aspects of real worlds which TAP intends to model. This includes the understanding of the nature of entities occurring in business processes, e.g. what the task performance is, what the enaction of tasks is and what the meta-task is, etc. By axiomatizing formally, the ontology has the ability to deduce facts about entities and relationships in TAP business process models. We can evaluate the expressivity and adequacy of business process model TAP by the ability of deduction. From an engineering viewpoint, we can elucidate ambiguous points in syntax and semantics of TAP. We can automatically check the syntactic and semantic consistency of TAP specifications e.g. the consistency of start times and end times of task performances or the consistency of object types and relationships, etc. Therefore we can use business process model TAP in planning, controlling, improving and monitoring business processes with confidence. We can verify the properties of processes rigorously by checking whether the models of some TAP specification satisfy some property. Both the TAP specification and the property are expressed in our logical language. We want to extend our ontology to cover features such as generalization and classification which was not treated in this paper. Another important topic is to compare our ontology and other business ontologies in formal level. For this, we should formalize these ontologies. References [1] Mark S. Fox and Michael Gruninger. Enterprise modeling. AI Magazine, 19(3):109–121, 1998. [2] Guido L. Geerts and William E. McCarthy. An accounting object infrastructure for knowledge-based enterprise models. IEEE Intelligent Systems and Their Applications, 14(4):89–94, 1999. [3] Jaap Gordijn and Hans Akkermans. Value based requirements engineering: Exploring innovative ecommerce idea. Requirements Engineering Journal, 8(2):114–134, 2003. [4] Naoko Izumi and Naoki Yonezaki. A logic of ontology for object oriented software component. In Proceedings of the 11th European-Japanese Conference on Information Modeling and Knowledge Bases, pages 83–99. Amsterdam, IOS Press, 2001. [5] W. E. McCarthy. The REA accounting model: A generalized framework for accounting systems in a shared data environment. The Accounting Review, 57(3):554–78, 1982. [6] Mike Uschold, Martin King, Stuart Moralee, and Yannis Zorgios. The enterprise ontology. The Knowledge Engineering Review, 13(1):31–89, 1998. [7] Naoki Yonezaki, Tapani Kinnula, Motoshi Saeki, and Jan Ljunberg. TAP: A new model for software process: Tasks-Agents-Products. In Proceedings of the 5th International Conference on Software Engineering and Knowledge Engineering, pages 346–350, 1993.
298
Information Modelling and Knowledge Bases XIX H. Jaakkola et al. (Eds.) IOS Press, 2008 © 2008 The authors and IOS Press. All rights reserved.
A proposal for student modelling based on ontologies Angélica DE ANTONIO1, Jaime RAMÍREZ1, Julia CLEMENTE2 1 Facultad de Informática. Universidad Politécnica de Madrid. 28660 Boadilla del Monte, Madrid, Spain e-mail: {angelica, jramirez}@fi.upm.es 2
Universidad de Alcalá. Escuela Universitaria Politécnica Departamento de Automática Campus Universitario. Ctra. Madrid-Barcelona, Km. 33,600 28871 Alcalá de Henares, Madrid, Spain e-mail: [email protected]
Abstract. The advances in the educational field and the high complexity of student modelling has provoked it to be one of the more investigated aspects in Intelligent Tutoring Systems (ITSs). The Student Models (SM) should not only represent the student's knowledge, in a wide sense, but rather they should be, insofar as it is possible, a snapshot of the student's reasoning process. In this article, a new approach to student’s modelling is proposed that benefits of the Ontological Engineering advantages, so widely used at the present time, to advance in the pursue of a more granular and complete knowledge representation. The goal is to define an ontological basis for SMs characterized by a high flexibility for its integration in varied ITSs, a good adaptability to the student’s features, as well as to favor a rich diagnostic process with nonmonotonic reasoning capacities, allowing the treatment of the contradictions raised during the student's reasoning and diagnosis.
1. Introduction In spite of the tendencies in the educational field, in constant evolution, with new approaches to Intelligent Tutoring Systems research pushing this constant progress, the construction and maintenance of ITS’s modules is complex and it still presents many lacks; Artificial Intelligence and Software Engineering are bound to play a crucial role for their resolution and continuous improvement. Some authors like Mizoguchi & Bourdeau [1] have attributed the current limitations of these systems, primarily, to a lack of a explicit representation of the conceptualization on which each system is based. An extensive revision of the state of the art in student's modelling, a distinctive feature of ITSs, has leaded us to corroborate this statement. The purpose of this article is proposing a new approach to student’s modelling based on Ontological Engineering, following Mizoguchi and Bordeau [1], but, beyond the approach taken by these authors, introducing a new student's modelling taxonomy that has been built after a rigorous analysis of the types of knowledge about the student that can be represented in a SM. This generality will enable the adaptation of the student’s model to different types of ITS, and will facilitate the construction of ITS which are truly adaptive, with tutoring moulding to the student's individual features. Our approach also facilitates an appropriate and powerful cognitive diagnosis, with nonmonotonic reasoning capacities.
A. de Antonio et al. / A Proposal for Student Modelling Based on Ontologies
299
The present article starts with a brief description of the state of the art in Student Modelling, proceeds with an analysis of the motivation and general objectives of our work, and continues with a description of the adopted solution. We have centred our focus in pedagogic design, upon which our solution is sustained, and in the ontology proposed for the SM, with only a sketch of how the diagnostic process is approached. The conclusions and the current and future work lines related to the proposal put an end to the paper. 2. Previous Work in the Area So far, numerous approaches to SM have been proposed in the field of ITS, representing different information types and using different methods to infer the student's cognitive state [2], [3]. They can be classified as: x SMs that just represent the state of the student’s knowledge about the subject matter, including SMs that only represent correct knowledge (Overlay Models such as in [4], [5], etc., or Differential Models such as in [6], [7], etc., present the drawback that the student's knowledge is usually not strictly a subset of an expert’s knowledge) and SMs that also represent wrong knowledge with different approaches to the development of the error library [8], [9], etc. (the consideration of possible errors improves the understanding of the student). x SMs that also represent the student’s reasoning process: according to Clancey [10], these can be divided into Behavior simulation models, that only describe the actions the student is carrying out ([11], [12]), and Functional simulation models, that describe the student’s beliefs and goals, what the student knows and what he’s trying to do ([13], [14]), etc.). Some taxonomies for the student's knowledge modelling have been contributed in this field similar to the one used for the previous classification ([15], [16], etc.) and others deserve to be highlighted by their interesting contributions such as: a) the taxonomy in the De Koning and Bredeweg approach [17] based on the multi-stratified framework KADS [18] distinguishes the strategic knowledge (allowing the representation of the goals in problem solving, how to reach them, and the knowledge required for reasoning with them). In this respect, it is more and more recognized that the student's metacognitive process should be considered in the educational process (an example is the system TAPS [19]). b) the McCalla and Greer’s taxonomy [20], sustained in the idea of granularity based reasoning (level of detail in the vision of a concept). This feature, incorporated explicitly in an ITS, can facilitate the diagnosis of the student's behaviour. It is also important to point out there are not many works that consider the student's personal features to carry out an adaptive teaching-learning process. Some examples are [21], or the Chen and Mizoguchi’s proposal [22], where an ontology and an agent for the SM are defined. It is this latter work the one that has served as a starting point for the proposal presented in this article. However, their ontology suffers from important limitations, such as: a) lack of information related to the student's learning objectives; b) scarce information on most of the considered knowledge types and; c) in general, lack of clarity in the description of the concepts as well as in its organization.
300
A. de Antonio et al. / A Proposal for Student Modelling Based on Ontologies
3. Objectives and motivation After an analysis of the state of the art in SMs, shortly described in the previous section, we observed that most approaches don't consider a complete taxonomy of knowledge about the student; also, most of them have validity only in certain domains or they are hard to be adapted for different ITSs. At the same time, most of them neither consider the student's individual features in detail nor do they facilitate a complete cognitive diagnosis, with non monotonic reasoning capacities, in line with the nature of human reasoning. In order to face those limitations, we propose the design and implementation of a SM mechanism that presents a distinctive group of features: Wide student knowledge taxonomy, capable of expressing many types of knowledge about the student, that will allow the tutoring module to carry out a more adaptive tutoring, including: a) The “student's profile”, depicting the student's psychological profile, learning style, previous experience in the targeted area by the course, etc.; b) The explicit representation of the learning objectives that the student should reach (an important feature included in the model) in several domain levels (knowledge, affective and psychomotor). This will facilitate the design of a new cognitive diagnosis method which is based, not only in the model of the student’s knowledge, but also in the trace of the student’s motions and physical actions throughout the specific activity that he is conducting, being able, at the same time, to provide better explanations and help during the learning process; c) The representation of various aspects of student's learning, some of them dependent on the activity he’s carrying out and other ones independent; and d) The knowledge on the student's cognitive state, related to the diagnostic phase. A powerful knowledge representation formalism that allows a rational concept representation (with different abstraction levels), and that also supports sharing and reusing knowledge. Ontologies have helped us to achieve these goals in the formalization of the SM, representing different knowledge granularity levels explicitly [18]. In this way, SMs can be easily developed, extended, and reused in other learning environments. A new diagnosis method of the student's knowledge state, with nonmonotonic reasoning capacities that are adjusted to the also non-monotonic nature of student’s modelling. Among the possible non-monotonic reasoning techniques, assumption-based reasoning has been selected as the support for our diagnosis method. It allows managing incomplete knowledge (about the student's cognitive state) by formulating hypotheses, so that the reasoning process can go on. However, if some of the hypotheses that have been assumed are ever refused during the reasoning process, these hypotheses must be retracted, and all the conclusions derived from them must be removed. In order to make more efficient this process and others related to the conflict resolution, an Assumption-based Truth Maintenance Systems –ATMS- has been employed. 4. Adopted solution The development of the proposed solution for the SM was inspired, from the beginning, in the pedagogic design approach that is shown schematically in the Figure 1.
A. de Antonio et al. / A Proposal for Student Modelling Based on Ontologies
Instructional design for the subject matter (X) Defining of: x A group of activities. x The objectives that the student should achieve in each activity.
x
x
Figure 1. Proposed architecture for the ITS
301
Automated planner Setting of the steps or actions (applications of operators) that should be carried out to conclude the activity correctly. Allowing dynamic construction of solution plans taking into account the current state of the learning environment and the possible student’s actions.
When the student executes a certain action (operator), this execution is registered according to the SM ontology, which not only contains different concepts but also relationships among them (such as the ones that relate the learning objectives -meaningful for the tutoring module- and the knowledge objects -meaningful for the expert module- that the student should acquire in order to be able to reach those objectives). The diagnosis about the SM is divided into two modules: the Pedagogic Diagnosis (PD) and the Cognitive Diagnosis (CD). Based on what action the student performs and how (registered in the ontology) and on the objectives that have already been reached when the action is executed, PD will take on the responsibility of determining the new objectives reached by the student. For that purpose the PD uses a group of diagnostic rules. On the other hand, based on the reached objectives and on the knowledge objects associated with them, CD infers the concrete knowledge state of the student. During the diagnostic process diverse types of contradictions can arise that the Conflicts Manager must solve. This capability will be based on an ATMS system and a conflict solver. 4.1 Detailed description of the Ontology To represent the SM the solution that has been adopted is based on ontologies, using OWL as the representation language, and Protégé [23] as the tool for its construction. Next, the top level classes of the ontology are defined in detail: Student_Activity_Record and its subclasses describe the trace that a student generates during a session (Figure 2). The properties of this subclass represent the start and finalization time of the register, respectively. Description Describes the trace of a certain variable that can be observed in the student's behaviour with a certain frequency (samplingfrequency). Subclasses: Emotional_Trace and Trayectory_Trace Describes a certain variable that can be observed in the student's behaviour (for instance, the student's Position, View, etc.)
Figure 2. Student_Activity_Record hierarchy of subclasses on the SM ontology
302
A. de Antonio et al. / A Proposal for Student Modelling Based on Ontologies
Student_Profile. Represents student's personal information. In the Figure 3 its Description general subclasses can be observed. The demographic data of the student (age, civil state, name and sex) The student’s experience level with computers, and his activity in the area Physical aspects that can affect the student’s learning (corporal sizes, disablements, etc.) The preferences of the student towards different interaction means, entry and exit devices Student’s preferences in order to face learning (practice oriented, principle oriented, example oriented, etc.) Its properties specify the name of the area targeted by the course and the student’s experience in this field The main student’s psychological features (personality features, the student’s disposition towards what he’s going to learn, etc.)
Figure 3. Student_Profile hierarchy of subclasses on the SM ontology
Learning_Objectives. To specify objectives for a course in one or several domains (Figure 4). Three taxonomies have been considered to define the subclasses: Krathwohl’s taxonomy (affective level [24]), Bloom’s taxonomy (cognitive level [25]) and Harrow’s taxonomy (psychomotor level [26]). Each objective has the following properties: identification, an associated valuation and the knowledge objects associated to it. Description
Subclasses
Deals with emotional abilities (fear control, empathy, self-regulation, etc.), and attitudes. Refers to knowledge structures Deals with physical and movement capacities, as well as coordination
Objective_Knowledge Objective_Comprehension Objective_Application Objective_Analysis Objective_Synthesis Objective_Evaluation
Figure 4. Learning_Objectives hierarchy of subclasses on the SM ontology
Learning_Valuation describes, for a student, certain data derived from the student trace during the learning session. These data will be used mainly by the Tutoring Module. Its subclasses are shown in the Figure 5. Description
Valuation of the student’s ability: memory, reasoning, etc. Valuation of the student’s objectives degree of achievement, master level (beginner, intermediate, etc.) Data referred to: the variables observed, the actions performed (for instance: number of times that the operators were applied correctly according to the plan, number of times that the applied operators were not in the plan, number of times that the student tried to execute a operator using a wrong object, number of asking questions, etc., and the valuation of student’s actions, acting factor, obtained from the previous properties), the specific execution of the activity, the general knowledge of the activity, etc. in a learning session. Valuation of general data such as: success/failure rate, number of questions, number of correct/incorrect answers, etc.
Figure 5. Learning_Valuation hierarchy of subclasses on SM ontology
A. de Antonio et al. / A Proposal for Student Modelling Based on Ontologies
303
Knowledge_Object. Describes the main knowledge element types that can be acquired in a certain educational activity (Figure 6). Description Sequence of actions (property includeSequenceActions) Any operation type that a student can perform in the learning environment. Properties: operator, preconditions, postconditions and role.concept Subclasses: Collaborative_Action, Individual_Action (action The subclasses added to the hierarchy were, among other, without contactingStymulus with any and object), Interaction_Object_Action Point Sensorial Object The steps or plan elements that should be carried out to achieve a certain educational activity (property isFormedbyBlocks) A step inside the plan (property positionRelativePlan). Subclasses: Compound_Action, a set of plan elements (can be a Sequence_Block, when its elements should be executed in a strict order or Unordered_Block, when its elements can be executed in any order) and Application_Action, the application of a concrete action described by an instance of Puntual_Action and the time interval of the action application Its properties are domain, nameRelation, range, reflexive, symmetrical, transitive, etc. Subclasses: Definition, Exacts_Sciencies_Proposition, Natural_Sciences_Proposition and Theory (and its subclasses)
Figure 6. Knowledge_Object hierarchy of subclasses on the SM ontology
Knowledge_State describes the information derived from the student behaviour during the learning session. The Objectives_Diagnosis specifies the learning objectives that the student has demonstrated to have reached (deduced by means of the pedagogic diagnosis). Their subclasses are shown in the Figure 7. Description
Diagnosis based on the questions asked by the student Diagnosis based on the analysis of the student activity Diagnosis based on the attempts of action executions Comprises information both related to the right knowledge of the student and the contradictions detected in the student
Figure 7. Knowledge_State hierarchy of subclasses on SM ontology
4.2 Diagnosis Rules for the Student Model According to the design adopted for the Student’s Diagnosis, there is the need to define a group of rules to carry out the first phase of Pedagogic Diagnosis. These rules will infer the new learning objectives reached taking into account the actions performed by the student and the already reached objectives inferred from the previous student behaviour. Certain rules can infer that the student has not achieved a certain objective; in this case the information that the SM provides on the student's trace will be indispensable to determine if the student has forgotten some knowledge or if he has never achieved those objectives.
304
A. de Antonio et al. / A Proposal for Student Modelling Based on Ontologies
The pedagogic diagnostic rules have been grouped according to certain rule patterns: Diagnosis according to the type of action that a student performs. These rules will infer the learning objectives that can be assumed whenever the student executes correctly/incorrectly a given action depending on the relevancy and appropriateness of the action. There are rule patterns that consider if the action is correctly executed but it is not in the target sequence of actions, if the action is in the plan but the student executes it in the wrong order, if it is improper to execute the action because some of the preconditions associated to the operator of the action are not met, if the student tries to apply the right operator but to the wrong object, etc. (for example, if the student picks up a designated visible object correctly, it can be assumed that s/he is able to recognize the appearance of the object). Diagnosis based on the number and type of questions formulated by the student. This provides information on the degree of knowledge that the student has of the existent objects in the scenario, of the operators, or of the own activity, depending on the type of question (what is this object for? Where is the object X? What should I do next? Why can’t I do this? What would it happen if I do this?...). The diagnostic rules will also consider the hints and instructions provided by the tutor as knowledge that can be assumed that the student has already acquired. 5. Conclusions and Future Work This article has described a solution based on ontologies to model the student in an ITS. The general objective has been developing a SM with the following main characteristics: generality, adaptability, non-monotonic diagnosis, extensibility, and reusability. As a proof of concept, we are currently concluding the adaptation and extension of the ontology for its application to Virtual Learning Environments, after a previous instantiation for the case of learning the interaction with the graphical user interface of software applications. This extension involves expanding the SM diagnostic rules with new information about the path followed by the student during their navigation through 3D scenarios, non-verbal behaviour such as gaze direction, hints or instructions that the tutoring module can provide to the student, new question types that the student can ask in this kind of environments and how can influence what the student knows a priori, etc. The non-monotonic part of the diagnosis method is also being completed with the design of the Conflict Manager. The implementation of the proposed diagnosis method will rely on an ATMS and a reasoner like Jena or the toolkit SweetRules. The last step will be improving the tutoring strategies by exploiting the proposed SM. 6. References [1] Mizoguchi, R. and Bordeau, J. Using ontological engineering to overcome common AI-ED problems. International Journal of Artificial Intelligence in Education, Vol. 11. [2] Petrushin, V. A. Intelligent Tutoring Systems: Architecture and Methods of Implementation. Journal of Computer and Systems Sciences International, Vol. 33, No.1, pp. 117-139, 1995. [3] Holt, P., Dubs S., Jones, M. and Greer, J. The State of Student Modelling. Student Modelling: The Key to Individualized Knowledge-Based Instruction. Springer-Verlag, pp. 3-39, 1994.
A. de Antonio et al. / A Proposal for Student Modelling Based on Ontologies
305
[4] Clancey, W. J. GUIDON. Journal of Computer-Based Instruction, vol. 10, No. 1, pp. 8-14, 1983. [5] Carr, B. y Coldstein, I. Overlays: a theory of modelling for computer-aided instruction. International Journal of Man-Machine Studies, 5, pp. 215-236, 1977. [6] Holt, P., Dubs S., Jones, M. and Greer, J. The State of Student Modelling. Student Modelling: The Key to Individualized Knowledge-Based Instruction. Springer-Verlag, pp. 3-39, 1994. [7] Burton, R. R. y Brown, J.s. A tutoring and student modeling paradigm for gaming environments. ACM SIGCSE Bulletin, vol. 8, 1, pp. 236-246, 1978. [8] Clancey, W. J. Qualitative student models. Annual Review of computer Science, J. F. Traub, editor, 1, pp. 381-450, 1986. [9] Burton, R. R. Diagnosing bugs in a simple procedural skill. Intelligent tutoring systems. Ed. Sleeman, D. H, pp. 157-184, Academic Press, 1982. [10] Clancey, W. J. Qualitative student models. Annual Review of computer Science, J. F. Traub, ed, Vol. 1, pp. 381-450, 1986. [11] Kass, R. Student Modelling in intelligent tutoring systems –implictions for user modelling. User Models in Dialog Systems, eds. A. Kobsa y W. Wahlster, pp. 386-410, Berlin: Springer-Verlag, 1989. [12] Martin, B. Constraint-Based Modeling: Representing Student Knowledge. New Zealand Journal of Computing, vol. 7, No. 2, pp. 30-38, 1999. [13] Katz, S. and Lesgold, A. Modelling the student in SHERLOCK II. Student Modelling: the Key to Individualized Knowledge-Based Instruction, eds. Greer, J. E. and McCalla, G., vol. 125, pp. 99-125, Berling Heidelberg: Springer-Verlag, 1994. [14] Martin, Brent. Constraint-Based Modeling: Representing Student Knowledge. New Zealand Journal of Computing, vol. 7, 2, pp. 30-38, 1999. [15] Dillenbourg, P. and Self, J. A framework for learner modelling. Interactive Learning Environments, Vol. 2, No. 2, pp. 111-137, 1992. [16] Baffes, P. T. and Mooney R. Using theory revision to Model Students and Acquire Stereotypicar Errors. Proceedings of the 14 th Annual Conference of the Cognitive Science Society, pp. 617-622, Bloomington, 1992. [17] Koning, K. and Bredeweg, B. Using GDE in Educational Systems. Proceedings of the twelf International Workshop on qualitative reasoning, pp. 42-49, 1998 [18] Wielinga, B. J. Schreiber, A. Th. and Breuker, J. A. KADS: A modelling approach to knowledge engineering, vol. 4, No. 1, pp. 5-53, 1992. [19] Derry, S.J. Metacognitive models of learning and instructional systems design. Adaptative Learning Environments: Foundations and Frontiers, eds. Jones, M. y Winne P., NATO ASI Series F, Berling Springer-Verlag, Vol. 85, pp. 257-286, 1992. [20] McCalla, G.I. and Greer, J.E. Granularity-based reasoning and belief revision in student models. Student Modelling: The key to Individualized Knowledge-based instruction, eds. Greer, J.E. and McCalla, G.I, Springer-Verlag, pp. 39-62, 1994 [21] Del Soldato, T. Detecting and reacting to the learner’s motivational state. Proceedings of the 2nd International Conference on Intelligent Tutoring Systems, Montreal, Quebec, eds. Frassson, C., Gauthier, G. y McCalla, G., Lecture Notes in Computer Science, Vol. 608, pp. 567-574, 1992 [22] Chen, W. and Mizoguchi, R. Learner Model Ontology and Learner Model Agent. Cognitive Support for Learning –Imagining the Unknown. Eds. P. Kommers, IOS Press, pp. 189-200, 2004. [23] Knublauch, H., Fergerson, R. W., Noy, N. F. and Musen, M. A. The Protégé OWL Plugin: An Open Development Environment for Semantic Web Applications. Lecture Notes in Computer Science, Vol. 3298, pp. 229-243, 2004. [24] Krathwohl, D. Taxonomy of educational objectives: The classification of educational goals: Handbook 2: Affective domain. New York: Longman, Inc. [25] Bloom, B. S. Taxonomy of educational objectives. Published by Allyn and Bacon, Boston, MA. Pearson Education. [26] Harrow, A. J. A taxonomy of the psychomotor domain: a guide for developing behavioural objectives. Ed. Longman, 1972.
306
Information Modelling and Knowledge Bases XIX H. Jaakkola et al. (Eds.) IOS Press, 2008 © 2008 The authors and IOS Press. All rights reserved.
Ontology-based Support of Knowledge Evaluation in Higher Education Andrea Kő, András Gábor, Réka Vas, Ildikó Szabó Corvinus University of Budapest, Faculty of Business Administration Department of Information Systems; Veres Pálné u. 36., Budapest, 1053., Hungary Abstract: This paper demonstrates research activities regarding ontology-based knowledge evaluation in Higher Education. Conceptualization initiatives for educational domain are currently a hot topic and at the same time a challenge. In the paper we demonstrate an adaptive knowledge testing and evaluation system supported by the educational ontology, which helps the students to explore missing knowledge areas and guide them to the material which has to be further studied. The system is the outcome of fourteen higher institutes in Hungary, so we have gathered several experiences during the system test (which was organized and performed by the participating higher institutes and their students), which we highlight in the article. Current phase of the research concentrated to the curricula of the Business Informatics program, as a test environment. We outline our further improvement and refinement also.
1. Introduction First initiative, which proposed the creation of the Higher Education European area as a key enabler to promote citizens mobility and employability was the Sorbonne Joint Declaration of 25th of May in 1998. These statements emphasized by Bologna declaration (1999), which initiated reforms of European Higher Education, also pointing out its crucial role in social, economic and human growth of the Continent. Additional goals of the Bologna process are the following: • Adapting easily readable and comparable degrees, • Adapting a system essentially based on two main cycles, • Establishing a system of credits, • Promoting mobility, • Promoting European co-operation in quality assurance and • Promoting the necessary European dimensions in higher education. The Berlin Communiqué (2003) stressed the goal of introducing a common framework of transparent and comparable degrees that ensures the recognition of knowledge and qualifications of citizens all across the European Union. It extended the Bologna Process with a third, doctoral cycle. European Higher Education Area is structured around three cycles where each level has the function of preparing the student for the labour market, for further competence building and for active citizenship. They aimed at developing descriptors for Bachelor’s and Master’s that can be shared within Europe and be
A. K˝o et al. / Ontology-Based Support of Knowledge Evaluation in Higher Education
307
used for a variety of purposes, depending on particular national, regional or institutional contexts and requirements. This was one of the first initiatives, which provided support for facilitating the comparison of degrees. The launch of these Dublin descriptors also indicates that competences should have a key role in providing transparent and comparable curricula and qualifications. Hungarian Government joined to the above mentioned initiatives. A large-scale reform was decided several years ago that aims to modify the Hungarian higher educational system, both the educational structure, and the operating model. Our research entitled “HEFOP – “Development of Knowledge Balancing, Short Cycle e-Learning Courses and Solutions” aimed at developing competitive training evaluation system and promoting the transition between the different levels of higher education (BSc, MSc levels). According to the complex nature of the investigated domain we applied ontological background during educational domain conceptualization. A further goal of the research is to provide support for the adaptive knowledge testing and evaluating of students in order to help them to complement their educational deficiency. Educational Ontology played a crucial role in this process. The adaptive testing model itself consists of two main modules: the Test Module, which consists of the Educational Ontology, Testbank; and Adaptive Examination System (AES); and the e-Learning environment, which contains a Learning Management System and a Learning Content Management System (LCMS). Figure 1 depicts the architecture of this adaptive testing model. The curricula of Business Informatics will be analysed in this research. Several approaches are available for developing curricula on the field of computing (ACM, AIS and IEEE-CS, 2005). However these do not cover all local requirements and do not fit the above discussed Hungarian specialities. User
User interface (LMS)
LCMS
AES
Educational Ontology
Testbank
Figure 1: Adaptive Test Model in e-learning environment
The paper will focus on summarizing the evolution and results of the above mentioned research project. Accordingly Section 2 describes conceptual model of the Educational Ontology. In this conceptual phase we applied our own modelling approach, and this assures formalization of the given knowledge area in standard ontology languages (e.g.: in OWL DL). (Corcho and Gómez-Pérez, 2000; 2002) Section 3 discusses the theoretical background and characteristic of computerized adaptive testing, while Section 4 demonstrates the implementation environment of the ontology and the testing system. Results are summarized in Section 5.
2.
Educational Ontology
For the ontology development we applied the Sure –Studer methodology, but in this section we will discuss in detail the kick-off phase (Sure and Studer, 2003). A challenge of modelling is that the scope of curricula taught in Business Informatics training program is
308
A. K˝o et al. / Ontology-Based Support of Knowledge Evaluation in Higher Education
wide and curricula are substantively different in nature (e.g. modelling of Knowledge Management curriculum may require different approach, then the modelling of Mathematics). Moreover it should be also taken into consideration that the structure and content of a subject may be at least partly different in different institutions. Accordingly in the first cycle of development the research has concentrated on defining the major classes of the ontology and taxonomy, pointing out the role of competences and concentrating on facilitating comparability. Knowing the amount of work required to produce ontologies even for the simplest concepts, in the course of ontology development we focused on providing easily definable and applicable classes and precise determination of relations, also keeping the goal of knowledge testing in view. This section gives a description of all of the classes in the ontology. Competences have played central role in the first version of the ontology model to enable grabbing the common features of different curricula. In higher education, accreditation documents provide a list of goals of the given training program in the form of competencies. This means that competencies and curricula of the training program must be aligned. Accordingly classes of “Competence module” and “Curricula module” were formed and connected to each other with the “belongs to” relation in the ontology to enable tracing of knowledge and competences possessed by students. Modules represent standardized units (of curricula or competences) that facilitate the comparison of curricula and competences of different institutions and universities. Curricula are modelled by defining their major parts that we call knowledge areas. Knowledge Area” is the super class of the ontology, representing major parts of a given curriculum. Each “Knowledge Area” may have several ”Sub-Knowledge-Areas”. Not only the internal relations, but relations connecting different knowledge areas are also important regarding knowledge testing. The “is part of” relation is still an important element of the model connecting knowledge areas and sub-knowledge-areas in the model. At the same time a new relation has to be introduced, namely the “requires knowledge of” relation. This relation will have an essential role in supporting adaptive testing. If in the course of testing it is revealed that the student has severe deficiencies on a given knowledge area, then it is possible to put questions on those areas that must be learnt in advance. For the sake of testing all of those elements of knowledge areas are also listed in the ontology about which questions could be put during testing. These objects are called “Knowledge Elements” and they have the following major types: “Basic concepts”, “Theorems” and “Examples”. In order to precisely define the internal structure of knowledge areas relations that represent the connection between different knowledge elements also must be described. (Vas, 2006) Figure 2 depicts all the above-discussed elements of the ontology that together form the internal structure of knowledge areas. The following markings are used on Figure 2: • Rectangles sign classes. • Arrows depict 0-N relations (e.g.: a competence may have several prerequisites, and it is also possible that a competence does not have any prerequisites).
A. K˝o et al. / Ontology-Based Support of Knowledge Evaluation in Higher Education
Competence
prerequisite
requires is part of
Basic concept
ensures
Knowledge Area
is part of
refers to
element of
is part of premise
309
Competence Module prerequisite
element of belongs to requires knowledge of
Curriculum Module
is part of
prerequisite
refers to
Theorem
Example refers to
conclusion refers to
Test questions
Figure 2: Educational Ontology Model
Test question do not form a part the ontology, but at least one test question must be connected to all major components (knowledge area, basic concepts, theorems, examples) of the ontology. This way to depict the difference test questions are connected to the ontology components with dotted lines.
3. Adaptive Knowledge Evaluation and Testing The main principles of adaptive testing also have to be analyzed to enable the development of an adequate testing system and its connection with the ontology. The main idea of adaptive testing is that the test should tailor itself to the estimated ability level of test takers and take into account how each test taker has answered previous questions (Linacre 2000). The basic principles of computer adaptive testing are provided by Thiessen and Mislevy (1990): • Test can be taken anytime, no need of group-administered testing. • There are no identical tests, as every test is tailored to the needs and capabilities of the test-taker. • Questions are presented on a computer screen. • After the answer is confirmed there is no chance to change it. • The examinee is not allowed to skip any of the questions • The questioning process is fully and dynamically controlled. Our research project aims at implementing an interface, which is used in a customized qualification program development, based on the individual’s pervious qualifications, completed levels, corporate trainings and practical experiences, in case of entering a certain educational level. Two main groups of input are needed to build up a qualification program. On one hand the individual’s knowledge and abilities must be measured, on the other hand a definition must be given about the prerequisites of the targeted qualification, which depends on the quality assurance and the accreditation system of higher education. After testing the individual’s knowledge, a customized supplementary training program should be allocated. A corresponding adaptive test provides help to the individual, who draws on this service. If the candidate passes the exercises and tests successfully, than the prerequisites for the certain qualification are fulfilled, so the student may enrol to the targeted level. As an additional benefit, this solution may be used for correcting the deficiencies of a certain curriculum during the qualification, as an ad-hoc support of education. Beside the Educational Ontology another pillar of the testing system is the set of test questions. Main characteristics of test questions should be the following: • A question must be connected to one or more Knowledge Elements or Knowledge Areas. On the other hand a Knowledge Element or Knowledge Area
310
A. K˝o et al. / Ontology-Based Support of Knowledge Evaluation in Higher Education
may have more then one test question. This way the Testbank is structured by the Educational Ontology. • All questions should be weighted according to their difficulty. Test questions will be provided in the form of multiple choice questions. So parts of the question must be the following: (1) question, (2) correct answer, (3) false answers. Figure 2 also shows how test questions connect to the elements of the ontology. Test questions are connected with dashed lines to the ontology, indicating that they don’t form a part of the ontology. Last but not least the algorithm of the testing procedure was also worked out in the frame of this research project. The testing procedure starts the examination at the top of the hierarchy of knowledge areas. It gives the student a test set having so many questions that cover the given knowledge area. If he answers properly – viz. the sum of points received for his answers reaches a given level (for example 60%) – we put questions about all the basic concepts and all the sub-knowledge-area of this knowledge area. If the student does not know the answer for the question related to the basic concept the knowing of the knowledge area will be refused. If he answers badly for some sub-knowledge-area the knowledge area and these sub areas will be not accepted, but if there are some subknowledge-area whose questions were answered properly then the testing engine interrogate them again in the previous manner. Namely the testing engine executes a depth first graph search algorithm in such manner that it closes a branch if the student does not know the given knowledge area or its sub-knowledge-areas (all of them) or a given basic concept.
4. Implementation To implement this Educational Ontology model we have to choose an adequate ontology editor which meets the following requirements: • extensible: the training system has to meet the requirements of labour markets so it has to be developed continually; • treatment of high volume data: the curricula contents consist of several knowledge areas, basic concepts, theorems etc.; • interoperability: many teachers, lecturers may be involved in building this ontology so it is necessary to ensure the access and usage to the system; • user friendly interface Accordingly the prototype was implemented by using Protégé, which is which is the most known free ontology editor tool, also being the most widely used one. For practical reasons we had to make several changes in the prototype as compared with the conceptual model: (1) In the case of the prototype the relation called is part of had to be corresponded to the relation with specific name, for example has part (theorem), has part (basic concept) etc. Because in the course of filling the model with knowledge elements it was difficult to distinguish which is part of relation is applied to the relation between the knowledge area and the theorems or which one is applied to the relation between the knowledge area and the basic concepts, etc. (2) To test the procedure easily the database of multiple choice questions is built in the knowledge model in form of a class. This class is related to the basic concepts, theorem and knowledge area classes. (3) To realize the testing system the set of test items called Testbank has been built into the Protégé project. To verify the applicability of the testing procedures Java inference engine was developed. Protégé is a Java-based ontology editor also (Protégé, 2006). It provides an interface which makes the knowledge-base accessible to other applications. These applications do not need to use Protégé graphical interface. The protégé.jar file includes the
A. K˝o et al. / Ontology-Based Support of Knowledge Evaluation in Higher Education
311
getKnowledgeBase() method inside the edu.stanford.smi.protege.model.Project class. The Protégé java documentation contains information about this class and related classes. So the application written in Java facilitates the common availability and interoperability among the users of this system. It ensures a graphical user interface, where the marking of the answers is unambiguous. The Java functions of our program are divided into two Java classes: a class handling graphical interface and a class manipulating knowledge base. This partition allows the construction of a more complex system: the linking of Java testing procedures and Learning Management System (LMS) (Borbásné Szabó, 2006).
5. Results To get feedback and input for refinement and improvement of the knowledge evaluation system we organized several test cycles in which graduating students participating in Business Informatics training program were involved. Further test details are discussed in Table 1. The knowledge evaluation test was available through a web-based learning management system (CooSpace). During the knowledge evaluation and testing we collected valuable source of information for further analysis. Total number of responses was 74504. Table 1 depicts the Higher Institutes-related information. We denoted their responsibility to the knowledge area which they were processing. Number of investigated knowledge areas includes all the sub knowledge areas for a certain knowledge area. Number of questions means the total of questions for a certain knowledge area. Higher Education Institute
Knowledge Area
Pécsi Tudományegyetem
Databases Application Development Architectures Information Technology System Analysis and Development I. System Analysis and Development II. OO programming and Java Mathematics, Linear algebra, Operational research Management Accounting Management and Organization
Eötvös Loránd Tudományegyetem Dunaújvárosi Főiskola Berzsenyi Dániel Főiskola Budapesti Műszaki és Gazdaságtudományi Egyetem Széchenyi István Egyetem Nyugat-Magyarországi Egyetem Eötvös Loránd Tudományegyetem, Kaposvári Egyetem, Szegedi Tudományegyetem Budapesti Gazdasági Főiskola Miskolci Egyetem Total
Number of Investigated Knowledge Areas
Number of questions
134
285
40 341
70 306
104
150
97
133
182
81
260
267
365
497
135
177
183 1938
156 2255
Table 1: Knowledge evaluation observations
Table 2 summarizes the most important observations, average response time (in sec.) per knowledge areas, average effectiveness, standard deviation of results and number of responses. Average effectiveness is defined as a ratio between good answer for a certain question and the total number of questions.
312
A. K˝o et al. / Ontology-Based Support of Knowledge Evaluation in Higher Education
Tests of Knowledge Areas
Average response time (sec)
Average effectiveness
Standard deviation of results
Number of responses
Databases Application Development Architectures Information Technology System Analysis and Development I. OO programming and Java Mathematics, Linear algebra, Operational research Management Accounting Management and Organization System Analysis and Development II. Average
23.284 26.085 27.000 24.545 27.000
55.20% 66.07% 77.30% 85.28% 76.33%
49.73% 47.35% 41.90% 35.43% 42.52%
8175 5730 2982 23618 1259
26.217 26.176
74.25% 57.36%
43.73% 49.46%
8854 14201
26.069 27.000 26.875
70.77% 61.96% 65.93%
45.49% 48.57% 46.37%
2145 1112 6428
26.025
69.05%
45.06%
74504
Table 2: Knowledge evaluation observations
The “best” performance (average effectiveness) was produced at Information Technology and at the same time standard deviation is the lowest and the number of responses the highest at this knowledge area.
Knowledge Areas
Standard Deviation of Effectiveness
System Analysis and Development II. Management and Organization Management Accounting Mathematics, Linear algebra, Operational research OO programming and Java System Analysis and Development I. Information Technology Architectures Application Development Databases 0%
10%
20%
30%
40%
50%
Standard Deviation(%)
Figure 3: Standard Deviation of Effectiveness per Knowledge Areas
Number of responses is the second largest one at Mathematics, Linear algebra, Operational research group, but the performance (average effectiveness) is the worst and standard deviation is the second highest one. This result is quite common; mathematics in higher education is one of the most difficult subjects for the students. Standard deviation of effectiveness per knowledge areas is demonstrated on Figure 3.
6. Conclusion Current phase of the research concentrated to the curricula of the Business Informatics program, as a test environment. This program is modelled and uploaded to the Educational Ontology. The primary goal of the adaptive knowledge testing and evaluation system, based on the Educational Ontology, is to explore missing knowledge areas. At the same
A. K˝o et al. / Ontology-Based Support of Knowledge Evaluation in Higher Education
313
time the accumulation of acquired competences and knowledge can also be determined. These way employers will be able to reinforce the position of student, trainee and employee when these persons want to enter in the labour market, want to look for another job or continue their study. We got valuable feedback during the test period, which is used for further improvement and refinement. In the future by improving the ontology and extending its content a common understanding of levels of competences based on learning outcomes can be established and this way educational system can compare their positions. Qualifications of “higher” level embrace the competences of “lower” levels. This suggests that there is a hierarchy between competences. Improving the model and applying further relations can also model this hierarchy in the Educational Ontology.
7. References [1] ACM, AIS and IEEE-CS (2005) “The Overview Report Covering Undergraduate Programs [online] http://www.computer.org/portal/cms_docs_ieeecs/ieeecs/education/cc2001/CC2005-March06Final.pdf [2] Berlin
Communiqué
(2003)
“Realising
the
European
Higher
Education
Area”
[online],
http://www.bologna-berlin2003.de/pdf/Communique1.pdf [3] Bologna Declaration (1999) “The Bologna Declaration of 19 June 1999” [online], http://www.bolognaberlin2003.de/pdf/bologna_declaration.pdf [4] Borbásné Szabó (2006) “Educational Ontology for Transparency and Student Mobility between Universities” in Proceedings of ITI 2006 [5] Corcho, O., Gómez-Pérez, A. (2000) “Evaluating knowledge representation and reasoning capabilities of ontology specification languages” in: Proceedings of the ECAI 2000 Workshop on Applications of Ontologies and Problem-Solving Methods, Berlin [6] Linacre, J. M. (2000): “Computer-adaptive testing: A methodology whose time has come”, in Chae, S. Kang, U. – Jeon, E. – Linacre, J. M. (eds.): Development of Computerized Middle School Achievement Tests, MESA Research Memorandum No. 69., Komesa Press, Seoul, South Korea [7] Gómez-Pérez, A., Corcho, O (2002) “Ontology Languages for the Semantic Web” IEEE Intelligent Systems, Vol. 17, No. 1, pp. 54-60. [8] Sorbonne Joint Declaration (1998) “Joint declaration on harmonisation of the architecture of the European higher education system” [online], http://www.aic.lv/rec/Eng/new_d_en/bologna/sorbon.htm [9] Sure, Y.; Studer, R. (2003) “A Methodology for Ontology-based Knowledge Management” In Fensel, D.; van Harmelen, F.; Davies, J., (2003): Towards the Semantic Web - Ontology Driven Knowledge Management, West Sussex, England: John Wiley & Sons Ltd. [10] Thissen,D., Mislevy,R.J. (1990). “Testing Algorithms” In Wainer, H. Computerised Adaptive Testing, A Primer. Lawrence Erlbaum Associates, Publishers, New Jersey, pp. 103-135. [11] Vas, R. (2006) “Educational Ontology and Knowledge Testing”, 7th European Conference on Knowledge Management, Budapest
314
Information Modelling and Knowledge Bases XIX H. Jaakkola et al. (Eds.) IOS Press, 2008 © 2008 The authors and IOS Press. All rights reserved.
When Cultures Meet: Modelling Cross-Cultural Knowledge Spaces Anneli HEIMBÜRGER University of Jyväskylä Faculty of Information Technology Information Technology Research Institute P.O. Box 35 (Agora) FIN-40014 University of Jyväskylä, Finland [email protected] Abstract. Cross-cultural research projects are becoming a norm in our global world. More and more projects are being executed using teams from eastern and western cultures. Cultural competence might help project managers to achieve project goals and avoid potential risks in cross-cultural project environments and would also support them to promote creativity and motivation through flexible leadership. In our paper we introduce an idea for constructing an information system, a cross-cultural knowledge space, which could support cross-cultural communication, collaborative learning experiences and time-based project management functions. The case cultures in our project are Finnish and Japanese. The system can be used both in virtual and in physical spaces for example to clarify cultural business etiquette. The core of our system design will be based on cross-cultural ontology, and the system implementation on XML technologies. Our approach is a practical, step-by-step example of constructive research. In our paper we shortly describe Hofstede’s dimensions for assessing cultures as one example of a larger framework for our study. We also discuss the concept of time in cultural context
1. Introduction The Internet and ubiquitous technology have opened up new possibilities for us to promote research and development projects as well as our business activities to new geographical locations and cultures. It is almost as easy to work with people remotely as it is to work face-to-face. Cross-cultural communication is more and more the new norm for our collaborative operations. Increasingly, businessmen, project managers, researchers and other professionals are becoming involved in international negotiations and meetings. The meetings can for example be international business meetings or international research project meetings. In addition to meeting agenda, participants also share culturally integrated space. Sometimes it can be difficult to understand culture dependent behavior of other parties during a meeting. By understanding some of the main cultural dimensions and by adjusting to cultural differences, people can face the challenge and become better negotiators and project managers on behalf of their companies and research organizations. The objective of our research project is to design and implement an information system – a cross-cultural knowledge space – that provides cultural assistant for people attending in cross-cultural meetings or for people working in cross-cultural projects [8].
A. Heimbürger / When Cultures Meet: Modelling Cross-Cultural Knowledge Spaces
315
The system can be used personally or collaboratively both in virtual spaces and in physical spaces. The contribution of the paper is to: x introduce a cultural ontology based approach to construct an information system that could promote communication and mutual understanding in cross-cultural collaborative research project environments, especially between eastern and western cultures x describe Hofstede‘s framework for cultural dimensions which is based on questionnaire study in 74 countries and on statistical analysis of the survey data x discuss the concept of time in cultural context as an essential issue of timebased project management functions. The term "culture" is used in our paper as it is defined in [11]: “Culture is a collective phenomenon, because it is shared with people who live or lived within the same social environment, which is where it was learned. Culture consists of the unwritten rules of the social game. It is the collective programming of the mind that distinguishes the member of one group or category of people from others“. The concept "cross-cultural" is used in the paper to describe comparative knowledge and studies of a limited number of cultures. For example, when examining negotiation manners or attitudes towards time in Finland and in Japan than that is a cross-cultural study. The concept "knowledge space" in cross-cultural context is used to describe personal and collaborative information systems both in virtual worlds on the fixed or ubiquitous Web and in physical worlds like in meeting rooms. The remainder of the paper is organized as follows. In Section 2, we describe a framework for assessing cultures with five cultural dimensions. In Section 3 we discuss the concept of time in cultural context. In Section 4, we introduce an idea for constructing an information system that supports cross-cultural communication in virtual and/or in physical space. The system is based on cultural ontology. We also present technological tools for and their roles in the implementation. Section 5 is reserved for conclusions and issues for further steps.
2. A Framework for Cultural Dimensions All of us, who are working, for example in international research projects, are involved – in addition to the subject of the project itself – in another kind of development process. Cultural competence [15] is a developmental process that evolves step-by-step over an extended period. Both individuals and organizations are at various levels of awareness, knowledge and skills on the cultural competence continuum. Cultural competence is about respecting cultural differences and similarities. There exist several studies for assessing cultures [11, 15]. These studies consider relations between people, motivational orientation, orientation towards risks, definition of self and others, attitudes to time, and attitudes to environments. Hofstede’s framework for assessing cultures is one of the widely used frameworks [10, 11]. Hofstede’s approach proposes a set of cultural dimensions along which dominant value systems can be ordered. These value systems affect human thinking, feeling, and acting, and the behavior of organizations and institutions in predictable ways. The framework consists of five dimensions: individualism/collectivism, power distance, masculinity/femininity, uncertainty avoidance and long-term orientation/short-term orientation (Table 1). All dimensions are generalizations and individuals may vary from their society’s descriptors. Hofstede’s metrics provides on interesting, larger framework for our study. In addition to this larger framework there are several culture dependent characteristics which
316
A. Heimbürger / When Cultures Meet: Modelling Cross-Cultural Knowledge Spaces
persons can face in their everyday working life. One example is communication style which can be indirect, paraverbal and/or nonverbal [18]. Nor should the role of business domain and organization specific cultures be underestimated. Awareness of cultural dimensions together with culture-specific characteristics could help people to develop their cultural competence. Table 1. Summary of cultural dimensions according to Hofstede’s study Dimension Individualism/ Collectivism
Power distance
Masculinity/Femininity
Uncertainty avoidance
Long-term/short-term orientation
Description of the dimension Individualism/Collectivism describes the extent to which a society emphasizes the individual or the group. Individualistic societies encourage their members to be independent and look out for themselves. Collectivistic societies emphasize the group’s responsibility for each individual. Power distance describes the extent to which a society accepts that power is distributed unequally. When power distance is high, individuals prefer little consultation between superiors and subordinates. When power distance is low, individuals prefer consultative styles of leadership. Masculinity/Femininity refers to the values more likely to be held in a society. Masculine societies are characterized by an emphasis on money and things. Feminine cultures are characterized by concerns for relationships, nurturing, and quality of life. Uncertainty avoidance refers to the extent that individuals in a culture are comfortable (or uncomfortable) with unstructured situations. Societies with high uncertainty avoidance prefer stability, structure, and precise managerial direction. In low uncertainty avoidance societies are comfortable with ambiguity, unstructured situations, and broad managerial guidance. Long-term/short-term orientation refers to the extent to which a culture programs its members to accept delayed gratification of their material, social, and emotional needs. Business people in long-term oriented cultures are accustomed to working toward building strong positions in their markets and do not expect immediate results. In short-term oriented cultures the “bottom line” (the results of the past month, quarter, or year) is a major concern. Control systems are focused on it and managers are constantly judged by it.
The scores of cultural dimensions in different countries according to Hofstede’s research are given in [12]. The survey is extensively described in [10]. The figures should not be taken literally. However they provide interesting information because they show differences in answers between groups of respondents. 3. Time in Cultural Context Time is seen in a different way by eastern and western cultures and even within these groupings temporal culture differs from country to country. Also temporal identities of different organizations and teams in organizations may vary. In cultural context, there exist two general time models: linear and cyclic [15]. In linear time model (Figure 1a) past time is over, present time can be seized and parceled and make it work for the immediate future. One task is carried out at time. For example, Scandinavian people are essentially linearactive, time-dominated and monochronic. They prefer to do one thing at a time, concentrate on it and do it within a scheduled timetable. Southern Europeans are more multi-active and polychronic. Monochronic cultures differ from polychronic cultures in that the former encourage a highly structured, time-ordered approach to life and the latter a more flexible, indirect approach, based more upon personal relationships than scheduled commitments. In many Asian countries time has traditionally been considered as cyclic. For example, the Japanese traditional temporal culture can be presented by the Makimono
A. Heimbürger / When Cultures Meet: Modelling Cross-Cultural Knowledge Spaces
317
model of time (Figure 1b) [7]. In Makimono time, the future flows into the present, just as the past does. The present is a period that links the region of the past with the world of the future. Nowadays linear time model has also been integrated into Japanese society.
Past
Past
Present
Future
Present Future
(a)
(b)
Figure 1. Linear time model and cyclic time model according to Makimono time pattern. Makimono takes its name from the makimono, a picture story or writing mounted on paper and usually rolled into a scroll.
Cross-cultural projects involve teams and individuals with different concepts of time, and therefore a completely different frame of mind as far as planning, scheduling, punctuality and project deadlines are concerned. Tensions may arise quite easily. In such a case, it is the task of the project manager, on the basis of his/her cultural competence, to make sure such different attitudes do not become the source of major misunderstandings. Time contexts in project management are discussed more detailed in [9]. 4. Towards a Cross-Cultural Ontology The development process of cultural competence of project managers and project teams could be supported by culture-sensitive information systems both in virtual and in physical environments. In our system we first construct a cross-cultural ontology which will be the basis of the system. An ontology is the result of an attempt to formulate an exhaustive and rigorous conceptual schema about a certain domain. The domain does not have to be the complete knowledge of that topic, but an interesting part of it decided by the creator of the ontology. In our approach, the cultural dimensions discussed in the Section 2 can be grouped into three categories: relations between people, motivational orientation and attitudes towards time. These categories can be complemented with an application category which includes cross-cultural applications such as project negotiations [2, 16] and time-based project management. The four categories form the first hierarchy level of a cross-cultural ontology (Figure 2): (individualism, collectivism) Relations between people (masculinity, femininity, uncertainty avoidance, power distance) Motivational orientation (long-term orientation, short-term orientation, linear time, cyclic time) Attitudes towards time (project negotiations, project management) Applications
318
A. Heimbürger / When Cultures Meet: Modelling Cross-Cultural Knowledge Spaces
Relations_between_People
IND COL MAS
Motivational_Orientation
FEM PD
Cross_Cultural_Thing
UA Attitudes_Towards_Time
STO LTO
Applications
LTM CTM
APPL1_Temporality_of_Project_Management
APPL2_Project_Negotiations
Language Initial_Contact Relationship_Before Orientation_of_Time Hierachy_Status Maintaining_Harmony Concern_with_Face Formality_and_Rituals Communication_Style Presentation Decision_Making_Process Role_of_Contract Dress_Code Meeting_and_Greeting Forms_of_Address Gift_Giving_and_Receiving Wining_and_Dining Maintaining_Relationship
Instant Interval Duration_Description Date_Time_Description Temporal_Unit Day_of_Week
Figure 2. Cross-cultural ontology can be associated to cultural knowledge that is represented in XML documents. The case cultures in our project are Finnish and Japanese. For example in application concerning project negotiations there can be a collection of XML documents describing a Japanese Negotiator and a Finnish Negotiator. The following abbreviations are used in the figure: Individualism = IND, Collectivism = COL, Masculinity = MAS, Feminity = FEM, Power Distance = PD, Uncertainty Avoidance = UA, LongTerm Orientation = LTO, Short-Term Orientation = STO, LTM = Linear Time Model and CTM = Cyclic Time Model.
A. Heimbürger / When Cultures Meet: Modelling Cross-Cultural Knowledge Spaces
319
The idea of the system design is that it can be used both in virtual and in physical environments i.e. (a) as a personal assistant via mobile devices, (b) as a collaborative assistant in meeting rooms and (c) as a personal/collaborative assistant in a virtual project space. Basically, the same idea can be applied for example to cross-cultural business meetings, education, tourism, medical and social services. The functions of essential technologies for implementation are shortly summarized in Table 2. Table 2. Essential technologies for constructing cross-cultural knowledge spaces
Technology
Functions
Ubiquitous and Context Aware Computing
Ubiquitous computing refers to a new computing paradigm that focuses on offering user-friendly information services - anywhere and anytime [20]. The core function is to support users by means of a cross-cultural knowledge space that is aware of their presence and cultural context. Context is any information that can be used to characterize the situation of an entity. An entity can be a person, a place, a space, time or an object that is considered relevant to the interaction between a user and an application. The system is contextware if it uses context to provide relevant information and/or services to the user where relevancy depends on the user’s task or situation [3, 4]. Examples of contexts in cross-cultural environments are nationality (static situation), location and time (dynamic situation), preferences (static intension) and joint project activities (dynamic intension). OWL is a markup language for publishing and sharing data using ontologies on the Internet [24]. OWL is used to formulate a conceptual schema for cultural entities. OWL-Time presents an ontology of temporal concepts [23]. The ontology provides a vocabulary for expressing facts about topological relations among instants and intervals, together with information about durations, date and time. OWL-Time is used as a basis time ontology in cross-cultural time-based project management applications. In Semantic Web languages, such as RDF and OWL, a property is a binary relation. It is used to link two individuals or an individual and a value. However, in some cases, the natural and convenient way to represent certain concepts is to use relations to link an individual to more than just one individual or value. These relations are called nary relations [22, 25]. In our ontology we need for example to represent multicultural properties of an object. Kansei is an ability that allows humans to solve problems and process information in a personal way. In every action performed by a human being, traces of his/her Kansei can be noticed, as well as his/her way of thinking and solving problems. Kansei is related both to problem solving tasks and to information analysis and synthesis. [1, 5, 6]. In the design of information systems, the concept of Kansei is related to data definition and data retrieval [13, 14]. In our research we study how cultural dependent semantic attributes could be added to Kansei information processing and thus how culturesensitive information retrieval can be supported. An important function in cross-cultural virtual spaces is to express emotions. XML based language for emotions could be one approach for expressing emotional functions. In topic maps [21, 27], three constructs are provided for describing the subjects represented by the topics: topic names, occurrences, and associations. Topic can be typed. Occurrences relate topics to the information they are relevant to. Table 2 continues …
Web Ontology Language (OWL)
OWL-Time
N-ary relations
Kansei Information Processing
XML for Emotions
XML Topic Maps (XTM)
320
A. Heimbürger / When Cultures Meet: Modelling Cross-Cultural Knowledge Spaces
XML Topic Maps (XTM)
Virtual spaces Radio Frequency Identification
Table 2 continues … Occurrences use URI addresses to identify the information resources, such as XML documents, being connected to the topic. Associations represent relationships between topic, and like occurrences they can be typed. The relationships in traditional classification schemes have little semantic content, whereas in topic maps one generally tries to make the typing of associations as specific as possible. In our project, Topic Map approach is used to design the user interface. An intelligent virtual space platform will be selected for the project RFID is an automatic identification method, relying on storing and remotely retrieving data using devices called RFID tags or transponders [17, 19]. A typical RFID solution consists of a data gatherer, RFID reader, and a data carrier (RFID tag) that is attached to an item (mobile device) or location (meeting room). Meeting rooms in organizations can be regarded as being context-sensitive areas and appropriately equipped by means of RFID technology for cross-cultural applications.
5. Conclusions and Next Steps In our paper we introduced an idea for constructing an information system that could support cross-cultural communication and project management functions in collaborative virtual or physical spaces. Our system will be designed by means of a cross-cultural ontology and will be based on XML and agent [26] technologies. The plan for our next steps is: Phase 1: System design, demonstrator implementation, testing in laboratory Phase 2: Qualitative evaluation in selected test sites Phase 3: Focus our design towards cross-cultural agent (CCA) applications. Cultural competence can be regarded as a set of congruent functions such as behaviors, attitudes, and policies that work in an information system and/or among professionals and enable the system and the professionals to work effectively in cross– cultural situations. From operational point of view, cultural competence is the integration and transformation of knowledge about cultures, groups of people and individuals into specific standards, policies, practices, and attitudes. These are used in appropriate cultural settings to increase the quality and context-sensitivity of information systems. Projects that use effective cross-cultural human-computer systems could provide a source of learning experiences and innovative thinking to enhance the competitive position of the participating organizations. Acknowledgements We express our deep thanks to the Satakunta High Technology Foundation and to the Scandinavia-Japan Sasakawa Foundation for funding the preliminary phase of our research project. References [1] [2]
Camurri, A., Trocca, T. and Volpe, G. 2002. Interactive Systems Design: A KANSEI-based Approach, Proc. NIME2002, Dublin, Ireland, May 2002. De Mente, B. 2001. Etiquette. Guide to Japan. Singapore: Tuttle Publishing. 132 p.
A. Heimbürger / When Cultures Meet: Modelling Cross-Cultural Knowledge Spaces [3] [4] [5] [6] [7] [8] [9]
[10] [11] [12] [13]
[14] [15] [16] [17] [18] [19] [20] [21] [22] [23] [24] [25] [26] [27]
321
Dey, A., Kokinov, B., Leake, D. and Turner, R. (eds.) 2005. Modeling and Using Context. LNAI 3554. Berlin: Springer-Verlag. 572 p. Dey, A. K. 2001. Understanding and using context. Personal and Ubiquitous Computing, Vol. 5, No. 1, p. 4 – 7. Harada, A. 1997. The Framework of Kansei Engineering. Report of Modelling the Evaluation Structure of Kansei. University of Tsukuba. Pp. 49 – 55. Hashimoto, S. 1997. KANSEI as the Third Target of Information Processing and Related Topics in Japan. In: Camurri, A. (Ed.) Proceedings of the International Workshop on KANSEI: The Technology of Emotion, Italian Computer Music Association and DIST-University of Genova, pp. 101 – 104. Hay, M. and Usunier, J-C. 1993. Time and Strategic Action. A Cross-Cultural View. Time & Society, Vol. 2, No. 3. Pp. 313 – 333. Heimbürger, A. 2006. Cross Cultural Interactive Spaces. The Position Paper for the W3C Ubiquitous Web Workshop, 9 - 10 March, 2006, Keio University, MITA Campus. 5 p. Heimbürger, A. et al. 2006. Time Contexts in Document-Driven Projects on the Web: From TimeSensitive Links towards an Ontology of Time. In: Kiyoki, Y., Kangassalo, H., Jaakkola, H. and Duzi, M.. (eds.). Proceedings of the 16th European-Japanese Conference on Information Modelling and Knowledge Bases (EJC 2006), May 29 – June 2, 2006, Trojanovice, Czech Republic. 158 – 175 p. Hofstede, G. 2001. Culture's Consequences, Comparing Values, Behaviors, Institutions, and Organizations Across Nations. Thousand Oaks, CA: Sage Publications. 596 p. Hofstede, G. and Hofstede, G. J. 2004. Cultures and Organizations: Software of the Mind: Intercultural Cooperation and Its Importance for Survival. New York: McGraw-Hill. 300 p. Hofstede, G. 2003. Geert Hofstede™ Cultural Dimensions (referred 13th Aug. 2007)
322
Information Modelling and Knowledge Bases XIX H. Jaakkola et al. (Eds.) IOS Press, 2008 © 2008 The authors and IOS Press. All rights reserved.
Process Dimension of Concepts Vaclav REPA Department of Information Technologies, University of Economics, W.Churchill sqr. 4,130 67Prague 3, Czech Republic
Abstract. This article discusses the problem of process description in conceptual models. We argue for the idea that also the conceptual elements of the reality should be viewable dynamically – as a process. The article outlines the sense and the methodical way of describing two basic types of the Real World processes: business processes and object life cycles. In more detail the article analyzes basic kinds of the models coherency introducing two main criteria of completeness and correctness of models together with the concept of the structural coherency of models. It also discusses possible ways of describing dynamic aspects of the Real World and outlines some general conclusions.
Introduction There are several approaches to the conceptual modeling in the area of object-oriented methods. Each of them reduces the Object Model (represented by the Class Diagram) to the model of objects and relationships between them, represented by their attributes, but not by their methods. This reduction is present also in Roni Weisman´s approach [10] even if he regards besides “Entities” also “Control Object”. Just the fact of distinguishing between “static” and “dynamics ensuring” objects is the best demonstration of such a reduction. The common understanding of the term “conceptual” thus tends to the synonym for “static”. However such an approach contrasts with the basic principle, and the main contribution, of the object-oriented paradigm – unity of data and operations. This principle evokes the idea that it is necessary to model not only static aspects of the Real World but also its dynamics. The existence of the object as the collection of data (attributes) and functions (methods) is to be the right reason for data processing operations control - strictly speaking: the object life cycle. Figure 4 illustrates the object life cycle as a complement to the Class Diagram. It is visible that all methods of the conceptual object should be ordered into one algorithm which describes the place of each method in the overall process of the object’s life. This placement of the method defines the conceptual meaning of it.
1. Types of processes in the Real World The problem of dynamics in the Real World model is usually closely connected with the phenomenon of business processes. Hence the model of business processes is usually regarded as the only significant description of the Real World dynamics. Consequently the conceptual model is usually regarded as just a static description of the Real World. Another
V. Repa / Process Dimension of Concepts
323
extreme opinion regards the Class Diagram as the sufficient tool for business process description and reduces the natural need for describing the process dynamics to the description of the business processes global attributes, and relationships among them (the standard UML profile for BP modeling [8], for example). Experience shows that above stated opinions inadmissibly reduce the substance of the problem of the Real World dynamics and finally lead to the incorrect conclusions. Figure 1 describes two main dimensions of the Real World model: • the structure of the Real World (the view on the Real World as a set of objects and their relationships), • the behavior of the Real World (the view on the Real World as a set mutually connected business processes). Real World Behavior (Business Processes Model) Process Diagram Events and Actions
Events / Methods
Real World Structure (Object Model)
States / Attributes State Chart Data Structures
Class Diagram Attributes and Methods
Figure 1 Two Dimensions of the Real World Model At the figure it is clearly visible that the concept of “behavior” cannot be regarded as a synonym to the “dynamics”. Both dimensions have common intersection. Even inside the Real World structure it is thus necessary to regard some dynamics – the intersection contains, besides the static object aspects as attributes and data structures, also typical dynamic aspects as events, methods, and object states. Thus the description of dynamics is not just the matter of the behavioral model. It is the matter of the conceptual model as well. Obviously there are two types of dynamics in the Real World: • dynamics of the Real World objects, represented by their life cycles, • behavior in the Real World, represented by business processes. The Real World objects cannot be regarded as business processes because: • objects are not behaving – their life cycles are rather the description of business rules in a process manner, • the process of the object life has no goal (except the “death” of the object), nor product, it is rather the expression of the objective necessity, • although we describe the process of the objects life-cycles, that description still remains the structural one – whole context is described statically (structurally), it is subordinated to the Real World structure, • objects are typically taking different roles in different processes giving them the context (Real World rules). From the opposite viewpoint the business process is quite a different kind of process than the life-cycle of the object because: • business process has the goal, and the product, as typical expression of the human will • business process typically combines different objects giving them the specific meaning (roles of actors, products, etc.). For detailed discussion of the main differences between object life cycles and business processes see [3], [4], [5], and [6].
324
V. Repa / Process Dimension of Concepts
The above mentioned facts support the need for modeling the dynamics of the conceptual objects as something different from the behavior of the Real World, which is traditionally represented by business processes. Although in both cases we regard the modeling of processes, at the same time we have to take into the account the fact that modeling of the conceptual objects dynamics has its specific logic, different from the logic of the modeling business processes. This logic primarily reflects the specific nature of the object life cycles, discussed above.
2. Modeling Object Life Cycles For the purpose of object life cycles description the most suitable tool from the Unified Modeling Language (UML) is the State Chart [8], [9]. State Chart is not primarily intended for the description of life cycle, its roots are in the area of state machines theory, and it is closely connected with the concept of so called “real-time processing”. However the concept of the state machine in general is not substantially reducible just to the area of real-time processing. Also in the area of data processing there is the need for recognizing states and transitions among them. The best proof of this idea is the concept of the object life cycle itself – once we think about the objects generally (i.e. in terms of their classes), than we have to strongly distinguish between the class and its instance. In the case of the object life this requires to determine those points in the life of all objects of the same class, which we will be able to identify, and which it is necessary to identify in order to describe the synchronization of the object life with life cycles of other objects. Such points of the object life are its states. So each object instance lives its own life while the lives of all instances of the same class are described by the common life cycle. As it is visible at the Figure 4 State Chart describes possible (allowed) states of the object together with the possible transitions among them. Each transition is described with two attributes: • reason for the transition (upper part of the transition description), • method of the transition realization (lower part of the transition description). Each described life cycle has to correspond to the particular object class in the Class Diagram. Such a way the State Chart specifies the general mechanism of the life of all possible instances of the given class. Described states and transitions among them consequently correspond to the attributes and methods of the class. Life cycle states represent in fact the specific attribute of the class (this attribute is not present in the class description but it exists by the definition – it is necessary to distinguish among particular states / values of this “hidden” attribute). Each transition between life cycle states then represents the use of the particular class method. While the method of the transition realization corresponds to the specific method of the given class, reason for the transition corresponds to the specific event (external influence) which causes the transition. The concept of events, as a common concept existing in both main points of view on the Real World dynamics, allows linking of the description of object life cycles with the description of business processes (see below).
3. Modeling business processes Process Diagram Technique aims to offer the set of concepts, symbols and rules, using which the modeler is able to describe all substantial characteristics of the real world behavior in as simple way as possible. The key concepts of the technique, together with their relationships, are specified in the process meta-model (see OpenSoul project [2]). In the process model events, states, and activities of the process play the crucial role. They
325
V. Repa / Process Dimension of Concepts
serve as a "meeting point" of the two main points of view existing in the real world modeling: • object model (static - structural model of the real world) • process model (dynamic - behavioral model of the real world). Therefore we regard stimuli and activities as so important aspects of the process. They enable interconnection between object and process models as well as they enable the expression of appropriate integrity rules.
Order
Stock
Goods dispatched
Order receiving Order entry
Customer payment done
Order fulfilment Order accepted
Order clearance Goods delivered
Order rejection report Invoice Order deficiencies report
Delivery order
Order rejected
Order cleared
Figure 2: Example of Business Process Model (BPMN notation) Figure 2 illustrates the use of above stated technique. It shows how the process description emphasizes the most important aspects of the process: • events and their consequences – process activities and states (i.e. points of waiting for the event) on one hand, • inputs and outputs processed by the process including the main process product (i.e. the main reason for the process run).
4. Coherency of models Regarding the coherency of models let us introduce two basic criteria: • completeness of models • correctness of models Class Diagram correctness of the conceptual model completeness of the conceptual model correctness (completeness) of object relations
correctness (completeness) of object roles
correctness (completeness) of reasons
State Chart correctness of the Object Life Cycle
correctness (completeness) of actions
completeness of the Object Life Cycle
Business Process Diagram
correctness of the business process model completeness of the business process model
Figure 3: Criteria of Completeness and Correctness in Diagrams
326
V. Repa / Process Dimension of Concepts
As the Figure 3 illustrates completeness and correctness are mutually interconnected. On the level of particular diagrams each criterion has the specific meaning. But in the intersections of particular diagrams, even more in the intersections of all three diagrams, both criteria convene together. Exactly: correctness of the models has the form of completeness of the superior general concepts (relations, roles, actions, and reasons) in them. Specific kind of models coherency is the coherency of main types of structures, which occur in all viewpoints in several forms. I call this kind of models coherency structural coherency. The roots of the idea of structural coherency are in the work of Michael Jackson, in his method “JSP” [1]. For detailed explanation of the idea how to use Jackson’s ideas for this purpose see [7]. Basic rules for the structural consistency of objects in the conceptual model as follows: • Each association between two object classes must be reflected by the specific operation in each class life cycle. • The cardinality of the association must be reflected by corresponding type of structure in the life cycle of the opposite class: cardinality 1:n by the iteration of parts, cardinality 1:n by the single part of the structure. • The optionality of the association must be reflected by corresponding selection structure in the life cycle of the opposite class. • Each generalization of the class must be reflected by corresponding selection structure in its life cycle. • Each aggregation association between classes must be reflected by corresponding iteration structure in the life cycle of the aggregating class (container / composite class). Class Diagram
Goods
Order
Order OrderEntry CreateOrder Exemption Delete
Delivery() Exemption RemovefromtheAccountofStore
In accountof Store Exemption RemovefromtheAccountofStore
Exempted Exemption Delete
Created
ChangeAmount() OrderChange() OrderCancel()
Registered RegistrationinStore TakeintoaccountofStore
Order_accepted / CreateOrder
CreateOrder()
Created Registration Takeinto account
Order No.:....... Name:....... ................
Delivery / ChangeAmmount
DeleteOrder()
ChangeofGoods Change GoodsDelivery ChangeAmount
Orders
0..1
Is Ordered by
1..n
Goods Part No.:....... Name:....... ................ Create() Delivery() ChangeAmount()
Filled Delivery / ChangeAmmount Goods_delivered / SendInvoice
Fulfilled Customer_payment / DeleteOrder
Exemption() Delete()
Figure 4 Structural Coherency of Objects and their Life Cycles Figure 4 illustrates some examples of structural coherences in the conceptual model. Class diagram represents the static contextual view on reality, while the object life cycle describes the “internal dynamics” of the class. The internal dynamics of the class should be subordinated to the context (i.e. substantial relationships to the other classes), therefore
327
V. Repa / Process Dimension of Concepts
each class contains specific operation (method) for each association (it is obvious that some associations to other classes are missing in this example). The life cycle determines the placement of each particular operation in the overall life history of the object - the internal context of the operation. The internal context must be consistent with the external one, which follows from relationships described between classes in the Class Diagram (associations to other classes, generalizations etc.). Dashed arrows indicate basic consequences ofdescribed associations and their cardinalities in life cycles of both classes: • Optionality of the association (goods may not to be ordered at all) is reflected by the existence of the possibility that the whole sub-structure, representingordering of goods may be idle in the Goods life cycle. Also the fundamental conditionality of the delivery is the reflection of this fact. • Multiplicity of the association (one Order may contain several items) is reflected by the iteration of the structure “Filling” in the Order life history which expresses the fundamental fact that the order may be created, fulfilled by several supplies, or changed several times, separately for each ordered item. The knowledge of structural consequences helps the analyzer to improve the Real World models concerning their mutual consistency as well as their relative completeness (as the completeness is a main part of the problem of consistency). Figure 5 illustrates how the process model explains dependencies between objects and their life cycles giving them the superior sense. This explanation is based on the perception of object actions in terms of reasons for them – events and process states. Objects are playing roles of attendees or victims (subjects) of processes. For completeness it is necessary to regard the fact that one object typically occurs in more processes as well as one process typically combines the attendance of more objects. The orthogonality of those two points of view is also typical and substantial – it gives the sense to this coupling. Structure and behavior is the analogy of two basic dimensions of the real world – space and time. Process Diagram O rd e r
Goods dis patc hed
S to ck
Order receiving Order entry
Custom er paym ent done
Order fulfilm ent
Order clearance Goods delivered
Order ac cepted
O rd e r re j e cti o n re p o rt In vo i ce O rde r d e fi ci e n ci e s re p o rt
De l ive ry o rd e r
O rd e r re je cte d
Class Diagram
O rd e r cl e a re d
Order
STD - Goods
Order No.:....... Name:....... ................ CreateOrder() Delivery() ChangeAmount() OrderChange() OrderCancel() DeleteOrder() 0..1
Order Entry CreateOrder
?
Goods Delivery ChangeAmount
Goods
Goods Delivery ChangeAmount
Registration in Store Take into account of Store
Goods Delivered Fulfilment
Part No.:....... Name:....... ................
Exemption Remove from the Account of Store Registered
Filled
1..n
Exemption Delete Created
Registration Take into account
Accepted
Objednává
Create() Delivery() ChangeAmount() Exemption() Delete()
Order Entry CreateOrder
STD - Order
In account of Store Fulfiled
Customer Payment Done Delete
Exemption Remove from the Account of Store Exempted Unconceptual presumption: two times the same event
Exemption Delete
Iteration
Figure 5 Example of the coherency of models
Change of Goods Change Goods Delivery ChangeAmount
328
V. Repa / Process Dimension of Concepts
Even the specified consistency rules are working together in mutual coherency. This means that there should be regarded a number of additional second- and third-level consistency rules following from the combination of basic rules. For example: We suppose that each event specified in the Object Life Cycles is used in some Business Process(es) (the rule for correctness (completeness) of reasons), and in the same time we require each state transition in the object life cycle to be corresponding to some association to another object class (the rule for correctness (completeness) of object relations). From this combination of rules follows that we suppose that each event causes some business action (as it is defined in the business process model) and that it causes the state transition of some object (as it is defined in the object life cycle), and that it fulfills the link to some other object in the same time. In fact this means that each business process activity has logical consequences in mutual behavior of objects (and vice versa1).
5. Conclusions - possible ways of describing dynamics of the Real World Concluding from the previous chapters we can see that there are two main approaches to the description of the Real World dynamics: Business Process approach is characterized with the modeling Business Processes on one hand and Object Life Cycles on the other hand, taking care of their mutual consistency. In this approach Object Life Cycles are playing the role of process-manner description of “Business Rules” – process description of crucial restrictions given by business which are naturally static (in spite of the fact that they are described as processes (of object lives)). Two basic viewpoints of the modeled Real World (the intentional one - business process versus the static one – object life cycles) allow dramatical refinement of the set of rules defining the correctness (completeness) of models. On the other hand this approach is not open - all possible actions are described in the form of business processes, actors have no chance to behave out of these processes. It means that this approach always reduce the large-scale reality just on the subset defined by the models. This can cause serious restriction of the ability to change traditional rules which is still more important in our turbulent world. “Legislative approach” is characterized with the modeling the objects and their mutual relationships supposing them as the Real World agents with their own activity. We also should take care of the mutual consistency of objects and their life cycles. In this approach Life Cycles are playing the role of the description of basic Real World rules which have to be respected by any behaving object. Those objects from the model which represent the Actors are regarded as the Real World agents with their own activity. They are behaving actively and independently, respecting just the rules given them via their life cycles and mutual dependencies on other objects. Thus there is no need to model Business Processes – the Class Diagram together with models of life cycles just “delimitate the space” for objects behavior – basic “legislation”. Therefore I call this approach Legislative approach. This approach is more open and thus potentially closer to the reality than the Business Process approach. In the real world actors are usually acting according to the given rules by their own activity. All possible ways of acting are in the account. On the other hand the missing presumption of intentional sets of Real World actions, in the form of described business processes, reduces the possibility to formulate integrity rules which could reduce the set of possible agents actions to the set of just correct ones (in the sense of described business processes). For instance see the rules for completeness and correctness of object 1
In fact, here we deal with famous “chicken-and-egg dilemma“, deciding whether mutual behavior of objects is the consequence of business process activities or whether business process activities are rather given by the actor’s behavior. This problem is connected with two basic ways of describing dynamics of the Real World which are discussed below.
329
V. Repa / Process Dimension of Concepts
roles, and for reasons, which are completely useless when the business process description is missing.
high
vague management
wise management
Legislative Approach (active objects, agents, business processes not defined)
maturity in knowledge management
Business Process Approach (definition of business processes, passive objects)
low
directive management
poor management
low
preciseness of processes description
high
Figure 6 Two main approaches to the description of the Real World dynamics Figure 6 shows the context of both above discussed approaches. The “Legislative approach” is suitable when the level of maturity in knowledge management is high. As it represents the vague management style, it strongly depends on the self-organizing ability of the system. On the other hand the “Business Process approach” is suitable just when the ability for the Real World rules description is high – i.e. in the terms of relatively static and well structured environment. It seems that the right way lies in the combination of both approaches which allows overcoming of their limitations. Particularly it means the need to find the role of active objects in business processes at first. This is the main idea of further development of the methodology.
References [1] Jackson M.A.: JSP in Perspective; in Software Pioneers: Contributions to Software Engineering; Manfred Broy, Ernst Denert eds; Springer, 2002. [2] http://opensoul.panrepa.org. [3] Repa V. “Object Life Cycle Modeling in the Client-Server Applications Development Using Structured Methodology“, Proceedings of the ISD 96 International Conference, Sopot, 1996. [4] Repa V. “Information systems development methodology – the BPR challenge“, Proceedings of the ISD99 International Conference, Kluwer Academics, Boise, ID, 1999. [5] Repa V. “Process Diagram Technique for Business Processes Modeling“, Proceedings of the ISD2000 International Conference, Kluwer Academics, Kristiansand, Norway, 1999. [6] Repa V. “Business System Modeling Specification“, Proceedings of the CCCT2003 International Conference, IIIS, Orlando, FL, 2003. [7] Repa, V: Modeling Dynamics in Conceptual Models, ISD 2006 Conference Proceedings, Budapest, New York : Springer, 2007. [8] UML “ OMG Unified Modeling Language Specification, v. 1.5. document ad/03-03-01, Object Management Group, March 2003.” [9] UML Superstructure Specification, v2.0 document 05-07-04, Object Management Group, 2004.” [10] Weisman, R.: Introduction to UML Based SW Development Process: www.softera.com, 1999.
330
Information Modelling and Knowledge Bases XIX H. Jaakkola et al. (Eds.) IOS Press, 2008 © 2008 The authors and IOS Press. All rights reserved.
E-Government: on the Way Towards Frameworks for Application Engineering Marie-No¨elle TERRASSE1,4,6 , Marinette SAVONNET1,6 , Eric LECLERCQ1,6 , George BECKER2,6 , Thierry GRISON1 , Laurence FAVIER4,5 , and Carlo DAFFARA3,6 (1) LE2I (UMR CNRS 5158), University of Burgundy, E-mail: fi[email protected] (2) E-mail: [email protected] (3) Conecta Telematica, Italy, E-mail: [email protected] (4) Pr@tsic, Maison des Sciences de l’Homme, University of Burgundy (5) Centre Georges Chevrier (UMR CNRS 5605), University of Burgundy, E-mail: [email protected] - (6) OpenTTT European project Abstract. In this article we present high-level architectures for e-Government applications. These architectures depend on a country’s strategy for e-Government integration and they give rise to two major issues. The first issue is how to guarantee semantical quality of information regardless of the chosen architecture. The second issue is how to facilitate sound transition of eGovernment applications from one architecture to another under evolutionary pressures of a country’s political strategy. In order to address these two issues we use Model-Driven Engineering which places metamodels, models and their transformations at the core of the engineering process. Overall semantical quality is thus guaranteed by metamodels while model transformations guarantee soundness under evolution. We propose two adjustments to OMG’s architectures for Model-Driven Engineering of highly-complex application domains. In OMG’s architectures, a metamodel describes an application domain (reusable information) while a model describes an application (contextual information). By introducing a reusable model for a family of applications, we can share pieces of model-level information.
1
Introduction
E-Government applications should be able to evolve incrementally since they belong to a relatively stable domain. Legacy information systems of public administrations operate in well-known domains. They generally rely on stable and recognized vocabularies and they are used in the context of unchanging business processes. Yet, the spreading of new technologies and the expectations of various actors (citizens, administrative project leaders, politicians) push towards development of innovative information systems. In fact, E-Governement implies several major changes in administration business processes: • a citizen-centered approach to e-Government which is based on availability of services dedicated to life and business events (e.g., birth, marriage, as well as setting up a company, paying taxes, participating in procurement activities) and delivered through various channels [7, 1]; • a separate management of services and their delivery through multi-channel portals;
M.-N. Terrasse et al. / E-Government: Towards Frameworks for Application Engineering
331
• an integration of administration services with respect to national strategies and citizens’ expectations, administrative staffs’ working habits, and international strategies. Even though administrative portals are the most visible part of current developments, E-Government’s integrated services are not restricted to front-office evolution. Back-office reorganization [5, 7] in turn makes it necessary to harmonize and to make consistent all levels of administration: local, national/federal, international (e.g., pan-european services) in order to enable interoperability of e-Government information systems. Such interoperability is rather difficult to set up, since e-Government applications generally exhibit strong heterogeneities, such as data heterogeneity (formats ranging from alpha-numeric data to cadastral map images, quality, semantics), actor heterogeneity (members of various administrations, end-users, or politicians which are given authorizations to access data and to use services), and heterogeneity of applications’ objectives. Furthermore, e-Government applications generally do not have precise non-functional specifications (such as those regarding security, confidentiality, and performance) even though many interoperability domain-dedicated frameworks that have been built recently enable e-signature, personal identification and exchange of data between administrations, (e.g., IETF, OASIS, WS-I, UNCEFACT, e-GIF, OOI, RGI [10, 11, 12, 13, 14, 15]). Such domaindedicated frameworks can be used together with technical specifications and architecture components [6, 9] that were offered to web-enabled application designers either by an international consortium, or by national structures (e.g., the Security Assertion Markup Language, the Identity Federation Framework that provides Single Sign On facilities, the UN/CEFACT Modeling Methodology, and the COSPA project [16, 17]). Most E-government applications can be described in terms of a loosely coupled integration of administrative information systems (from various administrations) to which up to three extra components can be added. The first component provides core business integration, i.e., it enables data and process consolidation. The second component is a portal for administrative staff members providing a unified access to information and services of each administration. The third component is a portal for end-users that offers an integrated view of all administrations regardless of their actual organization. Depending on the chosen components we define an architecture schema which we call an application profile. We propose eight different application profiles, and present them in Figure 1. The technical aspects of e-Government applications show that various basic components are necessary. For example, end-user portals should rely on an identity federation framework while administrative portals should encompass a language for expressing security and authorization rules. Similarly, the integrated core business should rely on knowledge and business process descriptions (e.g., ontologies, metamodels, models [2, 4]). We define four basic rules for selection of framework components. First, end-user portals are supposed to federate identities from various legacy e-Government applications. Second, administrative portals must encompass authorization descriptions and enforcement, as well as common vocabularies (formulated in terms of shared ontologies). Third, core business integration cannot be carried out without at least a common vocabulary (formulated in terms of shared ontologies). Fourth, each architecture must include security components. 2
An MDE perspective on E-Government applications
OMG’s metamodeling architectures strive to structure an application description into four levels: instance, model, metamodel, and meta-metamodel. The meta-metamodel level describes how the real world is seen, which high-level languages are used to describe the real world (e.g., description of a semantics of space and time). The metamodel level defines
332
M.-N. Terrasse et al. / E-Government: Towards Frameworks for Application Engineering
Figure 1: Different profiles of e-Government applications.
which language will be used for modeling of a specific application domain (e.g., a metamodel extended with constructs for spatio-temporal descriptions). The model level describes a given application (e.g., a model of a GIS for state and territorial border management). The instance level contains objects which belong to such a GIS (e.g., the French-German border after World War I, the border between the Brooklyn and Staten Island boroughs in New-York in 1964). Metamodels that were originally introduced as languages for model description [3] turned into languages for application domain descriptions (Domain Specific Languages [8]). Reuse is the key concept for application domain descriptions. MDE expresses such reuse at the metamodel level. Yet, building a model from a metamodel in case of a complex application requires a huge amount of work. We desire to reuse part of the modeling work: we thus propose to describe a family of applications in terms of a reusable model. A definition of such a reusable model distinguishes abstraction separation between metamodels and models (Figure 2.a) from methodological separation between reuse and contextualization (Figure 2.b). Each specific application is then built as a specific instance of the reusable model. Figure 2.c presents an example metamodel and two reusable models for e-Government applications with two different types of confidentiality requirements. 3
Illustrative Example: Metamodeling a Data Protection Strategy
Enforcing data protection policies in order to satisfy legal and security requirements is a major issue for e-Government applications. In order to keep our example reasonably small, we limit ourselves to a simplified context. We use the following vocabulary. E-Government applications use resources which are mainly documents containing data. Data elaboration is limited to two categories: raw data (e.g., the grades obtained by a student) and aggregate data
M.-N. Terrasse et al. / E-Government: Towards Frameworks for Application Engineering
333
Figure 2: Metamodeling levels and reuse boundary
(e.g., yearly averages of student grades). Depending on the data they contain, resources are classified according to the level of data protection they require. Data protection falls under three categories: public, confidential, and private data. Public data can be read by everybody (e.g., a list of the diploma delivered by a university), access to confidential data is restricted to administrative staff members (e.g., the grades obtained by students), private data can only be read by specially authorized administrative staff members (e.g., medical record for a disabled student). In order to manage access to private data, authorizations are delivered either on an individual basis or statutorily. Statutory authorizations are delivered by an administration to its staff members. Individual authorizations are delivered under responsability of authorization granting authorities. Introductory model Let us consider the general case where the following rules apply: R1- Public resources cannot be associated with authorizations. R2- Resources containing only aggregate data cannot be private. R3- Confidential resources must be associated with authorizations. Figure 3 presents an introductory model of e-Government applications in terms of a UML class diagram. Resources and data are represented by classes. The class Resource is specialized into classes Private, Confidential, and Public. The class Data is specialized into classes Raw and Aggregate. Authorizations are represented by a class Authorization together with two specialized classes Individual and Statutory. Authorization granting authorities are represented by a class Authority. An association, called reading, links resources with authorizations. An association, called granting, links individual authorizations with granting authorities. In order to guarantee modeling accuracy, it is necessary to make sure that rules R1 to R3 are expressed in the model. Rule R1 can be expressed as a specialization of the reading association. This specialization links Confidential with Authorization and has multiplicity set to 1..* at
334
M.-N. Terrasse et al. / E-Government: Towards Frameworks for Application Engineering
Figure 3: Data protection policy: introductory model (class diagram)
the Authorization end. Rules R2 and R3 must be expressed in the form of OCL constraints (e.g, as invariants of the class Public and Resource, respectively). These two rules are given in Figure 3.b. Metamodel In order to express domain-related knowledge at the metamodel level, we define three major concepts within the application domain, namely data, resources, and authorizations together with their relations. We then define five stereotypes: a stereotype D for modeling data, a stereotype R for resource modeling and expressing rules R1 and R2, a stereotype A for modeling authorizations, a stereotype RD for modeling reading, a stereotype RA for grant modeling. We choose to express rule R3 within reusable models since it pertains to the set of data elements associated with a resource. Figure 4 depicts the proposed metamodel: in part a) the proposed stereotypes are depicted in light gray, part b) presents the OCL expression of constraints c1 and c2 (which express rules R1 and R2, respectively). Reusable models By using the above metamodel, we define two example reusable models corresponding to two families of applications that share the same data protection policy. The corresponding reusable models are given in Figure 5. Our first example is a family of applications centered on protection of personal data. In such applications authorizations are statutory (invariant r4 of the class Authorization), and confidential resources must be associated with authorizations (invariant r5 of the class Resource). Rule R31 must be enforced (invariant r6 of the class Resource). Our second example is a family of applications centered on protection of strategic data. In such applications, confidential resources must be associated with individual authorizations (invariant r8 of the class Resource) and aggregate data are not necessarily 1 Rule
R3: Resources containing only aggregate data cannot be private.
M.-N. Terrasse et al. / E-Government: Towards Frameworks for Application Engineering
335
Figure 4: Data protection policy: domain-metamodel for e-governement applications with confidentiality requirements
public though they may not be private (invariant r7 of the class Resource). Rule R3 is subsumed by invariant r7. As stated in the above sections, reusable models allow reuse within families of e-Government applications. One of the major challenges is to appropriately define such families, which is particularly difficult for highly complex application domains. We propose to use each of the profiles of e-Government applications from Figure 1 as a family. A reusable model thus describes bases on which the integrated core business and portals of an application profile can be built. The benefit that we obtain is sound evolution of e-Government applications from one profile to another since models can be transformed under the control of the shared metamodel and of their source and target reusable models. Furthermore, satisfactory semantical quality of each model can be guaranteed (by means of a reference metamodel and reusable model). 4
Conclusion
In this paper, we have discussed the possible architectures of e-government applications. Two majors requirements apply to such architectures. First, these architectures have to enable sophisticated interoperability of legacy applications. Beyond integration of business processes and information, interoperation of e-government applications must enable: 1) end-users to obtain services from the integrated system for their life events. 2) administration staff members to access and control information and services. Second, sound evolution of an e-government application architecture must be guaranteed so that the architecture conforms to the country/region/administration strategy.
336
M.-N. Terrasse et al. / E-Government: Towards Frameworks for Application Engineering
Figure 5: Data protection policy: reusable models for e-Governement applications with confidentiality requirements
In order to satisfy the two above requirements, an MDE perspective on e-government has been introduced. Architectures of e-Government applications are thus described in terms of metamodels and models. In order to emphasize model-level reuse and semantical quality of the integrated information, we have described families of applications in terms of reusable models. Choosing characteristics of families of applications that are described by reusable models is a major issue for improvement of model-level reuse. Our on-going work is to validate the criteria defined in this paper, namely data protection strategies (a family of applications is defined by a given data protection strategy). In order to perform such a validation, various experiments will be carried out (including a one-year experiment with the French health care system. References
[1] E-government strategy 2002. Technical report, Executive Office of the President Office of Management and Budget Washington D.C. 20503, 2002. [2] C. Atkinson and T. Khne. Model-Driven Development: A Metamodeling Foundation. IEEE Software, 20(5), 2003. [3] Colin Atkinson. Meta-Modeling for Distributed Object Environments. In First International Workshop on Enterprise Distributed Object Computing, EDOC’97, pages 90–101. IEEE, October 1997. [4] G. Brunet, M. Chechik, S. Easterbrook, S. Nejati, N. Niu, and M. Sabetzadeh. A Manifesto for Model Merging. In Proceedings of the 1st ICSE Int. Workshop on Global Integrated Model Management, 2006. China. [5] Reorganization of Back-offices for Better Electronic Public Services – European Good Practices. Technical report, Danish Technological Institute & Institut f¨ur Informationsmanagement, Bremen, 2004. Volume 1: main 15.
M.-N. Terrasse et al. / E-Government: Towards Frameworks for Application Engineering
337
[6] B. Elvesæter, A. Hahn, A.J. Berre, and T. Neple. Towards an Interoperability Framework for Model-Driven Development of Software Systems. In Proceedings of the 1st Int. Conf. on Interoperability of Enterprise Software and Applications, Switzerland, 2005. [7] Interoperability for Pan-European e-Government Services. Technical report, European Union. COM(2006) 45 final, February 13, 2006. [8] M. Mernik, J. Heering, and A.M. Sloane. When and How to Develop Domain-Specific Languages. ACM Computing Surveys, 37(4), 2005. [9] M. Soden, H. Eichler, and J. Hoessler. Inside MDA: Mapping MOF 2.0 Models to Components. In Proceedings of the First European Workshop on Model Driven Architecture with Emphasis on Industrial Application, University of Twente, The Netherlands, 2004. Available at URL http: //modeldrivenarchitecture.esi.es/mda\ workshop.html. [10] The Internet Engineering Task Force (IETF). www.ietf.org. [11] OASIS. www.oasis-open.org. [12] Web Services-Interoperability Organisation (WS-I). www.ws-i.org. [13] e-GIF. www.govtalk.gov.uk. [14] OOI. http://standarder.oio.dk/English/. [15] General Interoperability Reference (RGI). ww.adele.gouv.fr/article.php3?id\ article= 1064. [16] United Nations Centre for Trade Facilitation and Electronic Business (UNCEFACT). www. ebxml.eu.org/default.htm. [17] COSPA Project. http://www.cospa-project.org.
338
Information Modelling and Knowledge Bases XIX H. Jaakkola et al. (Eds.) IOS Press, 2008 © 2008 The authors and IOS Press. All rights reserved.
A Personal Web Information/Knowledge Retrieval System Hao Han and Takehiro Tokuda {han, tokuda}@tt.cs.titech.ac.jp Department of Computer Science, Tokyo Institute of Technology Meguro, Tokyo 152-8552, Japan
Abstract. The Web is the richest source of information and knowledge. Unfortunately the current structure of Web pages makes it difficult for users to retrieve the information or knowledge in a systematic way. In this paper, using the tree approach, we propose a personal Web information/knowledge retrieval system for the extraction of structured parts from Web pages. First we get the layout pattern and paths of extraction parts of a typical Web page in target sites. Then we use the recorded layout pattern and paths to extract the structured parts from the rest of Web pages in target sites. We show the usefulness of our approach using the results of extracting structured parts of notable Web pages.
1 Introduction Today Web information/knowledge retrieval by a personal user is usually done through the use of Web browsers with the help of search engines’ index information. However, if we would like to get information and knowledge from a collection of necessary partial information of Web pages in one Web site or a number of Web sites, the use of Web browsers may not be a good solution. For example, in the BBC country profiles site, there exists a collection of 200 or more country/region information including most recent basic information such as capital city, population and leader’s name. If we would like to retrieve a collection of necessary basic information of 200 or more countries/regions, the use of Web browsers would be a time-consuming tedious task. Similar personal information/knowledge retrieval tasks may be retrieval of a collection of disease names and corresponding parts of human body from health/medicine sites or retrieval of a collection of company names and corresponding industrial area from finance sites. The purpose of this paper is to present a system for personal Web information/knowledge retrieval. Our system allows users to automatically collect necessary partial information or whole information from Web pages in one or a number of Web sites. What users have to specify may be the starting Web page, crawling area, target parts selection for one typical Web page, and the resulting table organization. The organization of the rest of this paper is as follows. In Section 2 we give an overview of our system. In Section 3 and 4 we respectively explain the method of partial information extraction and the method for reuse of layout patterns and paths. In Section 5 we show two kinds of resulting tables to present the extracted information. In Section 6 we give examples
H. Han and T. Tokuda / A Personal Web Information/Knowledge Retrieval System
339
of Web information/knowledge retrieval using our system. In Section 7 we discuss related work and evaluate our system. Finally we give our concluding remarks in Section 8.
2 Overview Our personal Web information/knowledge retrieval system has following steps for a user to retrieve a collection of partial information or whole information of Web pages in one Web site or a number of Web sites. Step 1. Specification of start points and crawling scopes in the target Web sites Step 2. Definition of names of target parts and their data types Step 3. Acquisition of layout pattern and selection of partial information or whole information from a typical Web page in the target Web sites Step 4. Reuse of the selection pattern of partial information in ordinary Web pages of the target Web sites Step 5. Definition of the resulting table format The outline of our system is shown in Fig. 1. We use XML tree approach for the extraction of partial information.
Figure 1: Outline of our system
3 Extraction of Partial Information 3.1 Definition of Part Names and Data Types We define a name for each target part and its data type for the extraction and presentation of partial information. The data type includes two kinds of information: property and structure. Property is text or object. Text is the character string in Web pages such as an article. Object is one instance of the photo, video and other multimedia files. Structure is single occurrence or continuous occurrence. A single occurrence is a node without similar sibling nodes such as the title of an article, and the continuous occurrence is a list of similar sibling nodes such as the paragraphs of an article. There are four kinds of data types: single text, continuous text, single object and continuous object. For example, for a news article Web page, the news title is a single text with name ”title”, the news contents are continuous text with name ”paragraph” and one photo is a single object with name ”photo”.
340
H. Han and T. Tokuda / A Personal Web Information/Knowledge Retrieval System
3.2 Layout Patterns 3.2.1 Definition of Layout Patterns The HTML document of a Web page can be represented by a tree structure as shown in Fig. 2. One node can be represented by its path from the root. We define a layout pattern of a Web page for dealing with Web page layout similarity. A tree structure of a HTML document can be divided into a number of subtrees. A Layout Pattern is a list of paths from the root of the entire tree to the roots of these subtrees. For example, a Web page can be divided into a number of main parts as shown in Fig. 3. The layout pattern of this Web page is the list of paths from the root of the entire tree to roots of all main parts.
Figure 2: A tree structure and paths
Figure 3: A Web page and its divided parts
3.2.2 Layout Pattern Acquisition In order to acquire the layout pattern, we need to parse the tree structure of a given HTML document of a Web page. We use JTidy [2] to transform HTML documents into XML documents because of potential syntax errors such as missing end-tags in HTML documents. We need to define the default number of division of the entire tree into main parts and also the default method of the division of the entire tree. If the number of divided main parts of a Web page is too large or too small, then the list of paths to roots of these main parts may be too sensitive or too insensitive. We analyzed many typical Web pages and found that the square root of the sum of leaf nodes of the tree structure seems appropriate for our extraction of partial information. Our default method of Web page division is as follows. 1. 2. 3. 4. 5. 6. 7.
Node ROOT = root node of tree; int SUM = sum of leaf nodes; int MAX = sqrt(SUM); List nodelist = new List(); nodelist.add(ROOT); List L = new List(); Node nextnode = null;
8. while (L.size() + nodelist.size() < MAX){ 9. L.add(nodelist.getAllNodes()); 10. L.remove(nextnode); 11. nextnode = the node in L with the maximum leaf nodes; 12. nodelist = nextnode.getChildNodes(); 13. }
The nodes in List L are root nodes of the divided subtrees. Usually the visible information is embeded between the node and , so we can consider the node as the root node of tree structure of HTML document. Therefore, the layout pattern is a list of paths from the node to nodes in List L. The path takes a form: body : 0/N1 : O1 /N2 : O2 /.../Nn−1 : On−1 /Nn : On , where, Nn is the node name of the n-th node, On is the order of the n-th node among the sibling nodes, and Nn−1 is the parent node of Nn .
H. Han and T. Tokuda / A Personal Web Information/Knowledge Retrieval System
341
Figure 4: Layout pattern acquisition
3.3 Parts Selection We select the target parts to reach the partial information. We collect the paths of the selected parts using the following process. 1. We divide the Web page into parts by our default method during the layout pattern acquisition. 2. We judge whether a target part is one of the divided parts. We redivide the part if this part contains both the target part and other undesired parts until the target part becomes a single part. 3. We select the target parts and save the paths of parts as a form: body : 0 : ID/N1 : O1 : ID1 /N2 : O2 : ID2 /.../Nn−1 : On−1 : IDn−1 /Nn : On : IDn , where, Nn is the node name of the n-th node, On is the order of the n-th node among the sibling nodes, IDn is the ID value of the n-th node, and Nn−1 is the parent node of Nn . 3.4 Partial Information Extraction 3.4.1 Path Selection We need to select the layout pattern corresponding to the Web page using the following steps: 1. We transform the HTML document to the XML document. 2. We select the saved layout pattern one by one. 3. We find out a layout pattern corresponding to the XML document if all the paths in this layout pattern can be found in the XML document. Then, we regard that the list of paths corresponding to the found layout pattern is the paths of partial information of this Web page. 3.4.2 Subtree Extraction We extract the subtrees according to the corresponding paths, and every subtree represents a part of Web page. If the data type of a part is continuous occurrence, the corresponding sibling trees with the same node names and ID are extracted, too.
342
H. Han and T. Tokuda / A Personal Web Information/Knowledge Retrieval System
3.4.3 Text Extraction According to the defined data types, we extract the partial information from the extracted subtrees in text format excluding the tags of HTML document. For the single text type, the partial information is the node value of the corresponding single leaf node. For the single object type, the partial information is the attribute value of corresponding single node. For the continuous text type, the partial information is the list of extracted values from the corresponding list of single subtrees. For the continuous object type, the partial information is the list of values extracted from the list of continuous subtrees. For example, the extracted information of a photo is the value of attribute ”src” of node , and the extracted information of the list of paragraphs of an article is the list of values of continuous leaf nodes.
4 Reuse of Layout Patterns and Paths 4.1 Reuse of Layout Patterns If we can find the similar paths in a layout pattern of the HTML document of a Web page, this Web page may be similar to the Web page corresponding to this layout pattern and this layout pattern may be reused. We give a definition of similar paths of layout pattern. Similar Path of Layout Pattern: Two paths are similar to each other, if these two paths have the same forms ignoring the difference of orders of nodes among sibling nodes, and the difference of orders is within a defined deviation range. The form of path is as follows: body : 0/N1 : (O1 − h ∼ O1 + h)/N2 : (O2 − h ∼ O2 + h)/.../Nn−1 : (On−1 − h ∼ On−1 + h)/Nn : (On − h ∼ On + h), where, Nn is the node name of the n-th node, On is the order of the n-th node among the sibling nodes, Nn−1 is the parent node of Nn , and h is the deviation value. For example, body : 0/f orm : 0/table : 1/tr : 0/td : 0 is similar to body : 0/f orm : 0/table : 2/tr : 0/td : 0 as shown in Fig. 5. 4.2 Reuse of Paths If we find a layout pattern that can be applied to a Web page, the specified paths corresponding to this layout pattern can be reused to extract the partial information. Firstly, we give a definition of similar path of part and get a list of similar paths. Similar Path of Part: Two paths are similar to each other, if these two paths have the same forms ignoring the difference of orders of nodes among sibling nodes, and the difference of orders is within a defined deviation range. The form of path is as follows: body : 0 : ID/N1 : (O1 −h ∼ O1 +h) : ID1 /N2 : (O2 −h ∼ O2 +h) : ID2 /.../Nn−1 : (On−1 −h ∼ On−1 +h) : IDn−1 /Nn : (On − h ∼ On + h) : IDn , where, Nn is the node name of the n-th node, On is the order of the n-th node among the sibling nodes, IDn is the ID value of the n-th node, Nn−1 is the parent node of Nn , and h is the deviation value. Then, we use the ID value to choose the most appropriate paths with the minimum deviation value from the deviation range, and reuse them to extract the partial information.
H. Han and T. Tokuda / A Personal Web Information/Knowledge Retrieval System
Figure 5: Similar paths
343
Figure 6: Resulting tables
5 Resulting Tables We need a default resulting table to present the result after we extract the partial information. We have two types of resulting tables: horizontal type and vertical type. A horizontal type resulting table has a number of columns equal to the sum of the number of selected parts. A column is identified by the name of selected parts. The first row is the header row to display the column names. We also have vertical type resulting tables. Examples of resulting tables are shown in Fig. 6.
6 Examples In this section, we will give some examples to show the process of partial information extraction and presentation. We extract the partial information from the top news pages of Yahoo! News and CNN.com and present the extracted information in a resulting table. 1. We specify the top page of Yahoo! News with the crawling area. 2. We define the names and specify the data types of the target parts: news title part ”YahooNewsTitle” of single text type, news contents part ”YahooNewsContents” of continuous text type, and photo part ”YahooPhoto” of single object type. 3. We acquire the layout pattern of a typical Web page and divide the Web page to select the target parts.
Figure 7: Layout pattern of a typical Web page
Figure 8: Page division and parts selection
344
H. Han and T. Tokuda / A Personal Web Information/Knowledge Retrieval System
<Pattern>
4. We do the same operations as Step1 ∼ 3 for CNN.com. 5. Our system extracts the partial information and presents the extracted information in a resulting table as shown in Fig. 9. We can extract the partial information from all kinds of the Web pages, such as company names and industrial classifications from Yahoo! Finance [8], and country names and profile information from BBC Country Profiles [1] as shown in Fig. 10.
Figure 9: A resulting table
Figure 10: Extracted partial information
7 Evaluation Our method is one of tree-oriented approaches. There have been research groups that focus on the problem of extracting information from Web pages based on the tree-oriented approaches. Crunch [3] is a HTML tag filter to retrieve the contents from the DOM trees of Web pages. However, the users have to spend much time in configuring a desired filter after analyzing the source of HTML documents of the Web pages. Internet Scrapbook [4] is a system which allows users to create a personal page by clipping parts of Web pages by specifying parts of Web pages to be extracted, which can not be applied to similar Web pages. PSO [7] is an approach to extract the parts of Web pages. It keeps the view information of the extracted parts by using the designated paths of tree structures of HTML documents, and users need to find out the paths from the HTML document of Web page by hand. Similarly, ANDES [5] is an XML-based methodology to use the manually created XSLT processors to realize the data extraction. HTML2RSS [6] is a system to automatically generate RSS feeds from HTML documents that consist of time-series items such as blog, BBS, chats and mailing
H. Han and T. Tokuda / A Personal Web Information/Knowledge Retrieval System
345
lists. [9] can automatically identify individual data records and extract data items from them. The extraction ranges of [6, 9] are limited to the Web pages that consist of list of data items with similar data structures or special data structures. Our system allows users to compose one resulting table from various parts of Web pages in a number of Web sites. Also our system allows users to select target parts without noticing the explicit tree structure of Web pages. Table 1 shows the performance of our system. Table 1: Number of correctly extracted Web pages with the number of layout patterns Web site Deviation Total pages number 1 pattern 2 patterns 3 patterns Yahoo! News 1 76 74 75 76 CNN.com 1 31 29 30 31 BBC Country Profiles 10 267 251 265 267 Yahoo! Finance 1 215 214 215 /
8 Conclusion We have presented our personal Web information/knowledge retrieval system based on XML tree approach. Our system allows users to extract information and knowledge from the partial information of Web pages in one Web site or a number of Web sites. We can easily select the target parts of typical Web pages and reuse the extracted paths to reach the partial information of general Web pages having similar structures. The contents extracted from Web pages may be used for personal data backup or Web site analysis of data for public purposes. The reproduction or republication of extracted contents may not be allowed. It is important for users of the personal Web information/knowledge retrieval system to conform to all copyright rules of contents on the Web. Our future work would be to provide mechanisms of static or dynamic combination of tasks of partial information/knowledge retrieval including a task of retrieving metadata such as RDF.
References [1] BBC Country Profiles. http://news.bbc.co.uk/1/hi/country profiles/default.stm. [2] JTidy. http://jtidy.sourceforge.net/. [3] Suhit Gupta and Gail Kaiser. Extracting Content from Accessible Web Pages. In Proceedings of the 2005 International Cross-Disciplinary Workshop on Web Accessibility (W4A), 2005. [4] Yoshiyuki Koseki and Atsushi Sugiura. Internet scrapbook: Automating web browsing tasks by demonstration. In ACM Symposium on User Interface Software and Technology,pages 9-18, 1998. [5] Jussi Myllymaki. Effective Web Data Extraction with Standard XML Technologies. In Proceedings of the 10th international conference on WWW, 2001. [6] Tomoyuki Nanno and Manabu Okumura. HTML2RSS: Automatic generation of RSS feed based on structure analysis of HTML document. In Proceedings of the 15th international conference on WWW, 2006. [7] Tetsuya Suzuki and Takehiro Tokuda. Path set operations for clipping of parts of web pages and information extraction from web pages. In Proceedings of the 15th International Conference on Software Engineering and Knowledge Engineering, pages 547-554. Knowledge Systems Institute, 2003. [8] Yahoo! Finance. http://biz.yahoo.com/ic/ind index.html. [9] Yanhong Zhai and Bing Liu. Web data extraction based on partial tree alignment. In Proceedings of the 14th international conference on WWW, 2005.
346
Information Modelling and Knowledge Bases XIX H. Jaakkola et al. (Eds.) IOS Press, 2008 © 2008 The authors and IOS Press. All rights reserved.
$3HUVRQDO,QIRUPDWLRQ3URWHFWLRQ0RGHO IRU:HE $SSOLFDWLRQVE\8WLOL]LQJ0RELOH3KRQHV 0LFKLUX7DQDNDD-XQ6DVDNLD<XWDND)XQ\XDDQG
D
$EVWUDFW ,Q WKLV SDSHU ZH SURSRVH D ULQJ W\SH LQIRUPDWLRQ WUDQVIHU PRGHO IRU LPSURYLQJSHUVRQDOLQIRUPDWLRQOHDNDJHSUREOHPVRQZHEEDVHGVHUYLFHV,QRUGHU WR PDWHULDOL]H WKH PRGHO ZH SURSRVH DQ DSSOLHG PHWKRG E\ XWLOL]LQJ PRELOH SKRQHV DQG PDWUL[ FRGHV ,Q DGGLWLRQ ZH GLVFXVV OHDNDJH ULVNV RI WKH SURSRVDO PHWKRG DQG VKRZ WKDW WKH SURSRVHG PHWKRG FDQ UHGXFH WURXEOHV RQ SHUVRQDO LQIRUPDWLRQOHDNDJHZLWKRXWFKDQJLQJXVHUWHUPLQDOHQYLURQPHQWV
,QWURGXFWLRQ ,Q -DSDQ WKH 3HUVRQDO ,QIRUPDWLRQ 3URWHFWLRQ /DZ ZHQW LQWR HIIHFW LQ $SULO DQG OHDNDJHSUREOHPVRISHUVRQDOLQIRUPDWLRQKDYHEHHQFORVHGXSDVVRFLDOSUREOHPVVLQFHWKH HQIRUFHPHQW2QFHSHUVRQDOLQIRUPDWLRQLVOHDNHGLQWRWKH,QWHUQHWLWLVGLIILFXOWWRGHOHWHLW 7KHUHIRUHVHFXUHZD\VWRVHQGDQGUHFHLYHWKHLQIRUPDWLRQDUHVWURQJO\H[SHFWHG 'LIIHUHGIURPHQWHUSULVHVZKHUHFRXQWHUPHDVXUHVDJDLQVWVHFXULW\SUREOHPVDUHWDNHQ KRPH QHWZRUN HQYLURQPHQW LV H[SRVHG WRWKH OHDNDJH ULVN E\ PDOZDUH &RQVHTXHQWO\ WKH SUREOHP FDQ QRW EH VROYHG VLPSO\ E\ DQWLYLUXV VRIWZDUH $FFRUGLQJ WR D UHSRUW E\ WKH 0LQLVWU\RI,QWHUQDO$IIDLUVDQG&RPPXQLFDWLRQVRI-DSDQ>@LQDTXHVWLRQ³,I\RXXVHWKH LQWHUQHWZKDWZRUULHVRUDQQR\V\RXDERXWLW"´DQRSWLRQSURWHFWLRQSHUVRQDOLQIRUPDWLRQ ZDVPRVWVHOHFWHGE\ DERXWUHVSRQGHQWV1HYHUWKHOHVVWKHUDWLRRIWKHUHVSRQGHQWVZKR KDGWDNHQPHDVXUHVDJDLQVWLWLVDERXW,IZHFRQVLGHUWKHJDSEHWZHHQWKHZRUULHVDQG WKH PHDVXUHV ZH FDQ LPDJLQH WKDW VHFXUH DQG HDV\ PHWKRGV IRU VHQGLQJ DQG UHFHLYLQJ SHUVRQDOLQIRUPDWLRQDUHVWURQJO\UHTXLUHG ,Q WKLV SDSHU ZH SUHVHQW PRGHOV IRU PRUH VHFXUHO\ DQG HDVLO\ WUDQVIHUULQJ SHUVRQDO LQIRUPDWLRQ ZKLFK LV LQSXWWHG RQ EURZVHUV LQWR QHWZRUN VHUYLFH SURYLGHUV ZLWKRXW LQVWDOOLQJDQ\VRIWZDUHDQGKDUGZDUHLQWRXVHUV SHUVRQDOFRPSXWHUHQYLURQPHQW $3URSRVHG0RGHO 5LQJ7\SH,QIRUPDWLRQ7UDQVIHU0RGHO :H SURSRVH D ULQJ W\SH LQIRUPDWLRQ WUDQVIHU PRGHO DV VKRZQ LQ )LJXUH DV D SHUVRQDO LQIRUPDWLRQ SURWHFWLRQ DUFKLWHFWXUH IRU LPSURYLQJ WKH SUREOHPV LQ H[LVWLQJ LQIRUPDWLRQ WUDQVIHUPRGHO,QWKHILJXUHWKH163PHDQV1HWZRUN6HUYLFH3URYLGHUDQGLWLVVXFKDVD ZHEDSSOLFDWLRQWKH8$PHDQVXVHUDJHQWVXFKDVDZHEEURZVHURQXVHUV¶ WHUPLQDO
M. Tanaka et al. / A Personal Information Protection Model for Web Applications
347
8QFHUWDLQSHUVRQDOLQIRUPDWLRQ SURWHFWLRQHQYLURQPHQW +LJKOHYHO SHUVRQDO LQIRUPDWLRQ SURWHFWLRQ HQYLURQPHQW 3HUVRQDO ,QIRUPDWLRQ 0DQDJHPHQW 6HUYLFH 3URYLGHU
163 1HWZRUNDFFHVV IXQFWLRQ (QFU\SWHGFKDQQHOE\ 66/7/6EDVHGRQ3.,
8$
3HUVRQDOLQIRUPDWLRQ LQSXWIXQFWLRQ
3,063
3HUVRQDOLQIRUPDWLRQILOWHULQJ 3HUVRQDOLQIRUPDWLRQPDQDJHPHQW 3ULYDF\SROLF\PDQDJHPHQW 3XEOLFNH\FHUWLILFDWLRQPDQDJHPHQW
(QFU\SWHGLGHQWLILDEOH SHUVRQDOLQIRUPDWLRQ
)LJXUH5LQJWRSRORJ\WUDQVIHUPRGHOIRUSHUVRQDOLQIRUPDWLRQSURWHFWLRQ 7KH IHDWXUH RI WKLV PRGHO LV WKDW LI XVHUV ZDQW WR SUHYHQW LQIRUPDWLRQ OHDNDJH WKH LQIRUPDWLRQ VKRXOG EH GHWRXUHGWKURXJK PRUH VHFXUH HQYLURQPHQW ,W LV D[LRPDWLF WKDW ZH KDG EHWWHU RQO\ XVH WKH KLJK VHFXUH HQYLURQPHQW LI ZH WKLQN LW VLPSO\ +RZHYHU LI WKH XVDELOLW\LVORZHUHGDQGWKHXVDJHFRVWLVKLJKE\XWLOL]LQJWKHPHWKRGDQGWKHHQYLURQPHQW LWEHFRPHVWKHUHDOLVWLFVROXWLRQWRXVHLWRQO\ ZKHQXVHUVGHDOZLWKLPSRUWDQWLQIRUPDWLRQ ,QDGGLWLRQLIWKHLQIRUPDWLRQZKLFKXVHUVGRQRWZDQWWREHOHDNHGLVQRWWUDQVIHUUHG LQWR 163 GLUHFWO\ OLNH WKH OLQHU W\SH PRGHO DQG WKH LQIRUPDWLRQ LV GHWRXUHG WKURXJK FRPPXQLFDWLRQ SDWKV RI ZKLFK KLJKHU VHFXULW\ DUH JXDUDQWHHG LW FDQ EH UHOD\HG YLD D SHUVRQDO LQIRUPDWLRQ PDQDJHPHQW VHUYLFH SURYLGHU 3,063 7KH 3,063 FDQ PDQDJH SHUVRQDO LQIRUPDWLRQ EODQNHWO\ DQG LV DEOH WR ILOWHU WKH LQIRUPDWLRQ IRU H[DPSOH LW FDQ PDNHWKHLQIRUPDWLRQDQRQ\PRXVEHIRUHWKHSHUVRQDOLQIRUPDWLRQUHDFKHVDWDUJHW163 0LQLPXP,QIRUPDWLRQ8VDJHE\3XEOLFNH\(QFU\SWLRQ 7KHLWHPZKLFKLVWKHPLQLPXPXVDJHRISHUVRQDOLQIRUPDWLRQE\163VLVGHVFULEHGDVRQH RIWKHV\VWHPUHTXLUHPHQWLQWKHSUHYLRXVVHFWLRQ,QFRPSDULVRQZLWKWKHJHQHUDOWUDQVIHU PRGHO WKH ULQJ W\SH WUDQVIHU PRGHO LV HDV\ WR FRQWURO SHUVRQDO LQIRUPDWLRQ EHFDXVH WKH LQIRUPDWLRQFDQEHUHOD\HGYLD3,063LQDULQJQHWZRUNDQGEHFRPELQHGZLWKSXEOLFNH\ HQFU\SWLRQPHWKRGVDQGSULYDF\SROLFLHV7KHIROORZLQJVDUHWKHLWHPVRIWKHGHVLJQSROLF\ WRXVHSHUVRQDOLQIRUPDWLRQPLQLPDOO\E\163V •
7KHSODFHZKHUHSHUVRQDOLQIRUPDWLRQLVHQFU\SWHGLVLQDXVHUDJHQW
•
(QFU\SWLQJSHUVRQDOLQIRUPDWLRQLVWKHEDVLFUXOHH[FHSWWKHLQSXWWHGLQIRUPDWLRQLV QRWLGHQWLILHGDVSHUVRQDOLQIRUPDWLRQ
•
3HUVRQDO LQIRUPDWLRQ LV HQFU\SWHG ZLWK D SXEOLFNH\ RI D RUJDQL]DWLRQ D GHSDUWPHQWDGHYLFHRUDSURJUDPZKLFKGHDOVZLWKWKHLQIRUPDWLRQVXEVWDQWLDOO\ 7KHHQFU\SWLRQLVEDVHGRQDSULYDF\SROLF\ZKLFKLVSURYLGHGE\D163DQGZKLFK DXVHUDJUHHVWR
•
7KH HQFU\SWHG GDWD LQ WKH XVHU DJHQW LV VWRUHG LQ RQO\ SHUVRQDO LQIRUPDWLRQ UHSRVLWRU\RI3,063
348
M. Tanaka et al. / A Personal Information Protection Model for Web Applications
%\ PDNLQJ XVH RI SXEOLFNH\ HQFU\SWLRQ DV IDU DV D SULYDWHNH\ D SDLUNH\ RI WKH SXEOLFNH\ LVQRWVWROHQLWLVDOPRVWEH\RQGSRVVLELOLW\WKDWWKHGDWDLVGHFU\SWHGLOOHJDOO\ )XUWKHUPRUH LI WKH V\VWHP FDQ LVVXH GLJLWDO FHUWLILFDWH IRUWKH SXEOLFNH\ DOWKRXJK LW ZLOO WDNH D ORW RI FRVWV VLQFH SXEOLFNH\ IRUJHU\ FDQ EH SUHYHQWHG LW FDQ UHGXFH PDQ\ LQIRUPDWLRQ OHDNDJH WURXEOHV (VSHFLDOO\ 3,063 FDQ PDQDJH SHUVRQDO LQIRUPDWLRQ XQLILHGO\ +RZHYHU SHUVRQV ZKR FDQ GHFU\SW HQFU\SWHG GDWD DUHRQO\ WKRVH ZKR KDYH WKH SULYDWHNH\ ZKLFK LV WKH SDLU RI WKH SXEOLFNH\ ZKLFK LV XVHG IRU HQFU\SWLQJ 7KHUHIRUH HYHQ LI FRQWHQWVRI D SHUVRQDO LQIRUPDWLRQ UHSRVLWRU\ DUHRR]HG LW GRHV QRW PHDQ WKDWWKH SHUVRQDOLQIRUPDWLRQLVOHDNHGLPPHGLDWHO\ 5ROHVRID3,063 $V VKRZQ LQ )LJXUH WKH FRPPXQLFDWLRQ EHWZHHQ D 3,063 DQG HDFK RI 163V FDQ EH DOORZHG DIWHUWKH 66/7/6 ELGLUHFWLRQDO DXWKHQWLFDWLRQ LV GRQH DQG LQGLYLGXDO UHODWLRQV RI WUXVW DUH EXLOW $QRQ\PLW\ XQLILHG PDQDJHPHQW RI SHUVRQDO LQIRUPDWLRQ DQG UHFRUG DQG GLVFORVXUHRIDFFHVVORJVDUHWKHDGYDQWDJHVE\SXWWLQJWKLV3,063EHWZHHQ8$DQG163 3,063 0XOWLSOH XVHUV
0XOWLSOH SURYLGHUV 3ULYDF\ 3ROLF\
8$
3XEOLF.H\ &HUWLILFDWLRQ
3HUVRQDO 5HSRVLWRU\ 5HSRVLWRU\ ,QIRUPDWLRQ 66/7/6 66/7/6 5HSRVLWRU\ XQLEL 2QHWLPH ELGLUHFWLRQDO
GLUHFWLRQDO DXWKHQWLFDWLRQ
$FFHVV /RJ
7RNHQ '%
163
DXWKHQWLFDWLRQ
)LJXUH,QWHUQDODQG([WHUQDO&RPSRQHQWVDQGWKHUHODWLRQRQD3,063 $SSOLFDWLRQRI5LQJ7\SH,QIRUPDWLRQ7UDQVIHU0RGHO 0DNLQJ8VHRI0RELOH3KRQHV 7R UHDOL]H WKH ULQJ W\SH PRGHO ZH KDYH WR OLQH XS PRUH WKDQ WZR FRPPXQLFDWLRQ SDWKV 7RGD\PDQ\RI3&VDVXVHUWHUPLQDOVKDYHFRQQHFWHGWRRQO\RQHFRPPXQLFDWLRQFKDQQHO ,IZHDSSO\WKLVPRGHOWRFXUUHQWJHQHUDOHQYLURQPHQWLWZLOOWDNHDORWRIFRVWV 0HDQZKLOHLQ-DSDQDWWKHPRPHQWRI0DUFKLQWKHQXPEHURIPRELOHSKRQH VHUYLFHFRQWUDFWRUVLVDERXWPLOOLRQVDQGWKDWRI,3FRQQHFWLRQVHUYLFHRIPRELOHSKRQHV LV RYHU PLOOLRQV LW PHDQV PRVW RI SHRSOH KDYH PRELOH SKRQHV >@ 0RELOH SKRQHV DUH DYDLODEOH LQ PRVWRIDUHDRI-DSDQ7KHUHIRUHZH PDWHULDOL]HWKHULQJW\SHWUDQVIHU PRGHO ZLWKWKHFRPPXQLFDWLRQFKDQQHOYLDWKLVPRELOHSKRQHVDQGWKHLUFDUULHUQHWZRUN $QRWKHU EHQHILW RI PRELOH SKRQHV LV LWV KLJK VHFXULW\ ,Q -DSDQ NH\ ORJJHUV DQG VS\ZDUH KDYH EHHQ KDUGO\ GLVFRYHUHG LQ -DSDQHVH FDUULHU V PRELOH SKRQHV 2QH RI WKH UHDVRQVLVDFFHVVUHVWULFWLRQWRWKHPHPRU\DUHDLQPRELOHSKRQHVEHFDXVHWKHUHLVVRPDQ\ SHUVRQDOLQIRUPDWLRQRZQHUVRIWKHSKRQHVXVXDOO\VWRUHSKRQHQXPEHUVRUDGGUHVVOLVWRI IULHQGV DQG DFTXDLQWDQFHV RI WKHP )XUWKHUPRUH PRELOH SKRQHV ZKLFK KDYH IHDWXUHV RI ELRPHWULFV DXWKHQWLFDWLRQ VXFK DV ILQJHUSULQW RU IDFH DXWKHQWLFDWLRQ KDYH FRPH WR HPHUJH
M. Tanaka et al. / A Personal Information Protection Model for Web Applications
349
UHFHQWO\ 7KH WUHQG LV FRQYHQLHQW WR LPSOHPHQW WKH PHWKRG ZKLFK GHDOV ZLWK SHUVRQDO LQIRUPDWLRQVHFXUHO\DVXVHUDJHQWV $Q(QKDQFHG5LQJ7\SH,QIRUPDWLRQ7UDQVIHU0RGHO ,QWKHFDVHRIWKHPRGHORI)LJXUHD8$LVLOOXVWUDWHGDVRQHHOHPHQWDQGLWFRUUHVSRQGVWR D ZHE EURZVHU RQ D 3& ,I ZH XVH D PRELOH SKRQH DQG LWV FDUULHU V QHWZRUN DV D QHZ FRPPXQLFDWLRQFKDQQHOWKHFKDQQHO VKRXOGOHDGWRD3,063IURP WKH8$ 7KHUH DUH SUREOHPV WR EH VROYHG ZKLFK DUH KRZ WR JXDUDQWHH WKH VHFXULW\ RI LQSXWWLQJ SHUVRQDO LQIRUPDWLRQ LQ D ZHE EURZVHU RQ D 3& DQG KRZ WR FRQQHFW D PRELOH SKRQHWRD3&7KHIRUPHU LVGLIILFXOWWREHDVVXUHGLQKRPHXVHUV 3&HQYLURQPHQWV7KH ODWWHULVDOVRGLIILFXOWDQGZLOOWDNHDORWRIHIIRUWVDQGFRVWVEHFDXVHWKHFRQQHFWLRQEHWZHHQ WKH EURZVHU DQG WKH SKRQH QHHG D ZLUHOHVV GHYLFH RU D 86% FDEOH DQG LWV VRIWZDUH LQVWDOODWLRQ (YHQ LI WKH FRQQHFWLRQ LV HVWDEOLVKHG VPRRWKO\ WKHUH LV D SUREOHP RI KRZ WR VHOHFW D FRPPXQLFDWLRQ SDWK IURP WKH WZR SDWKV RQ D EURZVHU ,Q WKH FXUUHQW JHQHUDO EURZVHUV LW VHHPV WR EH LPSRVVLEOH WR FKRRVH WKH SDWK &RQVHTXHQWO\ ZH SURSRVH DQ HQKDQFHG ULQJ W\SH LQIRUPDWLRQ WUDQVIHU PRGHO E\ XWLOL]LQJ D PDWUL[ FRGH UHDGHU RQ D PRELOHSKRQHDVVKRZQLQ)LJXUH 7KH HQKDQFHG SRLQW LV WR GLYLGH D 8$ LQWR D 8$ DQG D 8$ 8$ LV D XVHUDJHQW VXFKDVDZHEEURZVHUDQG8$LVDXVHUDJHQWVXFK DVDQDSSOLFDWLRQRQDPRELOHSKRQH :H DVVXPH WKDW 8$ LV PRUH VHFXUH WKDQ 8$ )RU WKH KLJK VHFXULW\ WKH FRQQHFWLRQ EHWZHHQ3,063DQGHDFK163KDVWREHDXWKHQWLFDWHGELGLUHFWLRQDOO\E\66/7/6DQGWKH FRQQHFWLRQ EHWZHHQ 3,063 DQG 8$ RU EHWZHHQ 8$ DQG D 163 KDV WR EH VHUYHUVLGH DXWKHQWLFDWHGE\66/7/6DWOHDVW 2QHWLPHWRNHQVZKLFKDUHUHODWHGWRLGHQWLILHUVRID163DQGD8$DUHLVVXHGRQD 3,0637KHWRNHQVDUHSDVVHGOLNH3,063ĺ163ĺ8$ĺ8$ĺ3,0637KH3,063 YHULILHVWKHWRNHQVDQGXQLWHVWKH8$DQGWKH8$ 7
8$
163 7
(QFU\SWHGFKDQQHO E\66/7/6
8VHU$JHQW 1HWZRUNDFFHVV IXQFWLRQ
7
2QHZD\ DQDORJFRPP
3,063 ,VVXHDQGYHULILFDWLRQ RIDRQHWLPHWRNHQ
8$ 7
8VHU$JHQW 3HUVRQDOLQIRUPDWLRQ LQSXWIXQFWLRQ
)LJXUH(QKDQFHGULQJW\SHWUDQVIHUPRGHODQGWKHIORZRIRQHWLPHWRNHQV $IWHU D 8$ DQG D 8$ DUH XQLWHG WKH 8$ DQG WKH 163 LQWHUDFW LQGLUHFWO\ ZLWK HDFK RWKHU IRU H[FKDQJLQJ SHUVRQDO LQIRUPDWLRQ %DVLFDOO\ 8$ GRHV QRW GHDO ZLWK WKH SHUVRQDOLQIRUPDWLRQ2QHWLPHWRNHQVDUHGHOLYHUHGIURP8$ZKLFKPD\QRWEHVHFXUH
350
M. Tanaka et al. / A Personal Information Protection Model for Web Applications
LQWR 8$ ZKLFK LV FRPSDUDWLYHO\ VHFXUH E\ RQHZD\ DQDORJXH FRPPXQLFDWLRQ DV LI LW EHKDYHVOLNHDGLRGHWRSUHYHQWEDFNFXUUHQW7KLVLVPRUHVHFXUHWKDQWUDGLWLRQDOZD\V7KH ULVNZKLFK8$PD\KDYHPDOZDUHVXFKDVVS\ZDUHLVKLJK,QFRPSDULVRQZLWKWKHPXWXDO FRPPXQLFDWLRQEHWZHHQ8$DQG8$WKLVPRGHOGRHVQRWVHQGSHUVRQDOLQIRUPDWLRQDQG WKHRWKHULPSRUWDQWLQIRUPDWLRQGLUHFWO\DQGORJLFDOO\IURP8$LQWR8$7KHUHIRUHHYHQ LIPDOZDUHKDVEHHQLQVWDOOHGLQWRD3&WKHPDOZDUHFDQQRWJHWWKHSHUVRQDOLQIRUPDWLRQ 0DNLQJ8VHRI0DWUL[&RGHV $V FDQGLGDWHV ZKLFK DUH WKH PHWKRGV RI RQHZD\ DQDORJXH FRPPXQLFDWLRQ ZH FDQ WDNH VLJKW FRQILUPDWLRQ PHWKRGV DQG D PHWKRG RI PDNLQJ XVH RI D PDWUL[ FRGH UHDGHU RQ D PRELOH SKRQH ,Q -DSDQ PRELOH SKRQHV ZLWK WKH PDWUL[ FRGH UHDGHU DUH ZLGHO\ VSUHDG DPRQJSHRSOH$FFRUGLQJWR177'R&R0R VSUHVVUHOHDVH>@ZHDUHDEOHWRJXHVVWKDWWKH UDWLR ZKLFK PRELOH SKRQHV HTXLSSHG ZLWK WKH PDWUL[ UHDGHU LV PRUH WKDQ 7KH PRVW ZLGHO\XVHGPDWUL[FRGHW\SHLV45FRGH>@DQGLWFDQUHDGDODUJHDPRXQWRILQIRUPDWLRQ DVDPDWUL[FRGHTXLFNO\ZKLOHFRUUHFWLQJZLWKRXWHUURUV,QDGGLWLRQLWFDQOHDGDXVHUWRD VSHFLILFZHEVLWHTXLFNO\ZLWKRXWSXVKLQJEXWWRQVRQDPRELOHSKRQHWKHUHIRUHPDQ\XVHUV DQGEXVLQHVVFRPSDQLHVXVHWKHIHDWXUH :H KDYH DOUHDG\ SURSRVHG DQG LPSOHPHQWHG DQ DXWKHQWLFDWLRQ PHWKRG ZKLFK LV FDOOHG DV 68$1 6HFXUH 8VHU $XWKHQWLFDWLRQ ZLWK 1LMLJHQ FRGH DQG KDYH HYDOXDWHG LWV VHFXULW\ XVDELOLW\ DQG FRVWV IRU WKH LQVWDOODWLRQ DQG WKH PDLQWHQDQFH >@ 7KH V\VWHP FRPSRVLWLRQDQGWKH LQIRUPDWLRQ IORZDUHDULQJW\SH LQIRUPDWLRQWUDQVIHU PRGHODQG LW LV IRXQG WKDW WKH V\VWHP FDQ EH HQGXUHG XQGHU UHDO HQYLURQPHQW :LWK WKLV PHWKRG PRELOH SKRQHVZKLFKEHFRPHKDUGZDUHWRNHQVGRQRWQHHGWREHFRQQHFWHGWRXVHUWHUPLQDOV,Q DGGLWLRQXVHUVFDQXVHWKH PHWKRGZLWKRXWLQVWDOOLQJDQ\VRIWZDUHLQWRWKHXVHUWHUPLQDOV 7KHUHIRUHWKH\DUHFRVWHIIHFWLYHDQGZLGHO\DGRSWDEOH &RQVHTXHQWO\ WR PDNH XVH RI WKH PDWUL[ FRGH UHDGHU DV RQHZD\ DQDORJXH FRPPXQLFDWLRQEHWZHHQ8$DQG8$LVFXUUHQWO\WKHEHQHILFLDOPHDQV +RZWR8QLWH7ZR8VHU$JHQWV 7ZRXVHUDJHQWV8$DQG8$DUHWLHGZLWKHDFKRWKHUE\SDVVLQJWRNHQVLQWKHULQJW\SH FRPPXQLFDWLRQ SDWK 7KH FRQWHQWRI WKLV WRNHQ LV OLNH )LJXUH DQG FRQVWLWXWHV D SURWRFRO QDPHD163LGHQWLILHUDFRPPDQGQDPHDQGDWRNHQHQWLW\7KHWRNHQHQWLW\LVLVVXHGDV DQ XQSUHGLFWDEOHVWULQJE\RQHZD\KDVKIXQFWLRQVXFKDV6+$ZLWKSDUDPHWHUVOLNHD163 LGHQWLILHUDVHVVLRQLGHQWLILHUIRUDXVHUDJHQWVXFKDVDZHEEURZVHUDQGVRIRUWK VXLPHVKRSSLVHQGGGGIIIIFGFEIHFDHEGF 3URWRFRO 1DPH
163 &RPPDQG ,GHQWLILHU 1DPH
7RNHQ(QWLW\
)LJXUH([SUHVVLRQH[DPSOHRIDRQHWLPHWRNHQ
1LMLJHQFRGHVPHDQPDWUL[FRGHVLQ-DSDQHVH
351
M. Tanaka et al. / A Personal Information Protection Model for Web Applications
7KHIROORZLQJVDUHVWHSVIRUXQLWLQJEHWZHHQ D8$DQGD8$ 67(3,VVXLQJDQGGLVSOD\LQJWRNHQV)LJXUH 67(3$8$DFFHVVHVWRDQ1637KH163NHHSVWKHVHVVLRQ,'ZLWKDFRRNLH 67(37KH163UHTXHVWVWRD3,063ZLWKWKH163LGHQWLILHUDQGWKHVHVVLRQ LGHQWLILHU RIWKH8$IRUJHWWLQJDRQHWLPHWRNHQ 67(3 7KH 3,063 LVVXHV WKH RQHWLPH WRNHQ 7KH UHODWHG LQIRUPDWLRQ VXFK DV WKH FUHDWHGWLPHDQGWKHH[SLUDWLRQWLPHDOVRLVVWRUHGDWWKHVDPHWLPH 67(37KH3,063VHQGVWKHLVVXHGWRNHQLQWRWKH163 67(37KH163VHQGVWKHUHFHLYHGWRNHQLQWR8$ 67(37KH8$GLVSOD\VWKHUHFHLYHGWRNHQ 67(39HULI\LQJWKHWRNHQDQG8QLWLQJWKH8$DQGWKH8$ )LJXUH 67(37KH8$UHDGVWKHWRNHQZKLFK LVGLVSOD\HGRQWKH8$WKH8$UXQVD KDVK IXQFWLRQZLWKWKHUHFHLYHGWRNHQDQGWKHSDUDPHWHUVVXFKDVXVHUWHUPLQDOQXPEHU 67(37KH8$VHQGVWKHWRNHQDQGWKHUHVXOWRIWKHKDVKIXQFWLRQWRWKH3,063 67(3 7KH 3,063 TXHULHV WKH VDPH WRNHQ IURP WKH WRNHQ GDWDEDVH ,I LW H[LVWV WKH 3,063YHULILHVZKHWKHUWKHH[SLUDWLRQWLPHLVYDOLGRUQRW,IWKHVDPHWRNHQGRHVQRWH[LVW LQWKHGDWDEDVHRUWKHWLPHLVLQYDOLGJRWR67(3 67(3,IWKHYHULILFDWLRQLVVXFFHVVIXOWKH3,063VWRUHVDUHFRUGZKLFKLVWLHGZLWKWKH VHVVLRQLGHQWLILHUVRIWKH8$DQGWKH8$DQGWKH163LGHQWLILHU 67(37KH3,063QRWLILHVWKHUHVXOWWRWKH163 67(3¶7KH3,063QRWLILHVWKHUHVXOWWRWKH8$ 67(3 8$ LG
163 67(3
8$
8$
163
67(3
67(3 163LG 8$ LG
67(3 5HVXOW
7 67(3 67(3 /LQNLQJ
3,063
67(3 ,VVXHG 7RNHQ
7
7
8$ 8$
67(3 3,063
7
8$
)LJXUH3URFHVVIORZRI67(3
7 67(3¶ 5HVXOW
7 0DWFKLQJ 67(3 9HULILFDWLRQ
8$
7 67(3
)LJXUH3URFHVVIORZRI67(3
+RZWR6HQG3HUVRQDO,QIRUPDWLRQ ,I ZH FRQVLGHU WKH ORZ XVDELOLW\ RI SXVKLQJ EXWWRQV RQ PRELOH SKRQHV IRU LQSXWWLQJ FKDUDFWHUV LW LV QRW SUHIHUDEOH WR UHLQSXW WKH VDPH SHUVRQDO LQIRUPDWLRQ HYHU\WLPH XVHUV KDYH WR VHQG WKHLU LQIRUPDWLRQ &RQVHTXHQWO\ WKH V\VWHP KDV WKH LQSXWWHG SHUVRQDO LQIRUPDWLRQEHFRPLQJUHXVDEOH7KH IROORZLQJVDUHWKHH[DPSOHVWHSV IRUXVHUVWRUHJLVWHU WKHLUSHUVRQDOLQIRUPDWLRQLQWR163VVXFKDVZHEVKRSV 67(3$XVHUDFFHVVHVWRDXVHUUHJLVWUDWLRQSDJHIRUDZHE VLWHRIDQ163 67(3 7KH XVHU XWLOL]HV WKH PDWUL[ FRGH UHDGHURQ KHUKLV RZQ PRELOH SKRQH8$ DQG
352
M. Tanaka et al. / A Personal Information Protection Model for Web Applications
UHDGVWKHPDWUL[FRGHZKLFKLVGLVSOD\HGRQWKHXVHUWHUPLQDODQGWKHXVHUDJHQW3&DQG 8$ $IWHUWKHFRGHLVUHDGVXFFHVVIXOO\WKHV\VWHPXQLWHVWKH8$DQGWKH8$ 67(3 7KH 163 VHQGV DQ LWHP OLVW RI SHUVRQDO LQIRUPDWLRQ LWV SULYDF\ SROLF\ DQG WKH UHODWHGSXEOLFNH\FHUWLILFDWHVWRWKHPRELOHSKRQHRIWKHXVHUYLDWKH3,063 67(3 8$ GLVSOD\V LQIRUPDWLRQ RI WKH WDUJHW 163 VXFK DV D 85/ DQG FRQILUPV DQ DJUHHPHQWRIWKHSULYDF\SROLF\ WRWKHXVHU 67(3 7KH XVHU VHOHFWV KHUKLV SURILOH RI SHUVRQDO LQIRUPDWLRQZKLFK LQFOXGHV DGGUHVVHV SKRQHQXPEHUVDQGVRIRUWK E\RSHUDWLQJWKHPRELOHSKRQH 67(3,IWKHXVHUDJUHHVZLWKWKHSROLF\WKH8$HQFU\SWVDVHWRIWKHWDUJHWLQIRUPDWLRQ ZLWKWKHSXEOLFNH\ RIWKH163RUWKHSURJUDPZKLFKXVHVWKHLQIRUPDWLRQVXEVWDQWLDOO\ 67(37KH8$WUDQVPLWVWKHGDWDLQWRWKH3,063DQGWKHGDWDLVVWRUHGLQWKH3,063 67(3 :KHQ WKH 163 QHHGV WKH SHUVRQDO LQIRUPDWLRQ LW UHTXHVWV DQG UHWULHYHV WKH HQFU\SWHGGDWDIURPWKH3,063 DQGGHFU\SWVWKHGDWDZLWKWKHSULYDWHNH\ 'LVFXVVLRQ 3KLVKLQJ XVXDOO\ VWHDOV XVHU QDPHV DXWKHQWLFDWLRQ LQIRUPDWLRQ VXFK DV SDVVZRUGV RU WKH RWKHUSHUVRQDOLQIRUPDWLRQE\OHDGLQJXVHUVWRSKLVKLQJVLWHVYLDPDLOOLQNVWRWKHVLWHV,Q WKH FDVH RI WKH SURSRVHG PHWKRGLI D XVHU LV OHG LQWR SKLVKLQJ VLWHV SHUVRQDO LQIRUPDWLRQ FDQ QRW EH VHQW IURP PRELOH SKRQHV WR D 3,063ZLWKRXW D YDOLG PDWUL[ FRGH (YHQ LI WKH YDOLG WRNHQ LV GLVSOD\HG LQ WKH SKLVKLQJ VLWH WKH SODFH ZKHUH WKH PRELOH SKRQHV VHQG SHUVRQDO LQIRUPDWLRQ WR LV RQO\ DQ 163 ZKLFK KDV WKH WUXVW UHODWLRQVKLS ZLWK WKH 3,063 7KHUHODWLRQVKLSEHWZHHQHDFKRI163VDQGWKH3,063LVEDVHGRQ66/7/6ELGLUHFWLRQDO DXWKHQWLFDWLRQ ,Q RUGHU WR JHW FHUWLILFDWH IRU WKH 66/7/6 VHUYHU DXWKHQWLFDWLRQ 163V XVXDOO\ KDYH WR EH DVVHVVHG ZKHWKHU WKH RUJDQL]DWLRQ 163 LV YDOLG RU QRW +RZHYHU UHFHQWO\ SKLVKLQJ VLWHV ZKLFK KDYH D YDOLG FHUWLILFDWH EHFRPH WR HPHUJH 7R SUHYHQW EXLOGLQJYLFLRXVUHODWLRQVKLSWKHSURSRVHGPHWKRGQHHGVRWKHUVSHFLILFLQVSHFWLRQV ,Q -DSDQ WURXEOHV RQ VS\ZDUH DQG NH\ ORJJHUV LQ PRELOH SKRQHV KDYH KDUGO\ EHHQ GLVFRYHUHG EHFDXVH RI WKH UHVWULFWLRQ RI LQIRUPDWLRQ DFFHVV RQ DSSOLFDWLRQV RI PRELOH SKRQHV 3HUVRQDO LQIRUPDWLRQ LV LQSXWWHG DQG VHQW RQ D 8$ PRELOH SKRQH ZKLFK FRUUHVSRQGV WR WKH DSSOLFDWLRQ DQG LW LV GLIILFXOW IRU SHUVRQDO LQIRUPDWLRQ WR EH OHDNHG EHFDXVH D 8$ ZHE EURZVHU ZKLFK KDV KLJKHU ULVN WR EH LQIHFWHG ZLWK VS\ZDUH WKDQ D 8$ ,Q DGGLWLRQ SHUVRQDO LQIRUPDWLRQ LV QHYHU VHQW ORJLFDOO\ IURP D 8$ LQWR DQ 163 7KHUHIRUHSHUVRQDOLQIRUPDWLRQLVQHYHUOHDNHGRQWKH8$'XHWRWKHUHDVRQVZKLFKDUH PHQWLRQHG DERYH WKH SHUVRQDO LQIRUPDWLRQ OHDNDJH ULVN LV YHU\ ORZ 5HVXOWV RI WKH ULVN VLPXODWLRQVLQGLFDWHWKDWWKHSURSRVHGPRGHOFDQUHGXFHWKHOHDNDJHSUREDELOLW\>@ $VIRUXVDELOLW\PDWUL[FRGHUHDGHUVDUHHDV\WRXVHEHFDXVHXVHUVRQO\SXVKDEXWWRQ DQG KROG D PRELOH SKRQH WR D PDWUL[ FRGH ,Q DGGLWLRQ RQFH XVHUV LQSXW SHUVRQDO LQIRUPDWLRQ WKH\ QHHG QRW LQSXW WKH LQIRUPDWLRQ UHSHDWHGO\ EHFDXVH WKH SHUVRQDO LQIRUPDWLRQFDQEHVWRUHGDQGUHXVHGRQPRELOHSKRQHV
M. Tanaka et al. / A Personal Information Protection Model for Web Applications
353
&RQFOXVLRQ ,Q WKLV SDSHU ZH SURSRVHG WKH ULQJ W\SH LQIRUPDWLRQ WUDQVIHU PRGHO ZKLFK FDQ UHGXFH SHUVRQDO LQIRUPDWLRQ OHDNDJH ZKHQ FRQVXPHUV LQSXW SHUVRQDO LQIRUPDWLRQ LQWR ZHEIRUPV LQZHESDJHVDQGVHQGLWLQWRZHEVHUYHUV,QDGGLWLRQDVWKHFRQFUHWHPHDQVEDVHGRQWKH ULQJ W\SH LQIRUPDWLRQ WUDQVIHU PRGHO ZH LQWURGXFHG WKH V\VWHP PRGHO E\ PDNLQJ XVH RI PRELOHSKRQHVDQGPDWUL[FRGHV ,Q WKLV ULQJ W\SH WUDQVIHU PRGHO ZH VKRZHG WKH SRVVLELOLW\ WR PDQDJH SHUVRQDO LQIRUPDWLRQXQLILHGO\DQGVHFXUHO\E\HQFU\SWLQJWKHLQIRUPDWLRQLQPRELOHSKRQHVZLWKWKH SXEOLFNH\ ZKLFK WKH NH\ LV XVHG E\ DQ RUJDQL]DWLRQ RU D SURJUDP VXEVWDQWLDOO\ )XUWKHUPRUHRQHRIWKHIHDWXUHVRIWKLVPRGHOLVWKDWZHQHHGQRWFKDQJHDQ\XVHUWHUPLQDO HQYLURQPHQW VXFK DV 3&V LQ RUGHU WR PDWHULDOL]H WKLV PRGHO 1HYHUWKHOHVV ZH FDQ DYRLG PRVW RI NH\ORJJHUV DQG SKLVKLQJ WURXEOHV EHFDXVH QR SHUVRQDO LQIRUPDWLRQ FDQ EH WUDQVIHUUHGEHWZHHQDQ163DQGDXVHUV¶3&,QDGGLWLRQSHUVRQDOLQIRUPDWLRQPDQDJHPHQW VHUYLFH SURYLGHUV FDQ EH UHGXFHG EHFDXVH WKH SHUVRQDO LQIRUPDWLRQ LV HQFU\SWHG DW XVHUV¶ PRELOHSKRQHVZLWKWKHSXEOLFNH\(YHQLIWKHGDWDDWWKHSURYLGHULVVWROHQLWFDQQRWEH GHFU\SWHGXQOHVVWKHSULYDWHNH\LVVWROHQ 'LIIHUHGIURPWUDGLWLRQDOPHWKRGVIRUSURWHFWLQJSHUVRQDOLQIRUPDWLRQOHDNDJH>@>@ WKLVPHWKRGFDQEHXVHGZLWKRXWDQ\VRIWZDUHDQGKDUGZDUHLQVWDOODWLRQLQWRXVHUWHUPLQDOV H[FHSWPRELOHSKRQHV7KHUHIRUHWKHSRVVLELOLW\ RIWKHVSUHDGLVKLJKDQGWKHHQKDQFHPHQW LVVWURQJO\H[SHFWHGIRUDFRXQWHUPHDVXUHDJDLQVWSHUVRQDO LQIRUPDWLRQOHDNDJHSUREOHPV 5HIHUHQFHV >@0LQLVWU\RI,QWHUQDO$IIDLUVDQG&RPPXQLFDWLRQVRI-DSDQ&RPPXQLFDWLRQV8VDJH7UHQG6XUYH\LQ &RPSLOHG KWWSZZZMRKRWVXVLQWRNHLVRXPXJRMSWVXVLQBUL\RXGDWDHQJBWVXVLQBUL\RXBSGI >@ 7HOHFRPPXQLFDWLRQV&DUULHUV$VVRFLDWLRQ7KHQXPEHU RIVXEVFULEHUVRI0RELOH7HOHSKRQH3+6 ,QWHUQHW 3URYLGHU6HUYLFHVDQG5DGLR3DJLQJ KWWSZZZWFDRUMS >@177'R&R0R,QFKWWSZZZQWWGRFRPRFRP >@'HQVR:DYH,QFKWWSZZZTUFRGHFRP >@0LFKLUX7DQDNDHWDO $0HWKRGDQG,WV8VDELOLW\IRU8VHU$XWKHQWLFDWLRQE\8WLOL]LQJD0DWUL[&RGH 5HDGHURQ 0RELOH3KRQHV3URFHHGLQJVRI,QWHUQDWLRQDO:RUNVKRSRQ,QIRUPDWLRQ6HFXULW\$SSOLFDWLRQV SS$XJXVW >@0LFKLUX7DQDNDHWDO $3HUVRQDO,QIRUPDWLRQ3URWHFWLRQ0HWKRGIRU:HE$SSOLFDWLRQVE\8WLOL]LQJD 0RELOH3KRQHVRQD5LQJ7RSRORJ\,QIRUPDWLRQ7UDQVIHU0RGHO,(,&( 7UDQVDFWLRQVRQ,QIRUPDWLRQDQG 6\VWHPV9RO-'1RSS)HE >@35,96+(/7(5,QF3ULY6+(/7(5 KWWSZZZSULYVKHOWHUFRP >@0LQ:XHWDO:HE:DOOHW3UHYHQWLQJ3KLVKLQJ$WWDFNV E\5HYHDOLQJ8VHU,QWHQWLRQV3URFHHGLQJVRIWKH 6HFRQG6\PSRVLXPRQ8VDEOH3ULYDF\DQG6HFXULW\-XO\
354
Information Modelling and Knowledge Bases XIX H. Jaakkola et al. (Eds.) IOS Press, 2008 © 2008 The authors and IOS Press. All rights reserved.
Manufacturing Roadmaps as Information Modelling Tools in the Knowledge Economy Augusta Maria PACI EPPLab ITIA - CNR Via dei Taurini 19, 00185 Rome, Italy
Abstract. Roadmaps are the authorative medium-high tech viewpoints for the competitiveness and sustainability of industrial and public organizations. The paper provides an overview of possibilities to use roadmaps as virtual tools which contribute to support, coupled with knowledge management, industrial innovation and new industry processes. In the age of digitization, virtual roadmaps as a full participatory process supports new conceptual modelling of the manufacturing domain and at the same time it enables knowledge workers to share and elaborate innovative concepts. Hence roadmaps enable the design of new collaborative knowledge management environments. In medium time horizon, considering the global dimension of manufacturing, roadmaps can be reference tools for cooperation agreements in bilateral and multilateral projects. The paper provides a case study of roadmaps for advanced manufacturing and an example of the open innovation model.
1. Roadmaps in the Age of Digitization In the last ten years, public research organizations and manufacturing industries have developed roadmaps to foster industrial innovation and new industry. Roadmapping is often connected with foresight studies [1,2,3,5]. In the industrial economy, firms exploit technology roadmaps without previously contributing to their development. Conversely, approaching the European knowledge economy, technology roadmapping means a full participatory process of both research organizations and firms to identify R&D solutions to industrial targets for innovation. This process is a continuous cycle that assesses through feedback short, medium and long-term development plans. This paper explores the use of roadmaps tools for investigation in industrial innovation through research based innovation. In the age of digitization [6], virtual roadmaps go beyond the description of time scaled plans and priorities, which are the main result of traditional paper roadmaps1 [3,4,14]. 2. Virtual Roadmaps for Knowledge Management Roadmaps are a new type of predictive tools, that facilitate new communication and organizational learning processes “on doing things right first time”. Roadmapping, as a full participatory process, supports the transformation of an organization culture, enabling inflows and outflows of knowledge from individuals and groups to the organization level and the creation of new knowledge. Virtual roadmaps are tools that allow people to share knowledge and communicate through a common language in any type of organization. As such, these tools are essential in the global market for new solutions 1
According to IMTI Roadmapping methodology: “Roadmaps define the desired future vision, identify the goals required to achieve the vision, and define requirements and tasks to achieve the goals. This approach serves to develop programmes that involve the organization to respond to challenges.”
A.M. Paci / Manufacturing Roadmaps as Information Modelling Tools in the Knowledge Economy 355
responding to the needs of virtual and networked enterprises and research institutions. In the industrial technology domain, new needs are: definition of the knowledge domains and multiple tiers, consensus building among relevant communities, influence on decision makers, policy convergence among different stakeholders, prioritization of interventions, time horizon of targets, investments planning, dissemination of best practices and flagship projects. The digital roadmaps are becoming new communication means in any type of organization and community: industry, research, education, public institutions, national and local government, market. Particularly, virtual and networked enterprises intensify collaboration with partners, suppliers, advisors and other towards the future. Roadmapping - as a full participatory process - facilitate information modelling according to the SECI model: [7] x socialization: sharing tacit knowledge that is built upon existing experiences x externalization: articulating knowledge and developing the organization “intellectual capital” through dialogue x combination: expliciting the borders of expectations in terms of “competitive advantage” x internalization: participating in a learning process. In this way, new roadmaps support the development of collaborative knowledge management and dynamically manage knowledge sharing. Therefore, roadmaps become a “social process” and a “collective framing” to encapsulate the intangibles elements that transmit tacit and explicit elements of the organizational knowledge; they bridge the dualism currently existing in ICT-based knowledge management between the “organization of knowledge” and the “business strategies”. They also provide expectations in market impact and support the increasing role of performance measurement and scientific management. Virtual roadmaps expand the field of knowledge management and serve as meta-description of future industrial and technology areas. Looking at roadmaps as new tools implies that they: x are authorative medium-high tech viewpoints - detailing future vision, goals, requirements and targets filtered through complex participatory process - for the competitiveness and sustainability of industrial and public organizations; x represent a hierarchy of complex contents in macro-area, sub-areas, detailed topics, detailed technologies, which are relevant for looking-forward approaches to foster and sustain the organization’s new products and services; x communicate through different styles and channels, from complex schemas for technical analysis to simple presentation for effective and immediate impact; x are agents for the diffusion of priorities for innovation among people -bridging high level business decision and practical high tech work- fostering the application of research results to practice in the knowledge economy innovation chain; x describe and disseminate concepts and goals in an uniform language; x contribute to the collection of data, value creation that can be measurable within organizations in terms of efficiency and market impact. According to the specific industry’s high tech development plan and strategy, any roadmap represents the targets of the organization’s medium and long term strategy: quantitative values, time horizons, high level requirements, market diffusion and impact, resource allocation, supporting activities, infrastructures, facilities and best-practices inside and outside the organization.
356 A.M. Paci / Manufacturing Roadmaps as Information Modelling Tools in the Knowledge Economy
These new roadmaps predict how dynamically create the conditions for intelligent business in industrial domains. These new tools can leverage the Seven Knowledge levers2, facilitating the knowledge creation process, handling the daily situations within turbulent environments, managing the human dimension and sense-making interpretations. 3. Case study on advanced manufacturing Referring to the above-mentioned main principles, the manufacturing high-tech domains have been studied as a Case study of Manufacturing Roadmaps. The most recent and comprehensive new roadmapping concept in manufacturing technologies is the authoritative high-level representation of the five pillars of the ManuFuture industrial transformation reference model [8,9]. These macro-domains concern transectoral RTD areas that require solutions based on key and emerging technologies for new production systems and business models. The ManuFuture roadmaps aim to achieve European industrial innovation for high added value products and services providing time-scales and prioritized topics (Fig. 1) [8].
TRANSFORMATION
Agenda objectives
Drivers
TRANSFORMATION OF INDUSTRY OF
Goals
MAKE/DELIVERY HVA PRODUCTSSERVICES
INNOVATING PRODUCTION
R&D
INNOVATING RESEARCH
Competition Rapid Technology Renewal Eco-sustainability Socio economic Environment
New Added Value Products and Services
New Business Models
Advanced Industrial Engineering
ShortMedium-Term
Medium Term
Emerging Manufacturing Sciences and Technologies
Infrastructures and Education
Regulation Values -public acceptability TIME SCALE
Continuous
Long Term
Long Term
Fig. 1: ManuFuture industrial transformation reference model (source: ManuFuture Strategic Research Agenda, September 2006)
The stakeholders who contributed to the ManuFuture Platform have set out plans to use these transectoral technology macro-domains and corresponding roadmaps. Many other strategic sources, like platforms’ Strategic Research Agendas, roadmaps and studies have been analysed to set the targets of knowledge-based industrial development for European manufacturing. Within the European manufacturing community, wide consultations were carried towards industrial and research bodies to gain relevant contributions. Later on, after an intensive work, further roadmapping for Manufuture [10,12,13] developed specific transectoral technology roadmaps that were presented in the Manufuture Conference in Tampere for further validation and comments (http://manufuture2006.fi/) [11]. 4. Towards a Collaborative Knowledge Management The roadmapping process in virtual environments supports the design of new collaborative knowledge management, consolidating, exploiting and maintaining the knowledge produced and consolidated in the process. This new collaborative knowledge management may exploit the SECI modalities fostering: 2
The Seven Knowledge levers are Customer knowledge, stakeholder relationship, business environment insights, organizational memory, knowledge in processes, knowledge in products and services, knowledge in people
A.M. Paci / Manufacturing Roadmaps as Information Modelling Tools in the Knowledge Economy 357 x
x
x
the combination modality enabling the knowledge conversion, the two-ways interaction between: high-level management of public and private organizations aiming at developing technology policy to win the market competition; people who learned in the process which knowledge and targeting goals are envisaged by the organization. This modality consolidates the transfer of the roadmaps concepts among knowledge workers (individuals and groups), through social interactions based on ICT technologies. the internalization modality enabling to internalize and practice the roadmapping concepts. This avoids a passive acceptance by knowledge workers, and triggers a participative and continuous validation process with verification procedures and control measures. the socialization modality enabling to widespread the understanding and use of the roadmaps as agents for diffusion of culture and innovation.
5. Open Model for Collaborative Knowledge Management The new collaborative knowledge management, that integrates the roadmapping process, provides an example of the Open innovation model [14]. In this example (figure 2), input from roadmaps provide specific elements for innovation while information modelling provide specific elements knowledge management. The combination of roadmaps and information modelling operating a convergence between prediction and responsiveness permits the creation of a new collaborative environment. Collaborative knowledge management (source An open Innovation Paradigm. CHESBROUGH, 2006 Elaboration EPPLab, 2006)
BUSINESS/TECHNOLOGIES POLICY STRATEGIES
ROADMAPS VALUE CREATION
PERFORMANCE MEASUREMENT
INFORMATION MODELING
KNOWLEDGE CREATION KNOWLEDGE WORKERS
PREDICTION
RESPONSIVENESS
Fig. 2: Collaborative knowledge management based on Open model
Therefore expectations and future goals are integrated with inflows and outflows through a continuous participatory process. This process responds to a fast changing environment and to the need of alignment of people capacity toward innovation. 6. Global Dimension In medium time horizon, considering the global dimension of manufacturing and innovation strategies, the new roadmaps can be applied as virtual reference tools within cooperation agreements and bilateral and multilateral projects. In the knowledge economy, these roadmaps will represent the high-tech manufacturing language. Like super highways, new roadmaps are the communication infrastructure for industrial innovation and new industry. They will allow the info-mobility of knowledge workers along complex high-tech concepts and innovation projects.
358 A.M. Paci / Manufacturing Roadmaps as Information Modelling Tools in the Knowledge Economy
In this spirit, Japan public research organizations say that: “By combining the knowledge of industry, government, and academic fields, METI established our country's first "Strategic Technology Roadmap" in 20 different fields. Strategic Technology Roadmap indicates the technical goals and demands of pro-ducts/services necessary for the production of new industry. Hereafter, Strategic Technology Roadmap will be offered to industry, government, and academic fields to promote cooperation of one another, and also to be used for managing METI research & development.” [15] 7. Conclusion In the knowledge economy, new roadmaps as reference tools support new ICT- based knowledge management, playing a role to achieve successful results in industrial innovation and new industry. They contribute to the concept design of collaborative environments for global knowledge creation and sharing. In the age of digitization, roadmaps enable to optimize the learning and the knowledge transfer allowing knowledge workers to cooperate remotely around common and strategic innovation goals. References [1] [2] [3] [4] [5] [6] [7] [8] [9] [10] [11] [12] [13] [14] [15]
DREHER C., ManVis Main Report, Fraunhofer-ISI, 2005. FUTMAN PROJECT, The future of manufacturing in Europe 2015-2020: Main report, 2004. IMTR/IMTI, Roadmapping Methodology, http://www.imti21.org/resources/docs/roadmapping.htm. INSTITUTE OF MANUFACTURING, Informan EUREKA Project, 2003. MANUFUTURE HIGH LEVEL GROUP, ManuFuture A Vision for 2020. Assuring the future of manufacturing in Europe, Report of the High Level Group, EU DG Industrial research, 2004. MACKENZIE OWEN J., The scientific article in the age of digitization, Springer, 2007. NONAKA I., The Knowledge-Creating Company , in: Harvard Business Review on Knowledge Management. Harvard Business School Press,. pp. 21-45, 1998 EUROPEAN COMMISSION MANUFUTURE PLATFORM, ManuFuture Strategic Research Agenda: September 2006, ISBN 92-79-01026-3. (www.manufuture.org). TOKAMANIS C., Improve the competitiveness of European Industry. ManuFuture Conference, Tampere, Oct. 2006, http://manufuture2006.fi/presentations/. PACI A.M., A collaborative industry-research frame for roadmapping in Production Engineering Conference, Wroclaw 7-8 December, pp 5-10, 2006. WESTKAEMPER E., Manufuture RTD Roadmaps: from vision to implementation, ManuFuture Conference, Tampere, Oct. 2006, http://manufuture2006.fi/presentations/. JOVANE F., PACI A.M., et al., Area Tecnologie di gestione e produzione sostenibile. In: II Rapporto sulle priorità nazionali della ricerca scientifica e tecnologica, Fondazione Rosselli (ed.), Milano, Guerini, pp 310-349, 2005. WILLIAMS D., Road mapping - A personal perspective, in Seminar: “Supporto alla ricerca in collaborazione con l’industria nell’area Sistemi di Produzione: strumenti e metodologie, CNR, Rome, 28 nov. 2006. CHESBROUGH H., Open innovation researching: a new paradigm. Oxford University Press, 2006, http://www.openinnovation.eu/. NEDO (New Energy and Industrial Technology Development Organization) Roadmap http://www.nedo.go.jp/roadmap/index.html.
Technical support provided by dr. Cecilia Lalle (EPPLab ITIA - CNR)
Information Modelling and Knowledge Bases XIX H. Jaakkola et al. (Eds.) IOS Press, 2008 © 2008 The authors and IOS Press. All rights reserved.
359
Metadata Extraction and Retrieval Methods for Taste-impressions with Bio-sensing Technology Hanako Kariya† Yasushi Kiyoki†† †Graduate School of Media and Governance, Keio University ††Faculty of Environmental Information, Keio University 5322 Endoh, Fujisawa, Kanagawa 252-8520, Japan {karihana,kiyoki}@sfc.keio.ac.jp Abstract. In this paper, we present a new generation food information retrieval by Taste-impression equipped with bio-sensing technology. The aim of our method is to realize the computing environment for one of the un-discussed basic human perception: sense of taste. Our method extracts Taste-impression metadata automatically by using sensor outputs retrieved from a taste sensor, according to 1) user’s desirable abstraction levels of terms expressing Taste-impression and 2) characteristic features of foods, such as a type, a nationality, and a theme. We call those characteristics of foods and drinks, “Taste Scope”. By extracting Taste Scope dependent metadata applying a bio-sensing technology, our method transforms sensor outputs expressing primitive taste elements into meaningful Taste-impression metadata and computes correlations between target foods (or drinks) and a query described in Taste-impression. Users can intuitively search any kinds of information regarding foods and drinks on the basis of abstract Taste-impression preferences with user’s desired granularity. We clarify the feasibility and effectiveness of our method by showing several experimental results.
1 Background Issues In recent years, a lot of recipe and drink databases are accessible through global area computer networks. These information resources are rapidly added and deleted according to the dynamic transition in food industry. There are currently two significant issues in foods and drinks search behavior for food consumers and creators (Target users of our retrieval method). First issue lies in the side of consumers. Consumers unfortunately do not have any attractive approaches to find his or her favorite food products, on the basis of their preferences of taste sense. Exisiting food data retrieval systems support users’ finding their favorite products or recipes, by merely providing product names and brands searches. Therefore, users’ relying on their own experiences is the only solution to reach favorite foods of his/her favorite tastes among numerous data replaced rapidly. The second issue is in the creator side. For example, food developers strrugle in designing new products on a daily basis. In order to design reputable and sustainable product, food developers need to understand desirable food or drink images of consumers. This would only be realized not by existing method such as advertisement but taste design itself. Furthermore, the product development needs to be performed for various foods and drinks concurrently in a limited period of time. Therefore, search environments for integrating the anonymous food data altogether according to his or her objective taste design vision are essential for competitive food products development. In order to solve such difficulties, an information retrieval system for “Impression” on the basis of user’s taste preferences should be of clear benefit to overall food business and consumers. In this paper, we propose metadata extraction and retrieval methods for Taste-impressions with bio-sensing technology, focusing on a metadata extraction method. Our impressionbased retrieval is realized by query expressed as verbal expression of impression keywords
360
H. Kariya and Y. Kiyoki / Metadata Extraction and Retrieval Methods for Taste-Impressions
Figure 1: Target User Categories with Query Expressions of Our Method
such as “rich”, “fresh” or taste pattern as shown in Figure 6. Our metadata extraction method automatically generates Taste-impression metadata for target data, according to features of each food or drink, such as a type, a nationality, and a theme by applying sensor outputs retrieved from the taste sensor. 2 Basic Concept Our approach to Taste-impression-based retrieval is based on two concepts (Figure 2). 1. “Taste Scope” adoption to a metadata extraction mechanism for optimizing sensor outputs makes it possible to manipulate the complexity of Taste-impression. 2. Application of bio-sensing technology to metadata extraction method, i.e. transforming cognitive data into metadata expressing verbal queries allows the impression-based retrieval with user’s optimal granularity in taste expression.
Figure 2: Basic Approach and Concept of our Method
H. Kariya and Y. Kiyoki / Metadata Extraction and Retrieval Methods for Taste-Impressions Automatic Metadata extraction for target data from sensor outputs
Beer Search engine
Target Data with Sensor Outputs
Soup Scope Beer Scope
361
Metadata for Soup Scope
㫌㫄㪸㫄㫀 㫊㪸㫃㫋㫐 䊶䊶䊶 㪸㪺㫀㪻 Metadata 䌦or Beer Scope
Beer &
Rich
㪹㫀㫋㫋㪼㫉 㫊㫎㪼㪼㫋 䊶䊶䊶 㪹㫆㪻㫐 Individual Feature Sets (Metadata) to represent different Taste Scopes
Query
Figure 3: Overview of Taste Scope adoption to Metadata Extraction
Feature1: Definition of Taste Scope in order to utilize anonymous contexts for terms expressing Taste-impression. One of the most important premises of Taste Scope is that, foods and drinks showing exact same taste patterns of sensor outputs, do not necessarily mean that Taste-impression is always the same and vice versa. An impression word “rich” is one of the common expressions for both soups and Japanese Sake, for instance. However, the main feature (taste elements) of each impression is in “bitterness” and in “umami (palatability)”. These taste elements are completely different, but have significant impact for underlying meaning definition in this same Taste-impression word. In order to deal with such Taste-impression-specific complexity, it is indispensable to define metadata for Taste-impression in verbal expression by transforming sensor outputs according to these viewpoints, that is, “Taste Scope”. For reflecting Taste Scope, our metadata extraction method for foods and drinks data introduces two modules named “Tasteimpression Metadata Generation Module” and “Standardization Module”, which perform new optimization operations by reflecting the Taste Scope intelligence in our metadata processing (Figure 3). Feature2: Application of bio-sensing technology to metadata etxraction, in order to transform sensory information to verbal query expressing Taste-impression with user’s desired granularity A multi-channel taste sensor, namely known as an electronic tongue [8] [6], computes and outputs taste senses for various foods and drinks quantitatively and provides the objective scale for human sensory expression in food developing and quality control. Unlike existing sensors such as temperature and pressure sensors, which respond to single physical quantities, a multi-channel taste sensor can measure many kinds of chemical substances in each food synthetically and transform these substances into meaningful quantities of basic tastes such as saltines, sweetness and its continued stimuli (Hereinafter called after taste). This sensor has been developed on the basis of mechanisms found in the biological system, such as parallel processing of multidimensional information or by the use of biomaterial and hence called bio-sensing technology (Figure 4). By applying bio-sensing technology to Taste Scope intelligence in our metadata extraction, our impression-based retrieval makes it possible to compute the correlation of basic components of sensory information and verbal query expressing Taste-impression with user’s desired granularity of taste expression.
362
H. Kariya and Y. Kiyoki / Metadata Extraction and Retrieval Methods for Taste-Impressions
Figure 4: Taste Sensor developed by bio-sensing technology [8, 5]
3 Related Work Whereas main objective of our study is to extract metadata focusing on taste sense interpretation and its expressions as Taste-impression keywords, several related work could be found in terms of Sensor and Kansei Database fields perspectives. In this section, we classify related works into two categories: 1) Kansei and Impression-based retrieval systems, 2) Sensor Database systems, and present the main difference of these studies from our method. 3.1 Kansei and Impression-based retrieval system Kansei databases are studied in various fields to realize intuitive search environment for images, music [9], video streams [7] and so on. Just to name a few,“ A Metadata System for Semantic Search by a Mathematical Model of Meaning ”[15] realizes impression-based retrieval for images by automatically computing the color scheme and its correlated impression word. The aim of these studies is to deal with the global impression of impression words of digital images in database. The paper [14] presents an extraction method of boundaries with impression changes by using color information and N-gram for video streams. These approaches are applicable and effective method for impression-based retrieval of images and video streams, whose impression are unique identifiable. In contrast to these solution for extracting global impressions for media data, our method extracts the metadata of Taste-Scope-dependent impression to solve the complexity and diversity of taste sense, as shown in table 1. 3.2 Sensor Database systems Concept for applying sensory information to a database system has been popular in numerous fields, and new applications are being explored constantly [10, 11, 3].
H. Kariya and Y. Kiyoki / Metadata Extraction and Retrieval Methods for Taste-Impressions
363
Table 1: Conventional Kansei and Impression-based Retrieval and our method
Pattern Matching
Retrieval by Impression Query
Retrieval by Scopes
Non-compliant
Non-compliant
Existing Impression-based retrieval
Compliant
Non-compliant
Our Method
Compliant
Compliant
For instance, application techniques of database systems to finger verification has been widely used. The aim of these studies is to realize exact pattern matching of finger prints with sensor information stored in heterogeneous databases, such as optical sensor and thermal sweeping sensors [12]. Location based sensor database studies have been very popular in ubiquitous computing fields as well [1]. There are successful applications of location-aware mobile computing, most notably navigation systems based on GPS sensors (e.g. [2]). Other examples are the NaviCam [16] and active badge systems [4]. Generally speaking, objectives of existing applications are detection of the presence of an object or a condition, object recognition, object identification/classification, tracking, monitoring, and change detection. Additionally, conventional approaches have been relied on simple physical sensor outputs such as distance and temperature sensors to achieve objectives above. On the other hand, our method applies sensory data of bio-sensing technology to database application as the raw information resource for metadata extraction in order to realize Tasteimpression expression. Sensory information with the combination of Taste Scope enables the automatic and meaningful information provision in our metadata processing. 4 An Example for Query Processing In this section, we first demonstrate the actual usage of our system with user scenarios in order to present the significance of our method. Next, we present an example of the metadata extraction and query processing method in order to show the data flow of our method. 4.1 User Scenarios There are two types of query options available in our Taste-impression search in order to satisfy different kind of target user needs. Beer information retrieval is shown as an example here. Assume that several beer makers offer local databases to introduce their products and general consumers (user scenario with Search Option 1) and drink developers (user scenario with Search Option2) are using our system. Our query processing and system architecture are shown in Figure 5 and user interface is shown in Figure 6. Query example for general consumers (Search Option1) A consumer unfamiliar to alcoholic beverages is seeking beers for refreshing. Search Option1 has prepared to satisfy needs of general consumers with elementary familiarity for taste flavors. Since the user does not have any detailed knowledge regarding taste preferences, the user holds only elementary level of expression ability for desired taste pattern. Such user submits a query with the Taste Scope “beer” and Taste-impression “fresh” to express his/her abstract favorite taste images.
364
H. Kariya and Y. Kiyoki / Metadata Extraction and Retrieval Methods for Taste-Impressions
Figure 5: Query Processing and System Architecture
Query example for foods and drinks developers (Search Option2) A drink developer needs to design marketable beer product for next season. Since the user does not have enough time to seek desirable taste by try and error for only one of their portfolios products, need the system which strongly support taste design of products in intuitive manner. In this case, the user should hold not merely ambiguous Taste-impression images but more concrete taste design images, if to a lesser degree of expert sake sommelier or a buyer in specialized food importers. Search Option2 has prepared to meet with such needs of taste design professionals with intermediate familiarity in taste. The user submits a query with the Taste Scope “beer” and directly addresses their objective taste pattern. The user is able to find similar beer item which could be the future rival products, in advance to physical product implementation. By understanding such information to differentiate with others in our system, the user is able to adjust the direction of product development with a cost effective solution. 4.2 Data flow of our method Taste Scope, such as “beer scope”, is described and committed as the query, and it is used to manage overall scenario of query processing (Figure 7). According to this Taste Scope, our method selects a candidate set of Taste-impression words as well as features, subsets of sensor outputs, functions to calculate the sensor outputs and aggregation functions for the intermediate values of sensor outputs. These functions are evaluated through following steps. Step-1 Mapping of a set of retrieval candidates for target data: Metadata extraction method maps a set of retrieval candidates for target data in the beer scope to the database, which consists of IDs of beer items and sensor outputs, with URLs as local information regarding beers.
H. Kariya and Y. Kiyoki / Metadata Extraction and Retrieval Methods for Taste-Impressions
365
Figure 6: User Interface
Target Data
Automatic Metadata extraction for target data from sensor outputs
㪠㫄㫇㫉㪼㫊㫊㫀㫆㫅㩷“㪩㪼㪽㫉㪼㫊㪿㫀㫅㪾”
㪠㫄㫇㫉㪼㫊㫊㫀㫆㫅㩷“㪟㪼㪸㫍㫐”
㪤㪛㩷㪽㫆㫉㩷㪙㪼㪼㫉㩷㪪㪺㫆㫇㪼 㪤㪛㩷㪽㫆㫉㩷㪪㫆㫌㫇㩷㪪㪺㫆㫇㪼
㪤㪛㩷㪽㫆㫉㩷㪙㪼㪼㫉㩷㪪㪺㫆㫇㪼 㪤㪛㩷㪽㫆㫉㩷㪪㫆㫌㫇㩷㪪㪺㫆㫇㪼
㪤㪛㩷㪽㫆㫉㩷㪮㫀㫅㪼㩷㪪㪺㫆㫇㪼
䊶䊶䊶
Query
㪤㪛㩷㪽㫆㫉㩷㪮㫀㫅㪼㩷㪪㪺㫆㫇㪼
䊶䊶䊶
㪤㪼㫋㪸㪻㪸㫋㪸㩷㪪㫇㪸㪺㪼㩷㪽㫆㫉㩷㪼㪸㪺㪿㩷㪫㪸㫊㫋㪼㩷㪠㫄㫇㫉㪼㫊㫊㫀㫆㫅 㫀㫅㫋㪼㪾㫉㪸㫋㪼㩷㪸㫅㫆㫅㫐㫄㫆㫌㫊㩷㪺㫆㫅㫋㪼㫏㫋㫊㩷㫆㪽㩷㪻㫀㪽㪽㪼㫉㪼㫅㫋㩷㪫㪸㫊㫋㪼㩷㪪㪺㫆㫇㪼㫊
Figure 7: Overview of Taste Scope
Step-2 Standardizing for sensor outputs: Optimizations for sensor outputs are automatically processed by the operation P2 in the standardization module Pz for the beer scope. Step-3 Extracting metadata from sensor data: Sensor outputs processed in Step-3 (intermediate values) are converted to metadata for target data, which consist of important feature sets for the Taste-impression definition in the beer scope by the operation G1 and G3 in the Taste-impression metadata generation module Ge . Step-4 Calculating correlation: The query processing method measures correlation values among metadata for beer items and keyword “fresh” (Search Option1) or addressed taste pattern (Search Option2) selected in Step-3, and outputs URLs as ranking results. 5 Metadata Extraction and Retrieval Methods for Taste-impressions with Bio-sensing Technology In this section, we present a framework of metadata extraction and retrieval method for Tasteimpressions with bio-sensing technology. The main functions of our method consist of a metadata extraction function for Taste-impression, and its query processing function. The execution model and basic algorithm outline of our method are shown in Figure 8 and 9. The execution model of our method is described in the following order. 1. The overall execution model 2. The metadata extraction method
366
H. Kariya and Y. Kiyoki / Metadata Extraction and Retrieval Methods for Taste-Impressions
Query Processing :
Λ(T , Wn, C x ) → R
Wn T aste Domain W1
A set of retrieval candidate R belongs to W1
T1
R2
Step-1 A cidic-bitterness
S1
Pz : R → S
Data E xtraction
T
A set of retrieval candidate S (Standardized R)
P1
R1
T2
Cx T aste Impression C1
Query1
R B ase-bitterness
schema example
P2
Step-2
S2
A set of R esults R1
G1 Ge : S → M G2
S
… A cidic-astringency
β ( M , C x ) → R1 Metadata M1 belongs to W1
Step-3 astringency
bitterness
… sourness
€ E xtr action for T ar get data M etadata
α (T , Wn ) → M
Figure 8: Execution Model
3. The query processing method The feature of our method is in the metadata extraction, where sensor outputs are transformed into meaningful Taste-impression metadata with user’s requested granularity. This feature brings two contributions in our method. First contribution is the adoption of Taste Scope to our metadata extraction method. Our method makes an interpretation of information retrieved from the Taste Scope, and develops metadata for Taste-impression. Our metadata extraction method makes it possible to recognize and express the abstract and subtle impression representation of taste sense by handling sensor outputs realizing modules, that are, Taste Impression Metadata Generation Module Ge and Standardization Module Pz . These modules are implemented by reflecting the specialized knowledge for defining subtle flavor of target Taste Scope. These modules make it possible to integrate diversified Taste-impression definition from anonymous Taste Scope. Second contribution is query expression dealing with the heterogeneous abstraction levels of verbal expression regarding sense of taste. We have set the abstraction level of verbal expression as granularity. We have implemented our method to meet with the desired granularity for different target users with low-intermediate knowledge regarding foods and taste sense. Namely, our method makes it possible to realize information provision for users balancing their familiarity level concerning taste and abstraction level for expressing target data verbally. Less familiar to taste knowledge, higher abstraction level for query is set (Figure 1). 5.1 The overall Query Execution Model In this section, we present overall query processing procedures and basic fuctions for metadata extraction and retrieval methods. In our method, the meaning of Taste-impression is determined by the indication of Taste Scope. Specification of the Taste Scope is executed with Wn of query, which is reflected to the metadata generation for Cx (Cx ∈ C(Wn )) and selection for target data. Wn (Wn ∈ W ) consists of appropriate feature sets to define the impression in each Taste Scope. As for query
H. Kariya and Y. Kiyoki / Metadata Extraction and Retrieval Methods for Taste-Impressions
367
selection by scope(T, Wn ) → T selection by (T ) → R; for each z · z + n{ selection by format(R, z) → Rz normalization(Pz , Rz ) → S for each Rzl in Rz { Pz (Rzl ) → Sl Append (S, Sl ); } MetadataExtraction(S) → m for each Sl in S{ Ge (Sl ) → Ml ; Append (m, Ml ); } }
Union(M, m); Figure 9: Algorithm Outline
options, we present Search Option1 as Q1 and Search Option2 as Q2. The structures of a query is defined as: Q1 = (Wn , Cx )
(1)
Q2 = (Wn , {d1 , d2 , · · · dn })
(2)
Wn = {SID, {f1 , f2 , · · · fn }}
(3)
Cx = {SID, {d1 , d2 , · · · dn }}
(4)
Execution model of our method F is only performed by inputs of query described with this data structure. Overall query processing F targets retrieval candidate T in and outputs retrieved results T out , by computing the correlation among Wn and Cx (Q1) and and sorting T in based on calculated correlation values. Otherwise, the user who understands exact taste pattern to be expected would directly specify the feature values as shown in Q2. Since data selection of the operation F is indicated by Wn , retrieval results T out is subset data of T in . Overall query processing operation F is defines as: F (T in , Wn , Cx ) → T out |T out ⊂ T in
(5)
5.2 The Metadata Extraction Method for Taste-impression In this section, we present overall outline of our metadata-extraction method for Taste-impression. Our metadata extraction method consists of three functions and executed by following order.
368
H. Kariya and Y. Kiyoki / Metadata Extraction and Retrieval Methods for Taste-Impressions
Step-1 Mapping a set of retrieval candidate to Rl Step-2 Feature Value Optimization by Standardization Module Pz Step-3 Schema Optimization by Taste-impression metadata generation module Ge First, we present and formalize functions and data structures of our method. Second, we show our metadata processing procedure demonstrating typical operation examples for Beer Scope. We clarify our method by introducing its 1) metadata schema in progress and 2) reflected Scope knowledge, which serve as the basis for each operation. 1. Mapping a set of retrieval candidate In this step, a set of retrieval candidates for target data Tl is extracted from all candidates T , on the basis of selected scope identifier SID. Tl consists of SID, its own identifier OID, and entity data (information resources in network) data. Each extracted Tl is joined with sensor data Rl , which is also extracted from all candidates R by SID. Sensor data are also described as the set of SID, OID, and sensor outputs data. These data are mapped and treated as baseline data for metadata generation. Data structure of target data Tl and sensor outputs Rl are defined as: Tl = {SID, OID, data}
(6)
Rl = {SID, OID, data}
(7)
Since each tuple in Tl has the SID, our mapping process consists of: Step-1: Selection of target data Tl with the Scope ID1 which is equivalent to Beer Scope, Step-2: Join of Tl with R1 , sensor data with SID Step-3: Mapping of selected R1 as raw data for creating metadata. 2. Standardization Module Pz Mapped sensor outputs Rl are automatically pre-processed by the standardization module Pz , in order to optimize feature values for target Taste Scope. Pz 1) selects adequate functions for target Taste Scope, 2) receives Rl and 3) outputs standardized values Sl . Therefore we could regard retrieved Sl as intermediate, pre-processed values for metadata. Data structure of Pz and Sl , function of the standardization module Pz are defined as: Pz : Rl → Sl
(8)
Sl = {OID, data}
(9)
normalization{Pz , Rl ||z ∈ {1, 2, 3, · · · , n}, l ∈ {1, 2, 3, · · · , n}}
(10)
Figure 10 is an example of sensor outputs optimization procedure activated by Taste Scope “Beer”. Our feature value optimization process consists of P1 operator which reflects Beer Scope specific intelligence. One of the actual operators is: • The threshold adjustment operation: In Taste Scope for Beers, it is widely known that slight difference of feature values significantly contribute to the flavor composition. For instance, only small multiplication of bitterness drastically changes the impression from “fresh” to “mild”. To deal with this issue, the threshold values adjusted by a specialist are subtracted for each feature value to describe the typical base line values of Japanese beers (taste pattern of reference solution).
H. Kariya and Y. Kiyoki / Metadata Extraction and Retrieval Methods for Taste-Impressions
369
Figure 10: Sensor Outputs Optimization process with threshold adjustment example
3. Taste-impression metadata generation module Ge Standardized sensor outputs Sl are automatically processed by Taste-impression metadata generation module Ge , in order to extract adequate feature sets for Taste-impression definition in target Taste Scope. Ge selects adequate functions for target Taste Scope, and selected functions receive Sl and outputs standardized values Ml . Sl is converted to Ml by using 1) the extraction and composition of features and 2) weighting of feature values on the basis of the denominator for target Taste Scope. Function of Tthe the taste-impression metadata generation module Ge and data structure of Ml can be defined as follows, where each Ml is composed of same features of Wn . Ge : Sl → Ml
(11)
Ml = {OID, v(Wn )}
(12)
v = {(f1 , d1 ), (f2 , d2 ), · · · (fn , dn )}
(13)
Figure 11 presents the Taste-impression-metadata generation phase with Taste Scope knowledge regarding “Beer”, whose intelligence are reflected as operator G1 to G3 . • The schema integration operation G1 : In Taste Scope for Beers, one of features of sensor outputs, salinity, does neither harm nor good on impression definition and indifferent to impression composition. Therefore, this feature will be omitted from the correlation matching target by multiplying feature values by 0. Acerbity (c5) and after taste of acerbity (c6) have merged with the union operator, in order to transform the abstraction level of feature words to suitable verbal expression for users. • The weighting operation for Acidic-bitterness (Sensor outputs from Channel ID3) G2 : In Taste Scope for Beers, Acidic-bitterness plays crucial role for impression
370
H. Kariya and Y. Kiyoki / Metadata Extraction and Retrieval Methods for Taste-Impressions
Figure 11: The taste-impression metadata generation example for Beer Data
composition. Since this feature could be described as one of the most essential taste elements for metadata definition, should be emphasized with strong weight naturally. In this experiment, we have tentatively set the weighting coefficient to 10. • The weighting operation for Base-bitterness (Sensor outputs from Channel ID4) G3 : In Taste Scope for Beers, Base-bitterness has negative effect on impression composition. Human interpret and feel this taste element in beers as if bitterness in medicines, whereas the function as “umami” if added adequate amount to tomato juice, for instance. For reflecting this fact, the weighing operation here turns feature values of Base-bitterness into negative. Aim of this operation is to realize pointdeduction scoring for feature bitterness when merged (union) with other sorts of outputs related to the bitterness. The standardization module Pz and the taste-impression metadata generation module Ge deserve recognition and expression mechanisms for defining the diversified impression representation on taste sense. Taste Scopes are eventually expresssed as metadata for each Taste-impression and has ability to express anonymous meanings of taste-impression in different Taste Scopes. Note that while these operation examples are realized as beer-specific operations, operations themselves are applicable and re-usable for several Taste Scopes, if same constraint applies. That is, each function reflects Taste-Scope intelligence for defining characteristic features of keyword for a Taste-impression applicable to several Taste Scopes. Such module application in our method realizes a search environment for various heterogeneous foods and drinks data in a comprehensive manner.
H. Kariya and Y. Kiyoki / Metadata Extraction and Retrieval Methods for Taste-Impressions
371
Figure 12: Correlation Calculation
5.3 The correlation calculation operator The correlation calculation operator β 1) computes correlations between Taste-impression metadata Me (Me ∈ M ) of target data and each Taste-impression words Cx (Cx ∈ C) and 2) outputs semantically close taste data Rl as retrieval results according to the user’s query (impression word or taste pattern and target scope). By employing our operation β, our method sorts the target data Tl in descending order on the basis of calculated correlation values, and enables ranking for target data according to impression words complied with Taste Scope. The data structure and function of Correlation calculation operator β are defined as: β(Ml , Cx ) → Rl
(14)
Our method provides two types of taste impression search options which eventually incorporated into operator β. Whereas Query1 correlate Wn with Taste-impression keyword with given feature values by the professionals for impression definition (The most abstract impression expression in our method), Query2 provides less intuitive search by directly addressing one’s desirable taste pattern as shown in Figure 12. 6 An application to the Beer and Japanese Food Scopes By realizing taste-impression-based retrieval by Taste-impression with our metadata extraction method, users can intuitively search any kinds of information regarding foods and drinks on the basis of abstract Taste-impression preferences. For extracting target data of the beer and Japanese foods by Taste-impression, we have applied our metadata extraction method to local Japanese recipe and drink databases. 6.1 A Metadata Extraction Method for Taste-impression In this section, we represent the implementation of our metadata extraction method. We have applied experimental data and defined functions for each module.
372
H. Kariya and Y. Kiyoki / Metadata Extraction and Retrieval Methods for Taste-Impressions
Figure 13: Principle of Taste Sensor (Offered by Insent, Inc.)
1. Sensor Outputs We show our information resources applied as row data for our experimental system. We present generation principle for sensor outputs 1 and an extraction method for an experimetal sensor data. In order to realize metadata extraction for the beer and Japanese Taste Scopes, we have applied real sensor outputs of beer data (28 tuples) and virtual outputs for Japanese food data (25 tuples). • Principle of Taste Sensor We have implemented our program applying taste sensing system proposed in [6] [8]. In a narrow definition of taste, human tongue receives taste as electric signal from foods and drinks composed of numerous chemical substances, whose 1) interaction has not been clear and 2) explanations are under developed. In order to deal with such difficulty for analyzing and evaluating taste, taste sensing technology applying human tong mechanism has developed as a multi-channel taste sensor and is widely used in food industry. Transducers of the sensor are composed of lipids immobilized with polyvinyl chloride. The multi-channel electrode is connected to a channel scanner through highinput impedance amplifiers. The electric signals are converted to a digital code by a digital voltmeter and then transferred to computer as shown in Figure 13. • Sensor Outputs of Taste Sensor The sensor output is not the amount of specific taste substances but the taste quality and intensity. The sensor has a concept of “global selectivity” which is the ability to classify enormous kinds of chemical substances into primitive taste elements such as saltiness, sourness, bitterness, umami and sweetness and its after taste (flavor stability on tongue). Electric signals obtained from the sensor are converted to taste quality based on the Weber-Fechner law which gives an approximately accurate generalization of the intensity of sensation. The base of logarithm is defined as 1.2. For example, 12.5 units means 10 times higher concentration than that of the original sample, and 125 units is 100 times higher concentration. Sensor outputs attributes consist of 16 features. Excerpt of sensor outputs for beers are shown in Table 2. 1 Description of Principle of Taste Sensor and Sensor Outputs of Taste Sensor have excepted and summarized according to [6] and [8].
H. Kariya and Y. Kiyoki / Metadata Extraction and Retrieval Methods for Taste-Impressions
beer brands
tartness
salinity
other bitterness
…
acidicbitterness
acerbity
Acerbity (after taste)
astringency
Kirin Rager
15.15
-4.87
-1.2
…
11.6
21.77
2.24
12.22
Kirin Ichiban-shibori
13.37
-5.34
-1.09
…
10.13
19.41
1.93
11.48
Sapporo Black Label
14.33
-6.49
-1.08
…
10.33
20.42
1.8
10.85
Suntory Malts
17.16
-4.83
-0.83
…
8.84
19.55
1.62
11.08
Asahi Super Dry
16.02
-8.73
-0.33
…
10.36
18.58
1.64
11.25
Kirin Tanrei
9.76
-8.55
0.16
…
9.79
16.26
1.63
12.47
373
Table 2: Subset data of Sensor Outputs for Beers
• Experimental Sensor data generated for Japanese Foods Similar virtual data are created by questionnaire for Japanese food data and applied to our experimental system tentatively. To generate virtual outputs, we have prepared 50 test subjects, 48 typical Japanese food items as experimental objects, and have conducted questionnaire by Semantic Differential method [13]. In this questionnaire, we have added free space for each target data so those tests subjective are able to write the impression words which they have came up with in his or her mind. We have applied these results for performance evaluation as well. Support rate has calculated as the ratio for the number of impression words written in each food among number of test subjective. We have set the threshold of support rate as 39% and eliminated 23 food data for convenience because main aim of this experiment is to search target data with impression, i.e. foods and drinks with low level of impression association for users are not as meaningful as target data of our system. 2. The standardization module Pz For sensor outputs conversion, we have implemented several functions for Pz as follows. P1 is defined as the comparative assessment for target Taste Scopes. It converts original feature values into suitable values for defining target Taste Scope respectively. We subtract the specific numbers for each feature value, which is adjusted by the specialist for target Taste Scope from original feature values. By this function, we are able to clarify the slight difference of feature values, and reflect subtle taste balance of flavor representation. P2 is defined as pre-processing function for feature values. This operation converts some of original feature values (sensor outputs) into absolute values. Since sensor outputs are expressed as electric potential, several features such as salinity and other bitterness have expressed in negative in original values. In order to comply with the semantic for vector expression for metadata, we have applied our P2 operation for features with such issues. P3 is defined as normalization function for feature values. Feature values are normalized to compute the norm of each vector between 0 and 1. By this function, we are able to resolve the big gap of gross average values between each feature marinating the balance of original important feature values. 3. Taste-impression metadata extraction module Ge For extracting adequate feature sets for Taste-impression definition in target Taste Scope, we have realized several functions for module Ge as follows. G1 is defined as integration function for feature values. Feature values are extracted and composed to define the target Taste Scope. By this function, we can produce suitable features for target Taste Scope.
374
H. Kariya and Y. Kiyoki / Metadata Extraction and Retrieval Methods for Taste-Impressions
G2 is defined as the emphasis assessment for feature values. We have tentatively multiplied feature values 10 times for emphasizing key feature value for impression composition of each Taste Scope. G3 is defined as another pre-processing function for feature values. This operation converts original feature values (sensor outputs) into negative. As described in the previous chapter, some of the taste elements such as base-bitterness in the beer case, has negative effect impression composition for beers. Aim of this operation is to realize point deduction scoring for feature when merged (union) with G1 . We can combine new functions for Pz and Ge in order to convert sensor outputs to suitable features and feature values for target Taste Scope, aside of the implemented functions above. 6.2 The Query Processing Method In order to realize an experimental information retrieval system, we have implemented the query processing method. The query processing method measures correlation values among metadata for target data (Ty ) and keyword described in Taste-impression (Cx in case of Search Option1), or directly insert the taste pattern (x1 to xm in case of Search Option2), and outputs the ranking results with URLs as local information for foods and drinks. We measure correlations between vectors of query and target data, using the operation of inner product. We have measured correlations using various ways, such as inner product, cosine correlation and comparison of vectors. In this paper, we have implemented the operation of inner product for measuring correlations. The Inner Product is a technique for calculating the amount of correlations between the query keyword and target data. Both of the query keyword and the target data are expressed respectively as vectors that have the same elements. The correlation function (Cx , Ty ) is defined as following formula:
(Cx , Ty ) =
mf
Wxi − Wyi
(15)
i−1
Cx = (Wx1 , Wx2 , · · · , Wxm )
(16)
Ty = (Wy1 , Wy2 , · · · , Wym )
(17)
7 Experimental Studies For evaluating feasibility of our system and its application, we have performed four experiments with following objectives. Experiment1: Feasibility evaluation for different Taste Scopes Experiment2: Performance evaluation for Japanese food scope Experiment3: Performance evaluation for Beer scope Experiment4: Functions Adjustment Evaluation for Beer Scope The overall experimental results have shown that our method has observed applicable to anonymous Taste Scopes as shown in Experiment1. Performance evaluation to the data in beer and Japanese food scopes assured retrieval results are reasonable in Experiment2 and 3. Furthermore, function adjustments in Experiment3 have allowed improvements in ranking
H. Kariya and Y. Kiyoki / Metadata Extraction and Retrieval Methods for Taste-Impressions
375
rank
target data ID
correlation
rank
target data ID
correlation
correlation
target data
support rate
correlation
target data
support rate
[1]
kyuuritowakameno-sunomono:
14.32
[1]
beer data1
26.54
1
20.97
chinjaoro-su:
73%
14
14.08
saba-no-misoni:
39%
[2]
ma-bo-toufu:
12.93
[2]
beer data10
25.04
2
20.93
ma-bo-toufu:
57%
15
14.05
niku-jaga:
[3]
chinjaoro-su:
12.75
[3]
beer data11
24.44
3
20.01
yaki-gyouza:
55%
16
13.35
karei-no-nituke:
18.38
butaniku-noshouga-yaki:
59%
17
[4]
beer data15
[5]
chirashi-sushi:
12.24
[5]
beer data16
23.14
[6]
butaniku-noshouga-yaki:
11.63
[6]
beer data17
22.52
[4]
yaki-gyouza:
12.29
24.09
[7]
ajino-shioyaki:
11.52
[7]
beer data2
21.77
[8]
mi-toso-supasuta:
11.13
[8]
beer data3
20.77
13.30
ro-ru-kyabetsu:
5
17.77
ebi-furai:
61%
18
12.44
kyuuritowakameno-sunomono:
6
4
16.74
ika-no-bata-sote-:
52%
19
11.87
kabochano-nimono:
7
16.61
kaki-furai:
64%
20
11.74
8
16.47
karubona-rapasuta:
61%
21
11.70
chirashi-sushi:
9
16.46
toriniku-no-teriyaki:
39%
22
10.37
ingen-no-gomaae:
potetosarada:
[9]
bi-fu-shichu-:
11.04
[9]
beer data4
20.42
10
15.50
buri-no-teriyaki:
50%
23
10.21
[10]
buri-no-teriyaki:
10.99
[10]
beer data 8
19.55
11
15.31
mi-toso-supasuta:
55%
24
9.12
chawan-mushi:
[11]
toriniku-no-teriyaki:
10.88
[11]
beerdata12
19.41
12
15.16
ajino-shioyaki:
25
8.08
houren-sou-no-ohitashi:
13
15.11
bi-fu-shichu-:
Results forJAPANESE FOOD Scope
Results for the BEER Scope
furofuki-daikon:
57%
Figure 14: Retrieval Results for different Taste Scopes Figure 15: Retrieval results (“Rich” for Japanese foods)
performance with the application of implemented modules for Beer Scope, hence verified the plagability of modules in our experimental system. 7.1 Experiment 1: Feasibility evaluation for anonymous Taste Scopes • Evaluation Method: Experiment1 is for applicability evaluation of our method to several Taste Scopes. For experimental studies, we submit a query with both Taste Scopes “Japanese food” and “beer” and selected Taste-impression “fresh”. • Experimental Results and Analysis: In this experiment, we have observed our experimental system have 1) selected the appropriate target data for each Taste Scope concurrently, and 2) ranked the target data with reasonable accuracy. For instance, “Kyuri-to-wakame-no-sunomono (Vinegared cucumber and brown seaweed)” and “Chirashizushi” (Vinegared rice arranged with various kinds of sliced raw fish) ranks in 1 and 5 in Japanese food scope. This result has suggested that the taste Scope-based metadata extraction method for impression retrieval is promising. Detailed performance evaluation for each scope is shown in following 2 experiments.
7.2 Experiment 2: Performance Evaluation for Japanese Food Scope • Evaluation Method: Experiment2 is for performance evaluation in Japanese food scope. As experimental objects, we have created the 25 virtual sensor outputs by questionnaire of 50 test subjects. In this experiment, we have committed the query with impression word “rich” Taste Scope “Japanese food”. Results are shown in Table 15. As impression words for query, we have implemented three impression expressions: “maroyaka (mellow or mild in Japanese)”, “sappari (fresh)” and “kotteri (heavy or rich)”. We have selected these impression words because frequent uses of these impressions have observed in our questionnaire. As target data, we have selected the 25 food items based on the support rate of keywords. • Experimental Results and Analysis: Target data with more than 39% support rate for impression word “rich” are indicated by boldface. Overall comparison of support rate and actual ranking results presents the reasonable correlation of our impression-based retrieval and keyword. Impression word “rich” define the attribute “fat” as the most significant feature for flavor definition (2.73) and then “salinity” (2.17). These attribute values are the 2nd and 3rd
376
H. Kariya and Y. Kiyoki / Metadata Extraction and Retrieval Methods for Taste-Impressions
Figure 16: Factor Analysis results (correct answers)
Figure 17: Function Combinations
greatest value among overall attribute values for feature values of impression words. Other values are relatively large as well, compared with other metadata of impression words. These facts demonstrate that the overall impression for the query “rich” for Japanese food is thick especially fat and salinity perspective. Since retrieving the target data with large attribute values globally is easier in inner product query processing, support rate for this query example is very promising. 7.3 Experiment 3: Performance Evaluation for Beer Scope • Evaluation Method: In experiment3, we have conducted performance evaluation for Taste Scope beer on the basis of extensive survey, evaluating the retrieval results of our method with prepared collect answers. As experimental objects, we have applied the 28 sensor outputs for the beer scope. The criterion for its performance evaluation is whether our experimental system highly ranks the correct answers as the ranking results. For preparing correct answers, beer data for each Taste-impression have defined by the marketing analysis survey for beer data. These answers have been generated by factor analysis for 178 test subjects, 30 beer data and 32 Taste-impression words 2 . We have sorted beer data according to this factor rating values in descending order and selected top 4 as the correct answers for each impression as shown in Table 16. • Experimental Results and Analysis: Results are shown in figure 7. Among 28 real beer data, we have observed correct answers are ranked in 1, 2, 6 and 11, demonstrating 50% with recall ratio in top5 and 75% for top10 target data. These experimental results have present feasibility in our metadata generation method. We will discuss the effect of module adoption to our method in the next experiments. 7.4 Experiment 4: Functions Adjustment • Evaluation Method Experiment3 is for evaluating plagability of modules in our experimental system. We have implemented and applied several function of the standardization module Pz and Taste Impression Metadata Generation Module Ge to metadata generation for sensor data in the beer scope.
2 The marketing data is offered by Masayuki Goto, Faculty of Environmental Information, Musashi Institute of Technology.
H. Kariya and Y. Kiyoki / Metadata Extraction and Retrieval Methods for Taste-Impressions
Query : Beer, Rich 㫉㪸㫅㫂 㪲㪈㪴 㪲㪉㪴 㪲㪊㪴 㪲㪋㪴 㪲㪌㪴 㪲㪍㪴 㪲㪎㪴 㪲㪏㪴 㪲㪐㪴 㪲㪈㪇㪴 㪲㪈㪈㪴 㪲㪈㪉㪴 㪲㪈㪊㪴 㪲㪈㪋㪴 㪲㪈㪌㪴 㪲㪈㪍㪴 㪲㪈㪎㪴 㪲㪈㪏㪴 㪲㪈㪐㪴 㪲㪉㪇㪴 㪲㪉㪈㪴 㪲㪉㪉㪴 㪲㪉㪊㪴 㪲㪉㪋㪴 㪲㪉㪌㪴 㪲㪉㪍㪴
㫋㪸㫉㪾㪼㫋㩷㪻㪸㫋㪸㩷㪠㪛 㪺㫆㫉㫉㪼㪺㫋㩷㪸㫅㫊㫎㪼㫉㩷㪋 㪺㫆㫉㫉㪼㪺㫋㩷㪸㫅㫊㫎㪼㫉㩷㪉
㪹㪼㪼㫉㩷㪻㪸㫋㪸㪈 㪹㪼㪼㫉㩷㪻㪸㫋㪸㪉 㪹㪼㪼㫉㩷㪻㪸㫋㪸㪈㪈 㪺㫆㫉㫉㪼㪺㫋㩷㪸㫅㫊㫎㪼㫉㩷㪈 㪹㪼㪼㫉㩷㪻㪸㫋㪸㪋 㪹㪼㪼㫉㩷㪻㪸㫋㪸㪌 㪹㪼㪼㫉㩷㪻㪸㫋㪸㪈㪌 㪹㪼㪼㫉㩷㪻㪸㫋㪸㪍 㪺㫆㫉㫉㪼㪺㫋㩷㪸㫅㫊㫎㪼㫉㩷㪊 㪹㪼㪼㫉㩷㪻㪸㫋㪸㪐 㪹㪼㪼㫉㩷㪻㪸㫋㪸㪈㪇 㪹㪼㪼㫉㩷㪻㪸㫋㪸㪈㪍 㪹㪼㪼㫉㩷㪻㪸㫋㪸㪈㪎 㪹㪼㪼㫉㩷㪻㪸㫋㪸㪉㪇 㪹㪼㪼㫉㩷㪻㪸㫋㪸㪈㪋 㪹㪼㪼㫉㩷㪻㪸㫋㪸㪈㪐 㪹㪼㪼㫉㩷㪻㪸㫋㪸㪈㪏 㪹㪼㪼㫉㩷㪻㪸㫋㪸㪉㪈 㪹㪼㪼㫉㩷㪻㪸㫋㪸㪏 㪹㪼㪼㫉㩷㪻㪸㫋㪸㪉㪉 㪹㪼㪼㫉㩷㪻㪸㫋㪸㪈㪉 㪹㪼㪼㫉㩷㪻㪸㫋㪸㪈㪊 㪹㪼㪼㫉㩷㪻㪸㫋㪸㪎 㪹㪼㪼㫉㩷㪻㪸㫋㪸㪊
㪺㫆㫉㫉㪼㫃㪸㫋㫀㫆㫅 㪉㪍㪅㪌㪋 㪉㪌㪅㪇㪋 㪉㪋㪅㪋㪋 㪉㪋㪅㪇㪐 㪉㪊㪅㪈㪋 㪉㪉㪅㪌㪉 㪉㪈㪅㪎㪎 㪉㪇㪅㪎㪎 㪉㪇㪅㪋㪉 㪈㪐㪅㪌㪌 㪈㪐㪅㪋㪈 㪈㪏㪅㪌㪏 㪈㪎㪅㪊㪋 㪈㪎㪅㪉 㪈㪍㪅㪉㪎 㪈㪍㪅㪉㪍 㪈㪋㪅㪌㪋 㪈㪋㪅㪊㪈 㪈㪊㪅㪎㪋 㪈㪊㪅㪊㪊 㪈㪊㪅㪇㪉 㪈㪉㪅㪍㪉 㪈㪉㪅㪌㪌 㪈㪉㪅㪉㪐 㪈㪉㪅㪈㪐 㪈㪈㪅㪎㪋
㫉㪸㫅㫂 㪲㪈㪴 㪲㪉㪴 㪲㪊㪴 㪲㪋㪴 㪲㪌㪴 㪲㪍㪴 㪲㪎㪴 㪲㪏㪴 㪲㪐㪴 㪲㪈㪇㪴 㪲㪈㪈㪴 㪲㪈㪉㪴 㪲㪈㪊㪴 㪲㪈㪋㪴 㪲㪈㪌㪴 㪲㪈㪍㪴 㪲㪈㪎㪴 㪲㪈㪏㪴 㪲㪈㪐㪴 㪲㪉㪇㪴 㪲㪉㪈㪴 㪲㪉㪉㪴 㪲㪉㪊㪴 㪲㪉㪋㪴 㪲㪉㪌㪴 㪲㪉㪍㪴
㫋㪸㫉㪾㪼㫋㩷㪻㪸㫋㪸㩷㪠㪛 㪺㫆㫉㫉㪼㪺㫋㩷㪸㫅㫊㫎㪼㫉㩷㪋 㪺㫆㫉㫉㪼㪺㫋㩷㪸㫅㫊㫎㪼㫉㩷㪈 㪹㪼㪼㫉㩷㪻㪸㫋㪸㪍 㪹㪼㪼㫉㩷㪻㪸㫋㪸㪈 㪺㫆㫉㫉㪼㪺㫋㩷㪸㫅㫊㫎㪼㫉㩷㪉 㪺㫆㫉㫉㪼㪺㫋㩷㪸㫅㫊㫎㪼㫉㩷㪊 㪹㪼㪼㫉㩷㪻㪸㫋㪸㪈㪈 㪹㪼㪼㫉㩷㪻㪸㫋㪸㪋 㪹㪼㪼㫉㩷㪻㪸㫋㪸㪈㪌 㪹㪼㪼㫉㩷㪻㪸㫋㪸㪈㪍 㪹㪼㪼㫉㩷㪻㪸㫋㪸㪌 㪹㪼㪼㫉㩷㪻㪸㫋㪸㪐 㪹㪼㪼㫉㩷㪻㪸㫋㪸㪉㪇 㪹㪼㪼㫉㩷㪻㪸㫋㪸㪈㪇 㪹㪼㪼㫉㩷㪻㪸㫋㪸㪉 㪹㪼㪼㫉㩷㪻㪸㫋㪸㪈㪐 㪹㪼㪼㫉㩷㪻㪸㫋㪸㪎 㪹㪼㪼㫉㩷㪻㪸㫋㪸㪈㪊 㪹㪼㪼㫉㩷㪻㪸㫋㪸㪈㪎 㪹㪼㪼㫉㩷㪻㪸㫋㪸㪉㪈 㪹㪼㪼㫉㩷㪻㪸㫋㪸㪈㪉 㪹㪼㪼㫉㩷㪻㪸㫋㪸㪈㪋 㪹㪼㪼㫉㩷㪻㪸㫋㪸㪉㪉 㪹㪼㪼㫉㩷㪻㪸㫋㪸㪈㪏 㪹㪼㪼㫉㩷㪻㪸㫋㪸㪊 㪹㪼㪼㫉㩷㪻㪸㫋㪸㪏
Exp.A (original)
㪺㫆㫉㫉㪼㫃㪸㫋㫀㫆㫅 㪉㪅㪈㪋 㪉㪅㪇㪇 㪈㪅㪊㪇 㪈㪅㪉㪐 㪈㪅㪉㪏 㪈㪅㪇㪐 㪈㪅㪇㪈 㪈㪅㪇㪇 㪇㪅㪐㪐 㪇㪅㪎㪎 㪇㪅㪍㪈 㪄㪇㪅㪊㪊 㪄㪇㪅㪊㪐 㪄㪇㪅㪌㪍 㪄㪇㪅㪎㪊 㪄㪇㪅㪎㪍 㪄㪈㪅㪌㪍 㪄㪈㪅㪎㪉 㪄㪈㪅㪐㪐 㪄㪉㪅㪊㪋 㪄㪉㪅㪌㪍 㪄㪉㪅㪍㪇 㪄㪊㪅㪇㪏 㪄㪊㪅㪈㪋 㪄㪊㪅㪏㪊 㪄㪋㪅㪈㪏
㫉㪸㫅㫂 㪲㪈㪴 㪲㪉㪴 㪲㪊㪴 㪲㪋㪴 㪲㪌㪴 㪲㪍㪴 㪲㪎㪴 㪲㪏㪴 㪲㪐㪴 㪲㪈㪇㪴 㪲㪈㪈㪴 㪲㪈㪉㪴 㪲㪈㪊㪴 㪲㪈㪋㪴 㪲㪈㪌㪴 㪲㪈㪍㪴 㪲㪈㪎㪴 㪲㪈㪏㪴 㪲㪈㪐㪴 㪲㪉㪇㪴 㪲㪉㪈㪴 㪲㪉㪉㪴 㪲㪉㪊㪴 㪲㪉㪋㪴 㪲㪉㪌㪴 㪲㪉㪍㪴
㫋㪸㫉㪾㪼㫋㩷㪻㪸㫋㪸㩷㪠㪛 㪺㫆㫉㫉㪼㪺㫋㩷㪸㫅㫊㫎㪼㫉㩷㪋 㪺㫆㫉㫉㪼㪺㫋㩷㪸㫅㫊㫎㪼㫉㩷㪈 㪹㪼㪼㫉㩷㪻㪸㫋㪸㪍 㪺㫆㫉㫉㪼㪺㫋㩷㪸㫅㫊㫎㪼㫉㩷㪊 㪹㪼㪼㫉㩷㪻㪸㫋㪸㪈 㪺㫆㫉㫉㪼㪺㫋㩷㪸㫅㫊㫎㪼㫉㩷㪉 㪹㪼㪼㫉㩷㪻㪸㫋㪸㪈㪌 㪹㪼㪼㫉㩷㪻㪸㫋㪸㪋 㪹㪼㪼㫉㩷㪻㪸㫋㪸㪈㪈 㪹㪼㪼㫉㩷㪻㪸㫋㪸㪈㪍 㪹㪼㪼㫉㩷㪻㪸㫋㪸㪌 㪹㪼㪼㫉㩷㪻㪸㫋㪸㪐 㪹㪼㪼㫉㩷㪻㪸㫋㪸㪉㪇 㪹㪼㪼㫉㩷㪻㪸㫋㪸㪈㪇 㪹㪼㪼㫉㩷㪻㪸㫋㪸㪈㪐 㪹㪼㪼㫉㩷㪻㪸㫋㪸㪉 㪹㪼㪼㫉㩷㪻㪸㫋㪸㪎 㪹㪼㪼㫉㩷㪻㪸㫋㪸㪈㪊 㪹㪼㪼㫉㩷㪻㪸㫋㪸㪈㪎 㪹㪼㪼㫉㩷㪻㪸㫋㪸㪉㪈 㪹㪼㪼㫉㩷㪻㪸㫋㪸㪈㪉 㪹㪼㪼㫉㩷㪻㪸㫋㪸㪈㪋 㪹㪼㪼㫉㩷㪻㪸㫋㪸㪉㪉 㪹㪼㪼㫉㩷㪻㪸㫋㪸㪈㪏 㪹㪼㪼㫉㩷㪻㪸㫋㪸㪊 㪹㪼㪼㫉㩷㪻㪸㫋㪸㪏
Exp.B
377
㪺㫆㫉㫉㪼㫃㪸㫋㫀㫆㫅 㪎㪉㪅㪌 㪎㪇㪅㪐 㪍㪍㪅㪋 㪍㪋㪅㪐 㪍㪊㪅㪐 㪍㪊㪅㪏 㪍㪊㪅㪏 㪍㪊㪅㪍 㪍㪈㪅㪏 㪌㪏㪅㪏 㪌㪎㪅㪉 㪌㪇㪅㪋 㪌㪇 㪋㪏㪅㪈 㪋㪋㪅㪈 㪋㪊㪅㪐 㪊㪏㪅㪊 㪊㪊㪅㪏 㪊㪈㪅㪏 㪊㪇㪅㪋 㪉㪍㪅㪉 㪉㪌㪅㪍 㪉㪊㪅㪉 㪉㪇㪅㪉 㪈㪊㪅㪊 㪐㪅㪐
Exp.C
Figure 18: Function Adjustments Results for Beers
Figure 19: Recall Ratio Improvements (Exp.4)
We compared these retrieval ranking results of Experiment A (only with fundamental operation, very close to original data of sensor outputs), Experiment B (threshold adjustment with G3 ) as and Experiment3 (Experiment A with weighting) as shown in Table 17. Here, we have committed the query with impression word “rich” applying these 3 optimization patterns. • Experimental Results Results are promising as shown in figure 18. Adjusting the feature values with these functions have accepted better ranking results, compared with those without standardization function. To be more specific, adoption of the beer-specific Taste Scope intelligence in these modules (Experiment B and C) achieves 40% recall ratio improvement for in top5, 25% for top10 target data compared with Experiment A, thus demonstrating the promise of the approach (Figure 19 ). These results suggest that function adjustment in our metadata generation method is effective for optimizing feature values for target Taste Scopes. 8 Conclusion and Future Work In this paper, we have presented Metadata Extraction and Retrieval Methods for Taste-impressions with bio-sensing Technology. Features of our metadata extraction method are 1) the metadata extraction method which transforms sensor outputs into meaningful Taste-impression metadata automatically and 2) the definition of the Taste Scope which utilizes anonymous meanings of each Taste-impression. The application of our methods to media data of the beer and Japanese food scope has been shown, and the feasibility of our system has been examined by several experimental studies. We are currently developing new Taste Scopes which deal with the view points of user groups in order to manipulate diversified sensitivity of people’s tongue, such as of the elderly and the young. These functions would be added to our proposed method and allow the further flexibility for extracting Taste-impression. Eventually, we are hoping to realize a sensor based metadata extraction method by several bio-sensing technologies such as odor sensor in order to improve the quality of metadata from other five senses aspects. Acknowledgements We would thank to Shuichi Kurabayashi and Dr. Naofumi Yoshida of Graduate School of Media and Governance, Keio University for valuable discussions and helpful comments on this study. I also would like to express my gratitude to researchers of Taste Sensor, Dr. Hidekazu
378
H. Kariya and Y. Kiyoki / Metadata Extraction and Retrieval Methods for Taste-Impressions
Ikezaki of Intelligentsensor technologies,Inc. and Prof. Kiyoshi Toko of Graduate School of Information Science and Electrical Engineering, Kyushu University for valuable comments for implementing the experimental system. References [1] Albrecht Schmidt, Michael Beigl, Hans-W. Gellersen,“There is more to Context than Location – Environment Sensing Technologies for Adaptive Mobile User Interfaces”, Proceedings of International Workshop on Interactive Applications of Mobile Computing (IMC98) [2] BMW. The BMW navigation system. BMW compass. http://www.bmw.com/compass/htdocs/BMWe/backissue/FORSCH2E.shtml, 1998. [3] B. Dasarathy, “Sensor Fusion Potential Exploitation - Innvative Architectures and Illustrative Applications”, Proc. of the IEEE, Vol. 85, pp. 24-38, Jan. 1997. [4] Beadle, P., Harper, B., Maguire, G.Q. and Judge, J.Location Aware Mobile Computing. Proceedings of IEEE International Conference on Telecommunications, Melbourne, Australia, April 1997. [5] Charles Zuker, “A Matter of Taste :Candidates for Taste Receptors Identified” Howard Hughes Medical Institute Bulletin, 2003 [6] H.Ikezaki, Y.Kobayashi, R.Toukubo, Y.Naito, A.Taniguchi, and K.Toko : “Techniques to Control Sensitivity and Selectivity of Multichannel Taste Sensor Using Lipid Membranes” Digest Tech. Papers Transducers ’99, pp.1634-1637, June, 1999 [7] Ijichi, A. and Kiyoki, Y.:“ A Kansei Metadata Generation Method for Music Data Dealing with Dramatic Interpretation ”, Information Modeling and Knowledge Bases, Vol.XVI, IOS Press, [8] K.Toko,“Biomimetic sensor technology”, Cambridge University Press, 2000 [9] Nobuko Miura, Shuichi Kurabayashi, and Yasushi Kiyoki: An Automatic Extraction Method of TimeSeries Impression-Metadata for Color Information of Video Streams. International Special Workshop on Databases For Next Generation Researchers (SWOD2005) in conjunction with ICDE2005, 2005, 54-57. [10] Pramod K. Varshney, “Multisensor Data Fusion”, Lecture Notes in Computer Science (Springer-Verlag Heidelberg), Vol.1821, 2000 [11] R. Antony, “Database Support to Data Fusion Automation”, Proc. of the IEEE, Vol. 85, pp. 39-53, Jan. 1997. [12] R. Cappelli, “SFinGe: an Approach to Synthetic Fingerprint Generation”, in proceedings International Workshop on Biometric Technologies (BT2004), Calgary, Canada, pp.147-154, June 2004. [13] Snider,J.G. and Osgood, C.E. : ”Semantic Differential Technique-A Sourcebook”, Aldine Pub. Company, 1969 [14] Tanizawa, K. and Uehara, K. : “Automatic Detection of the Semantic Structure from Video by Using Ngram” [15] Y. Kiyoki, T. Kitagawa, T. Hayama, “A Metadatabase System for Semantic Image Search by a Mathematical Model of Meaning”, Multimedia Data Management using metadata to integrate and apply digital media, McGrawHill, Amit Sheth and Wolfgang Klas(editors), Chapter 7, 1998. [16] Nagao, K., Rikimoto, J. Agent Augmented Reality: A software Agent meets the real world. Proceeding of the 2nd Conference on Multiagent systems (ICMAS-96), Dec 1996.
Information Modelling and Knowledge Bases XIX H. Jaakkola et al. (Eds.) IOS Press, 2008 © 2008 The authors and IOS Press. All rights reserved.
379
!! " #$%%&%'( Position paper In this paper we propose a cooperation modeling approach which aims to the alignment of system development with the organizational change where the system will operate. It consists in a semantic enrichment of cooperation documentation so that the intertwining interactions between organization, human and system views could be represented explicitly into the system development process. The proposed ontological framework plays crucial roles as a communication, learning and design artifact for different stakeholders.
1. Cooperation capturing and modeling Cooperation modeling is very decisive for the system development process when the application domain is characterized with complex cooperation. It is not only necessary to identify and to understand the actual work practices but also to capture and predict the changes the future system will initiate so that the system is kept adaptable to the permanent changing environment. These changes can be explicitly known such as those of technological nature, or not as easily identifiable such as those from social nature. The literature witnesses the emergence of manifold models technology supporting the cooperation such as CSCW (Groupware and Workflows), Business process re-engineering, etc. On one side, the difference of the origins of the approaches on which the models are based (theories of situated action, Communities of practice, Distributed cognition, studies on coordination mechanisms and “articulation work”, etc.) leads to the fact that there is no consensus regarding the set of concepts and abstraction levels underlying the cooperation modeling. On the other side, changing approaches are mainly from organizational point of view (organization work-oriented, system-oriented, collaborators-oriented, process-oriented approaches) and dealing thus with the nature of the work practices, for instance, if they are structurally opened or closed to be changed or not [10], after the embedding of the cooperative system.
!"# !$# !%# !$$# they do not consider explicitly the two levels of requirements dealing with the nature of the cooperative work as well as what will be changed and the way to change it. &
'
(
)* &
( (
(
( *
+ ( *
380
B. Lahouaria / An Ontological Framework for Modeling Complex Cooperation Contexts
The alignment of organization, human and system views for cooperation modeling Participatory and evolutionary system development approaches such as STEPS [12] are very convenient for cooperation support because of the fact that considering organization, human and system at the same level is an old tradition in such approaches. Modeling complex cooperation characterizing work practices in organizations requires unfortunately, not only the alignment of these analytical dimensions but furthermore that they should be explicitly integrated into the system development and embedding processes. This seems to be a very difficult task for the methods grounded on participation and evolution principles where learning processes are based on project-oriented software artifacts (such as Scenarios, Glossary, Prototypes, etc.). The project itself is unique and limited in the time so that the focus is finally only on the software as a product. We claim that a practical solution for the alignment of system development with the organizational change should assure that: • the participation of the users to the development process means that continuously new unknown participants with different interests could be introduced in order to deal with ( !,#
• the evolution of the system goes beyond the project’s context so that t
-
) * ( ! (
)*
An ontology-based cooperative work representation . ' (
/ ***0 ( *
1 23
5+ ! * * * * * * *
! *! - !6#*
; * ! / 7 0 !8# (
* .
9 ( *
& :
• ;
(
(
• Concretizing the evolution by creating an organizational memory information system, in order to have a common learning artifact.
Ontological framework for cooperative processes The whole environment (organizational, human and system) where the system is embedded does not exist physically but is represented by means of the cooperation ontologies represented at the cognitive context (see Fig. 1). . -
) / 0 ( -
) -
*
B. Lahouaria / An Ontological Framework for Modeling Complex Cooperation Contexts
381
.9 .9
; ;
)
'
; ;
; ;
; )
+
+
+
/(
=0
+
(
+ < )
+
+
)
Fig 1. An Ontological framework as a communication and learning artefact
We propose OFCP (an Ontological Framework for Cooperative Processes) consisting in one cooperation top-level ontology (see Fig. 2) and three cooperation foundational ontologies (see Fig. 3) underlying the different cooperation’s understandings from organization, human and system views. • Top-Level ontology Active entities, actions and passive entities which seem to be in agreement with the principles of FRISCO framework [8] supporting a constructivist view of system development approach are the basic constituents of the top-level ontology. An active entity is any kind of entity which is able to carry-out actions (including nonhuman entities) such as doctors, teams, etc…A passive entity is a special thing which is involved in a post-state of an action. They are created, modified, or only accessed for information purposes such as a document. - ( (
( *
A typical cooperation scenario is when an active entity carries out an action in order to change (consult) a passive entity. The action could be further-carried out by another active entity, the passive entity could be retransmitted to another active entity and the active entity could communicate with another active entity.
9
9 ? 9(
9
9
>
9
99>
9
Fig. 2 Top-level cooperation ontology
(
* •
; * ! !
382
B. Lahouaria / An Ontological Framework for Modeling Complex Cooperation Contexts
*
. (
* . (
( / 0 / 0 /0
!@#*
/(0 *
) *
( *
* .
*
+
( *
!
!
!
"
.
(A
# 79 + 9 ;9
79 + 9
79
;9 + 9(A
;9
Fig. 3 Organizational, human and system cooperation foundational ontologies
Table 1. Cooperation modeling levels in the system development process E 9 /9 0
Analyse
E 9 / 0 / 0
Operationalization Extensional level (instances)
The horizontal representation of cooperation views through the foundational ontologies (see Fig. 3) is useful for guiding the developers team to take into account different stakeholders with different interest, understandings and terminology about the cooperation, whereas the vertical representation of cooperation levels (see table 1) is useful for guiding them in their task of analyzing and generating contextual cooperative processes metamodels which should be adequate to their application domain in hand. 5. Application of OFCP to a hospital research project ?;= ( =BB9%DDD / ( 0
' 9* . '
( ( 9
B. Lahouaria / An Ontological Framework for Modeling Complex Cooperation Contexts
383
* =BB9%DDD FE
( ( )
FE
* & ?;= A (
!G# )
A '
*
*
?;= ( )
(
*
? =BB9
9
) * ( 23
( *
?
! ?;= ( )( for cooperation analysis process. Indeed, a cooperative process could be characterized in terms of network of dependencies among entities annoted through the set of concepts in OFCP. The process of analysis could begin or alternate from any type of entity (task-oriented, object-oriented, actor-oriented, resource-oriented analysis, etc.). References [1] Y. Engeström. Developmental work research: Reconstructing expertise through expansive learning. In: Nurminen, M. I., Järvinen, P. & Weir, G. (eds.), Conference on Human jobs and computer interfaces, Tampere, Finnland, June 26-28, 1991,s.124-143. [2] C. Floyd, Y. Dittrich, R. Klischewski, (Eds.). Social Thinking - Software Practice. Relating Software Development, Work and Organizational Change, Dagstuhl-Report Nr. 99361. Wadern: IBFI, 1999 [3] I. Wetzel. Information Systems Development with Anticipation of Change Focusing on Professional Bureaucraties. In proc. Of Hawai’, International Conference on System Sciences, HICCS-34, Maui, January 2000. [4] J. Ziegler. Modeling cooperative work processes- A multiple perspectives framework. Int. Journal of human-computer interaction, 14(2), 139-157, 2002 [5] C. Floyd. Software development as reality construction. In: Floyd, C. et al. (eds) : Software Development and Reality Construction. Springer Verlag, Berlin 1992 [6] D. Hensel: Relating Ontology Languages and Web Standards. In: J.Ebert, U. Frank (Hrsg.): Modelle und Modellierungssprachen in Informatik und Wirtschaftsinformatik. Proc. „Modellierung 2000“, FöllbachVerlag, Koblenz 2000, pp. 111-128. [7] N. Guarino. Foundational ontologies for humanities: the Role of Language and Cognition, in first int. Workshop “Ontology Based modeling in Humanities”, University of Hamburg, 7-8 April 2006. [8] E. Falkenberg, W. Hesse, P. Lindgreen, B.E. Nilsson, J.L.H. Oei, C. Rolland, R.K. Stamper, F.J.M. Van Assche, A.A. Verrijn-Stuart, K. Voss: FRISCO - A Framework of Information System Concepts - The FRISCO Report. IFIP WG 8.1 Task Group FRISCO. Web version: ftp://ftp.leidenuniv.nl/pub/rul/frifull.zip (1998) [9] WJ. Orlikowski & D. Robey. Information Technology and the Structuring of Organizations. Information Systems Research. Vol 2(2): 143-169.1991. [10] E* H : ; : ; + < ;+;& +
;E 1 ("9$% %DDG + ( ? + !$$# < E 7* * .9? I +: . ; ;
; -. *,$ 1 $% $""J* !$%# ;* ? ?*9E* 7* +: +.K=+ L : ;* 7 M* E9 < /*0: K+K;- J" + K + K ; 1;+ GJ8 * ,J96,L
+ $"J"
384
Information Modelling and Knowledge Bases XIX H. Jaakkola et al. (Eds.) IOS Press, 2008 © 2008 The authors and IOS Press. All rights reserved.
Information Modelling and Knowledge Bases for Interoperability Solution in Security Area Prof. Ladislav BUŘITA, Vojtěch ONDRYHAL University of Defense, Communication and Information Systems Department Kounicova 65, 612 00 Brno, Czech Republic
Abstract. The article presents an example of information interoperability solution in a security field. NEC, transformation concept based on wide ICT utilisation, forms a framework for such endeavour. Results of information modelling developed in MIP group, including IEDM and development method, are introduced. It looks however, that IEDM reached its limits. New forward-looking approaches, like domain modeling or knowledge technologies, will be applied in the near future. The aim of our project is a verification of knowledge approach possibilities based on ITM software.
1. Introduction Selected approaches of information modelling in security based projects are presented in the article. MIP research group results, authors’ experience and future model advancements are described. New means getting over model limitations like model simplification (domain approach) or new model strategy (knowledge technology) are examined. Our project deals with ITM software possibilities. All activities supporting information interoperability are covered by NEC concept.
2. Network Enabled Capability understanding To understand background of the project it is necessary to became familiar with the main strategy undertaken by the NATO in the communication and information areas. Information integration and sharing has become one of the key pillars of NATO Network Enabled Capability (NNEC) initiative. This capability involves the seamless linking together of sensors, decision makers, and weapon systems, as well as multinational military, appropriately linked with governmental, and non-governmental agencies in a collaborative, planning, assessment and execution environment. The NEC must provide for the timely exchange of secure information, utilising communication networks which are interconnected, interoperable and robust, and which will support the timely collection, fusion, analysis and sharing of information [4]. Many aspects are involved in NEC initiative, like operational needs, people, logistics etc., but from technology point of view required Networking and Information
L. Buˇrita and V. Ondryhal / Information Modelling and Knowledge Bases
385
Infrastructure (NII) is important - it is clear that the Alliance will only be able to achieve its operational ambitions if future force structures are well supported by flexible, adaptable, highly interconnected, communication networks and information systems [4]. Four maturity levels (deconflict, coordinate, collaborate and coherent) and identified technologies are displayed on the Figure 1.
Figure 1. Maturity levels and technology trends for NEC The Information and Integration component of the NII is characterized by the use of Service Oriented Architectures (SOA) to expose software functions as consumable services that can be discovered and invoked across the network. The use of the SOA approach requires that we adopt a common Net-Centric Data strategy (like MIP - described later in the article) to ensure that we make information visible, accessible, understandable and interoperable with other sources of information. One of the keys to the widespread use of XML-enabled technologies is meta-data standardization. Military specific vocabularies require the participation of military experts, not only to define the core vocabularies for various COIs (Communities of Interests) but to also define the semantic relationships that exist between the words themselves (i.e. Ontologies). This standardization activity is key to information interoperability at all levels of maturity, key to future concepts of information security and key to the use of machine based reasoning / agent based technology that will provide the foundation for meeting the longer term objectives for the NII in general. 3. Information Modelling in Multilateral Interoperability Programme The Multilateral Interoperability Programme (MIP) aims to deliver an assured capability for interoperability of information to support joint / combined operations. The aim of the MIP is to achieve international interoperability of Command and Control Information Systems (C2IS) in order to support multinational (including NATO), combined and joint operations and the advancement of digitisation in the international arena [5].
386
L. Buˇrita and V. Ondryhal / Information Modelling and Knowledge Bases
3.1 The Information Exchange Data Model The MIP solution enables information exchange between co-operating but distinct national C2 systems. The core of a MIP solution is the Information Exchange Data Model (IEDM). It is a product of the analysis of a wide spectrum of Allied information exchange requirements. It models the information that combined joint component commanders need to exchange. The MIP solution enables C2IS to C2IS information exchange and allows users to decide what information is exchanged, to whom it flows, and when. The MIP contribution is to facilitate the timely flow of accurate and relevant information, using the Information Exchange Mechanisms specified by MIP, between the different national C2IS. MIP will, therefore, is one of the factors contributing to the realization of NEC for the commanders within a combined joint force [5]. Joint Consultation, Command and Control Information Exchange Data Model (JC3IEDM) is the last version of IEDM and is intended to represent the core of the data identified for exchange across multiple functional areas and multiple views of the requirements. The purpose of the JC3IEDM is to provide the following [5]: • A description of the common data that contains the relevant data, abstracted in a well structured normalised way that unambiguously reflects their semantic meaning. • A basic document that nations can use to present and validate functional data model views with their own specialist organisations. • A specification of the physical schema required for database implementation. The overall goal is to specify the minimum set of data that needs to be exchanged in coalition or multinational operations. Each nation or agency or community of interest is free to expand its own data dictionary to accommodate its additional information exchange requirements with the understanding that the added specifications will be valid only for the participating nation, agency or community of interest. 3.2 The Information Modelling Concept Basic concept in data specification is an entity, properties or characteristics of an entity are referred to as attributes. This edition of the model contains nearly 300 entities, and the entire structure is generated from 15 independent entities. The content of the model in terms of attributes and sets of enumerated values represents the semantics of a given functional domain [5]. The IEDM is built for many years (more than 20) and it seems that the initial concept considers all possibilities. Model is very large, complex, confused and only few analytics are well known with the construction. There is problem by the model modifying and enlarging; some changes are compromise between nations and to find harmony and agreement is still more and more difficult. The serious problem is the backward compatibility. The international MIP community is gradually pressed to find a new future concept of C2IS interoperability. Some NATO exploratory teams try to find useful solution. There are in account Domain View, Knowledge Bases, Ontologies, Semantic Web, Intelligent Agents, etc. 3.3 Information Interoperability Domains The concept of “Information Interoperability Domains” originates from a simple ‘common sense’ idea: whenever a problem becomes too complex, split it up in relatively autonomous parts, which can be independently defined, but still fit, in the overall solution. The set of systems that interact by using the same exchange language is called an
L. Buˇrita and V. Ondryhal / Information Modelling and Knowledge Bases
387
information interoperability domain. A system is said to be part of a domain when it is able to interact with other systems by making use of the domain’s exchange language. 4. Project of Information System in the State Security 4.1 Project background Project of Information System in the State Security was started at the end oh the year 2006. Project is a part of the research program „Development, integration, administration and security of CIS in NATO environment“ and is prepared in cooperation with Institute for Strategic Studies (ISS). Project is based on application of commerce software Intelligent Topic Manager (ITM) of the company Mondeca (France) for the intelligent data organisation and retrieval. ITM is a unique tool that federates and organizes information and knowledge in a businessspecific reference repository for more effective navigation and searches. ITM functionality (see Figure 3) includes: • Ontology management, thesaurus, taxonomies, knowledge bases. • Navigation in a business-related representation. • Multi-criteria searches in bases and content. • Automatic content annotation and knowledge acquisition. • Collaborative work to capitalize on knowledge. • Reuse of content for composition, publishing and distribution. 4.2 Project goals and specification Current state of the information processing in the ISS could be specified as decentralize and individual. The information obtained and created in the ISS is currently saved in the PC of individual worker. The information is in the form of studies, articles, proceedings, presentations, academic documents and photos (Army Strategy, Doctrine, and Regulation). These come from the Czech Republic and also from international sources. The document formats are .jpeg, .gif, .doc, .rtf, .xls, .ppt, .pdf. Information subject classification is consistent with subject of individual group of ISS (security studies, warfare group, and resources and processes). The technical base is suggested as open software (RDMS PostgreSQL and application server JBoss) to achieve compatibility of SW ITM. Final state of the information processing in the ISS should be specified as centralize and integrated. Save consolidated information in accordance to subject of ISS group, central management and integration, intelligent searching. The prototype should allow conceptual searching, annotation creating, collaborating on knowledge, subject publishing according to selected criteria, exploitation of ontology and taxonomy.
388
L. Buˇrita and V. Ondryhal / Information Modelling and Knowledge Bases
GRAPH BASED REASONING
COLLABORATIVE CONTENT AND KNOWLEDGE EDITING CONTENT ANNOTATION
CONTENT AND KNOWLEDGE EDITING
ITM INDEXING
NATURAL LANGUAGE PROCESSING TERMINOLOGY EXTRACTION
ITM REASONING ONTOLOGY MANAGEMENT
LOGIC REASONING
SEARCH AND RETRIEVAL
ONTOLOGY
QUERY MANAGER
TERMINOLOGY THESAURUS MANAGEMENT
INDEXING MANAGER
PUBLISHING INDEX AGENT FOR DISTRIBUTED
KNOWLEDGE REPRESENTATION LINKS MANAGEMENT
ON DEMAND PUBLISHING TAXONOMY EDITOR
CONTENT METADATA MANAGEMENT
VISUALIZATION
ONTOLOGY REPOSITORY
WEB ASSISTANT
ITM CORE
YOUR FAVOURITE SEARCH ENGINE
TEXT MINING KNOWL. EXTRACTION
AUTOMATED PUBLISHING
ITM PUBLISHING
CLASSIFICATION
CONTENTS CMS
TEMIS IDE
END OF BOOK INDEX and TABLES GENERATION
CONTENT CRAWLER
YOUR FAVOURITE CMS
Figure 3. ITM functionality with context of other possibilities Project Phases: • Preparation phase, education in knowledge management, ontology, ITM etc. • Installation of DBMS PostrgreSQL, AS JBoss, SW ITM. • Ontology research and preparation. • Prototype building, implementation and verification. • Results demonstration and evaluation phase. Method of thesauri design: • Preparation of typical ISS document base. • Thematic vocabulary ad-hoc specification. • Analyse of document base (text mining, harvesting). • Thematic vocabulary corrections and thesauri definition. Future work: • Ontology definition. • Automatic annotation. • Information retrieval from Internet sources using thesauri and ontology. References [1] [2] [3] [4] [5]
BRUCE, Thomas A. Designing Quality Databases with IDEF1X Information Model. Dorset House Publishing, 1992. BURITA, L., ONDRYHAL V. NATO C3 Architectures and Difficulties of Application in National Environment. In EJC-2006. Trojanovice, CR: TUO, 2006, ISBN 80-248-1023-9, pp. 98-105. Information Exchange for Future Coalition Defence, Draft End Report v.44, December 2006, NATO-RTO-IST-ET-05, pp.48. NATO Network Enabled Capability Feasibility Study, Version 2. NATO: NC3A, October 2005, 623pp. THE JOINT C3 INFORMATION EXCHANGE DATA MODEL (JC3IEDM Main). Germany, Greding: MIP, 2006, 292 pp.
Information Modelling and Knowledge Bases XIX H. Jaakkola et al. (Eds.) IOS Press, 2008 © 2008 The authors and IOS Press. All rights reserved.
389
On the Construction of Ontologies based on Natural Language Semantic Terje AABERGE Western Norway Research Institute P.O. Box 216, N-6851 Sogndal, Norway Abstract. Ontologies based on natural language semantic are supposed to represent the mental models that speakers of a language possess of domains. They are thus commonly understood and may serve as a efficient means for communication between humans and computers. A menu based on a taxonomy extracted from such an ontology may therefore serve as interface for a web site communicating information about the corresponding domain. To determine an ontological representation of a mental model one must consider not only the meaning of isolated words, but also how they enter into true sentences and valid inferences. In this paper a method is presented that consists in identifying true sentences and using valid inferences to construct taxonomies for categorisation hierarchies. The method is discussed and then applied to the construction of a taxonomy for the domain of tasks provided by the Norwegian municipalities.
Introduction Ontologies are means to efficiently structure knowledge about the physical and mental universes. They are loosely speaking of three kinds, categorisation schemes based on the semantic categories of natural language, formal classification taxonomies and mathematical systems. They all define a semantic structure that endows the description languages of their domains of application with an implicit semantic and accordingly provide them with reasoning capabilities. But they differ both with respect to the way they are constructed and with respect to their intended purpose. Thus, presently the main motivation behind the construction of categorisation schemes is to make possible a more direct communication with the computer, communication based on the premises of the user rather than abstract programming language1. Such ontologies are often constructed on the basis of a set of nonconstraining guiding principles2,3. In this paper I present a principle for the construction of the taxonomic backbone of naturallanguage-based ontologies. The method consists of a number of steps. First, one identifies the vocabulary used to describe the domain, then one establishes a set of true, sentences about the domain, combine as many of the sentences as possible into valid inferences, and then orders the terms in a taxonomy that represents the logical structure of the inferences. I start the presentation by considering the theoretical background and then outline the methods of empirical investigation. The method is a simplification of the method of mathematics. In mathematics the set of axioms and definitions constitute an ontology. Its construction is a result of the study of the logical relations between statements and the decision of what should be axioms, definitions and theorems. Thus, while the method should be relatively well established it is not commonly applied, as can be seen from the (lack of) structure of the menus of many web sites.
390
T. Aaberge / On the Construction of Ontologies Based on Natural Language Semantic
The philosophical basis for this work is provided by Tarski4 and Wittgenstein5. I apply Tarski’s use of metalanguage, and I adhere to Wittgenstein’s picture theory from Tractatus in interpreting what a model is. In addition, the discussion of categorisation is inspired by Wittgenstein’s Philosophical Investigations6. In this work Wittgenstein supplements the standard definition of meaning by also referring to how words and expressions acquire meaning through use; for example, words acquire meaning not only through extension, but also through how they enter into true sentences and valid inferences. These ideas are taken up and partially justified by the results of cognitive linguistics7. 1. Domain of Investigation It is commonly assumed that the mental representations of categories, facts and systems are the same for all speakers of a language and that they are mirrored in the semantic of the language and supported by its syntactic and logical rules. This hypothesis justifies the search for and construction of ontologies that one might not be able to formulate explicitly, but that one immediately will recognise the referents of. The semantic of a language is constituted by the relations between an external reality, mental reality and the signs of language. To emphasise these relations one distinguishes between category, concept and predicate. A category is a naturally assembled set of individual systems. A concept is a cognitive entity that represents a category. And a predicate is a physical sign that represents both a category and the corresponding concept. Therefore, the concept is a mediator between a category and a predicate. This relationship is represented by the semiotic triangle concept
category
predicatete
where the arrows state directions. They stand for maps between sets of classes, concepts and predicates. Moreover, the arrow from class to predicate is a derived map being defined by the condition of commutativity of the diagram. Similarly, we distinguish between an atomic fact, a thought that represents the fact and an atomic sentence that represents both the fact and the thought. We also distinguish between a system, the mental model of the system and the model that represents both the system and the mental model. Concepts are more or less general. Fruit is a more general concept than Apple because the category Fruit contains the category Apple but also other categories like Plum. Thus, Apple and Plum are kinds of Fruit. The elements of certain sets of categories can therefore be arranged hierarchically according to generality of meaning. A categorisation hierarchy can be represented graphically by a taxonomy which pictures the hierarchy as an inverted tree. The nodes’ titles in the different levels of the linguistic representation represent, as for the nodes’ titles of classification taxonomies, degrees of generality. A taxonomy supplemented by relational sentences constitutes a categorisation ontology.
T. Aaberge / On the Construction of Ontologies Based on Natural Language Semantic
391
Taxonomies of categories are constructed by prototyping and semantic analysis. Prototyping simulates classification but without explicitly identifying and using the properties characterising the systems of the domain to define classes. One mentally associates systems that resemble the prototype and places them in the same category; “resembles” replaces the formal condition of “having the same properties as”, used in classification procedures. Therefore, tomato, despite being classified as Fruit, is commonly categorized as Vegetable because it is used as a vegetable; cooks do not put tomatoes into the fruit salads. 2. Empirical Investigations Discussing ontology construction, it is useful to distinguish between description language, theory, ontology, system description and metamodel. The description language for a domain is constituted by an appropriate vocabulary and a set of syntactic rules. A theory is a description language endowed with an ontology defining a semantic structure that makes inferring possible. An ontology is a set of implicit definitions of the words needed to describe the domain. They limit the scope of possible interpretations. It is the ‘model’ of the domain of the systems in the domain. System descriptions are specifications of the ontology that distinguishes the systems. They are representations of the systems in the description language determined by the ontology. The description depicts the system such that literate interpreters knowing the system recognise its referent. A metamodel, on the other hand, is a set of rules of interpretation expressed in the metalanguage; these rules must be known to understand the ontology and the model. Ontology construction is an iterative method that consists of making a preliminary ontology and then testing the result. Attempts to further improve the ontology follow testing and these attempts involve further testing, which consists of investigating if the ontology faithfully represents the domain considered and if itl can provide the basis for the construction of a user-friendly tool. To test if an ontology faithfully represents a domain consists of inspecting the domain and then compare it to the ontology. To investigate whether existing man-made tools satisfy requirements for usability, in other words, whether the implementation of scenarios supported by the ontology properly simulates user behaviour, one will have to ask the users of the tools. The results from such inquiries might allow one to identify criteria of usability for a category of tools. Such criteria can be related to the ontology; furthermore, they define an evaluation scheme for the category of tools belonging to the domain considered. There are several complementary methods to determine the mental models of humans and to test their representation in language8. For our purpose the most important methods are dialogs, group tests and user testing. 3. A Taxonomy for Municipalities The method has been applied to the modelling of the domain of tasks of Norwegian municipalities, excluding the administrative and political tasks. It was done together with domain experts with the aim to establish a user-friendly menu for the information site of a municipality. It took as a point of departure the vocabulary found in Norsk tenestekatalog, which is a catalogue established by the Association of Norwegian Municipalities. The catalogue lists all the tasks a Norwegian municipality is assumed to perform. It not only
392
T. Aaberge / On the Construction of Ontologies Based on Natural Language Semantic
provides keywords, but also describes some tasks in complete sentences. To assure usability of the result as a menu, the perspective chosen was that of the inhabitants. Examples of sentences that can be formulated in this vocabulary from this perspective are: child welfare service is provided by the municipality environmental plan is made by the municipality permission for construction is given by the municipality social subsidy is yielded by the municipality tax collection is performed by the municipality fire supervision is carried out by the municipality
These sentences give examples of the kinds of tasks a municipality performs for its inhabitants. They are formulated as relations between the variable of the category Municipality and the variables of the categories of tasks. The relations are ”provide”, ”make”, ”give”, “yield”, ”perform” and ”carry out”. The category Municipality consists of all the municipalities. In Norway there are 434, each having a unique name and occupying a well-defined geographical area, and all of which add up to mainland Norway. Temporary Foster Home, Foster Home, Preventive Measure and Child Care are examples of other categories of tasks performed by a municipality. They all consist of tasks that belong to the Child Welfare Service. This is expressed by the following sentences: temporary foster home is a Child Welfare Service foster home is a Child Welfare Service preventive measure is a Child Welfare Service child care is a Child Welfare Service
and one of the many valid syllogisms we can establish is the following: all temporary foster homes are Child Welfare Services all child welfare services are provided by the municipality all temporary foster homes are provided by the municipality It should e noticed that this is not a syllogism because neither “a child welfare service” nor “a temporary foster home” is a municipality. Moreover, we notice that all the expressions – “child welfare services”, “environmental plans”, “permissions for construction”, “social subsidies” and “fire supervision” – are composed. The first words of the composed expressions impose restrictions on the meanings of the second words. In fact, we can make valid syllogisms of the kind: all temporary foster homes are Child Welfare Services all child welfare services are Services all temporary foster homes are Services The second word thus represents a more general concept than the compound term. Taking this into account we get taxonomy branches like the following: Municipality Service Child Welfare Temporary Foster Home Foster Home Preventive Measure
T. Aaberge / On the Construction of Ontologies Based on Natural Language Semantic
393
Child Care
Notice that the relation between Municipality and the other predicates are not that of inheritance. I have still called this structure a taxonomy because it can be represented graphically as an inverted tree. The complete representation of the taxonomy of the domain Municipality is found in http://sfj.vestforsk.no/ht/municipality.html. The complete ontology will also contain true sentences expressing relations between non-vertical concepts. These relations supplement those of the taxonomy and may be represented by links in a web site describing a municipality. User tests were conducted on the taxonomy of Municipality. Thirteen persons participated, eight women and five men aged between 25 and 65. All were experienced web users. However, none had specific knowledge of the internal workings of a municipality. They were presented with between four and nineteen problems. The user tests did not uncover any flaws in the established taxonomy. 4. Concluding remark The goal of this paper has been to demonstrate how one can use valid inferences to construct a taxonomy once one has identified the appropriate vocabulary for the description language of a domain and a set of sentences that describe the domain. When these are identified by means of sentences from natural language that are true for the speakers of the language and the same is the case for the inferences, there is a good chance that one will manage to construct an ontology that represents the mental model of the users. This claim is based on the hypothesis that mental models are already implicit in the semantic of a language. The hypothesis is only a slight extension of the cognitive understanding of the relations between category, concept and predicate that are expressed by the semiotic triangle.
References [1] Daconta, MC, Obrst, LJ, and Smith, KT. The Semantic Web: A Guide to the Future of XML, Web Services and Knowledge Management, Wiley Publishing Company (Boston 2003) [2] Noy, NF, McGuinness DL. Ontology Development101: A Guide to Creating Your First Ontology, http://protege.stanford.edu/publications/ontology_development/ontology101.pdf [3] Smith, B. Against an Idiosyncracy in Ontology Development, http://ontology.buffalo.edu/bfo/west.pdf [4] Tarski, A. Logic, Semantic, Metamatematics, Hackett Publishing Company (Indianapolis 1955) [5] Wittgenstein, L. Tractatus Logico – Philosophicus, Routledge (London 1922) [6] Wittgenstein, L. Philosophical Investigations, Routledge (London 1952) [7] Croft, W, Cruse, DA. Cognitive Linguistics, Cambridge University Press (Cambridge 2004) [8] Speel, P-H, Schreiber, ATh, van Joolingen, W, van Heijst, G, Beijer, GJ. Conceptual Modelling for Knowledge-Based Systems, http://www.cs.vu.nl/~guus/papers/Speel01a.pdf
This page intentionally left blank
395
Information Modelling and Knowledge Bases XIX H. Jaakkola et al. (Eds.) IOS Press, 2008 © 2008 The authors and IOS Press. All rights reserved.
Author Index Aaberge, T. Aaltonen, J. Becker, G. Brumen, B. Buřita, L. Chen, X. Clemente, J. Daffara, C. de Antonio, A. Družovec, M. Duží, M. Favier, L. Funyu, Y. Gábor, A. Golob, I. Grison, T. Grzegorzek, M. Hackelbusch, R. Hagihara, S. Hai, P.V. Han, H. Hausser, R. Hegner, S.J. Heimbürger, A. Henno, J. Ito, S. Iwazume, M. Izquierdo, E. Jaakkola, H. Kangassalo, M. Kariya, H. Kidawara, Y. Kiriyama, S. Kiyoki, Y. Kő, A. Lahouaria, B. Leclercq, E. Leppänen, M. Liu, B. Locuratolo, E. Masuda, K.
389 142 330 276 384 40 298 330 298 276 21 330 346 306 276 330 190 114 290 208 338 1 79 314 170 290 282 190 v, 276 237 359 282 100 v, 40, 181, 282, 359 306 379 330 257 208 160 40
Moravec, R. Nakanishi, T. Noro, T. Oinas-Kukkonen, H. Ondryhal, V. Otani, N. Paci, A.M. Palomaki, J. Praks, P. Räisänen, T. Ramírez, J. Repa, V. Rozman, I. Ruuska, H. Salmenjoki, K. Saloheimo, M. Sasaki, H. Sasaki, J. Savonnet, M. Schewe, K.-D. Szabó, I. Takano, K. Takashima, A. Takebayashi, Y. Tanaka, M. Tanaka, Y. Tanttari, A. Terrasse, M.-N. Teshigawara, Y. Thalheim, B. Tokuda, T. Tuikkala, I. Tuominen, E. Uden, L. Válek, L. Vas, R. Vojtáš, P. Welzer, T. Yonezaki, N. Zettsu, K.
190 282 208 217 384 100 354 160 190 217 298 322 276 100 200 142 181 346 330 59 306 40 134 100 346 134 200 330 346 59 v, 208, 338 142 237 200 190 306 21 276 290 282
This page intentionally left blank