This content was uploaded by our users and we assume good faith they have the permission to share this book. If you own the copyright to this book and it is wrongfully on our website, we offer a simple DMCA procedure to remove your content from our site. Start by pressing the button below!
on s; Else if flag = 1 Then r :=roid(p); R'=Ru{(p,r,0)); return; Else r:=roid(p); Generate the Voronoi diagram near p; Search the I-order neighbor of p in the Voronoi diagram, s is first segment found; Insert a hook h =
K-Order Neighbor
35 36 37 38
1001
if not ExistsSegment(sJ then R':= R' u {(si,roid(si), 0)); SP := {(sc, {(s, roid(si))li€ { 1,. ..,n) )) I sc E scids(s)); End Insertpoint
Fig. 8. Algorithm of point insertion
4.2 Algorithm of Segment Insertion
Begin Transaction InsertPoint(R, p, 1, R', r, SP); InsertPoint(R, q, 1, R', r, SP); InsertSegment(R, s, R', RD, SP) (see [3] for detail, we omit parameter "ok"); Commit Transaction Fig. 9. The algorithm of segment insertion
Our algorithm of segment insertion is similar to those in stored realm and virtual realm. The process of inserting a segment s = (p, q) requires three steps: (1) insert p into realm; (2) insert q into realm; (3) insert s into realm. Because in point insertion, we distinguish if a point is an end point of a segment to be inserted, these three steps must be completed in one transaction (Fig. 9). Otherwise, the integrity rule of proper envelope will be violated.
5 Performance Analysis Taking stored realm as an example, this section presents an informal comparison of the performance of point insertions using K-order neighbor and not using it. If the distribution of spatial objects is dense, the advantage of our approach is very clear: the segments needed redrawing decrease greatly. Therefore, fewer segments are generated after redrawings. At the same time, the restriction of distortion provides a good foundation for the simplicity of transaction management. Table 1 summarizes that tasks while inserting a point using K-order neighbor and not, The column of difference indicates whether K-order neighbor increase (+) or decrease (-) the cost of performing input-output (110) or processing (CPU) tasks. Table 1. The influence of using K-order neighbor on point insertion in stored realm (modified from 181)
Stored realm
Retrieve required nodes spatial index (many entries)
Stored realm using K-order neighbor of Retrieve required nodes of spatial index (many entries)
Difference
-110, -CPU
1002 Y. Zhang et al.
Retrieve segments and points Retrieve segments and points -110 from MBR of inserted segment (possible many) Retrieve the corresponding part +I10 of basic Voronoi diagram Compute the local detail Voronoi +CPU
I Compute
the modification of 1 +CPU
I Voronoi diagram
I Write new segments into disk Compute changes to spatial
1 Retrieve I
I
the spatial objects related to the changed segments Replace the changed segments in spatial objects Write changed spatial objects to disk
I
Use Voronoi diagram to search the neicrhbors Compute changes in the realm Delete changed segments fiom disk Write new segments into disk Compute changes to spatial index Write changed index to disk Write changed basic Voronoi diagram to disk (maybe not necessary) Retrieve the spatial objects related to the changed segments Replace the changed segments in spatial objects Write changed spatial objects to disk Delete the local detail Voronoi diagram
I
I
+Cpu
-CPU - I10
I
I
I
- I10 -CPU - VO +I10 - VO - I/O - I/O
I
+CPu
The results presented in the table can be summarized with respect to the effect on I10 costs and CPU time: 110: (a) There are two factors to increase VO activities, the reason is that we store a basic Voronoi diagram in database; (b) There are eight factors to decrease I10 activities, the reason is that the total nwnber of points and segments in database is fewer. CPU: (a) There are four factors to increase CPU time, the reason is that the processes related Voronoi diagram are added; (b) There are three factors to decrease CPU time, the reason is that the total number of points and segments in database is fewer. Overall, there are two main reasons that influence I/O costs and CPU time: Voronoi diagram and the realm objects in database. It is hard to say that because of using Korder neighbor, I10 costs and CPU time are saved. The saving of 110 costs and CPU times depends on the distribution of spatial objects (i.e., the applications). However, we can conclude that using K-order neighbor decreases the realm objects and
I
K-Order Neighbor
1003
simplifies the transaction management. Furthermore, the stored Voronoi diagram can speed many ROSE algebra operations such as inside, intersection, closest and so on.
6 Conclusion This work is being carried out in the context of a project of 3D spatial database based on realm. We find that in some applications (especially the distribution of segments is dense), in spite of using stored realm or virtual realm [3] [8], point insertions and segment insertions bring out boring cascaded update. In these conditions, many segments have to be redrawn, which result many segments to occupy large storage space and make the management of transactions very complex. So we want to find an approach to restrict cascaded update. K-order neighbor [9] is a concept commonly used in Delaunary triangulation network. We use Voronoi diagram to describe this concept. K-order neighbor restricts cascaded update efficiently. Presently we have implemented the algorithm of K-order neighbor and used it in the data update of realm.
References 1. Greene, D., Yao, F.: Finite-Resolution Computational Geometry. Proc. 27th IEEE Symp. on Foundations of Computer Science (1986) 143-152 2. Gold, C.M.: The Meaning of Weighbow". In Frank, A. U., Campari, I. And formentini, U. (Eds) Theories of Spatial -Temporal Reasoning in Geographic Space. Lecture Notes in Computer Science 639, Berlin: Springer-Verlag ( 1992) 220-235 3. Guting, R.H., Schneider, M.: Realms: A Foundation for Spatial Data Type in Database Systems. In D. Abel and B.C. Ooi, editors, Proc. 3rd Int. Conf. on Large Spatial Databases (SSD), Lecture Notes in Computer Science, Springer Verlag, (1993) 14-35 4. Guting, R.H., Ridder, T., Schneider, M.: Implementation of the ROSE Algebra: Eficient Algorithms for Realm-Based Spatial Data Type. 4th Int. Syrnp. on Advances in Spatial Databases (SSD), LNCS95 1, Springer Verlag (1995) 2 16-239 5. Guting, R.H., Schneider, M.: Realm-Based Spatial Data Types: The Rose Algebra VLDB Journal, Vo1.4 (1995) 100-143 6. Cotelo Lema, J. A., Guting, R.H.: Dual Grid: A New Approach for Robust Spatial Algebra Implementation. Fernuniversity Hagen, Informatik-Report 268 (2000) 7. Cotelo k m a , J. A,: An Analysis of Consistency Properties in Existing Spatial and Spatiotemporal Data Models. Advances in Databases and Information Systems, 5" East European Conference ADBIS ' 200 1, Research Communications, A. Caplinskas, J. Eder (Eds.): Vol. 1 (2001) 8. Muller, V., Paton, N.W., Fernandesyy, A.A.A., Dinn, A., Williams, M.H.: Virtual Realms: An Efficient Implementation Strategy for Finite Resolution Spatial Data Types, In 7th International Symposium on Spatial Data Handling - SDH'96, Amsterdam (1996) 9. Zhang, C.P., Murayama, Y.J. : Testing local spatial autocorrelation using k-order neighbours. Int. J. Geographical Information Science, Vol. 14, No.7, (2000) 68 1-692
A Hierarchical Raster Method for Computing Voronoi Diagrams Based on Quadtrees Renliang ZHAO
Zhilin LI ', Jun CHEN
"
C.M.
old' and Yong ZHANG~
' ~ e ~ a r t m e of n t Cartography and GIs, Central South University, Changsha, China, 410083 { [email protected] } 'Department of Land Surveying and Geo-Informatics The Hong Kong Polytechnic University, Kowloon, Hong Kong { [email protected] } "ational Geometries Center of China, No. 1 Baishengcun, Zizhuyuan, Beijing, China, 100044 { [email protected] } 4~epartment of Computer Science and Technology, Tsinghua University ,Beijing, 100084
Abstract. Voronoi diagram is a basic data structure in geometry. It has been increasingly attracting the investigation into diverse applications since it was introduced into GIs field. Most current methods for computing Voronoi diagrams are implemented in vector mode. However, the vector-based methods are good only for points and difficult for complex objects. At the same time, most current raster methods are implemented only in a uniformed-grid raster mode. There is a lack of hierarchical method implemented in a hierarchical space such as quadtrees. In this paper such a hierarchical method is described for computing generalized Voronoi diagrams by means of hierarchical distance transform and hierarchical morphological operators based on the quadtree structure. Three different solutions are described and illustrated with experiments for different applications. Furthermore, the errors caused by this method are analyzed and are reduced by constructing the dynamical hierarchical distance structure elements.
1 Introduction A Voronoi diagram is one of fundamental geometric structure, actually describes the spatial influent region for each a generator and each point in the influent region associated with a generator is closer to the generator than the others [I],[14],seen as Fig. 1. Voronoi diagrams have been applied widely in various areas since they were originally used to estimate regional rainfall averages in 191 1 [I],[20].In G I s area, Voronoi diagrams are also taken as one useful tool and have been increasingly attracting the investigation into diverse applications of Voronoi methods [3],[5],[6],
[81,[101, [151,[181,WI.
Generally, Voronoi diagrams can be implemented in vector space and also in raster space. But most current methods are vector-based, for example, the classic divide and conquer method, incremental method, sweepline method [1],[9],[15] [20]. P.M.A. Sloot et al. (Eds.): ICCS 2002, LNCS 2331, pp. 1004−1013, 2002. Springer-Verlag Berlin Heidelberg 2002
A Hierarchical Raster Method for Computing Voronoi Diagrams 1005
However, such vector-based methods are good only for discrete spatial points, difficult for complex objects such as line and area objects.
Fig. 1. A Voronoi diagram for points
At the same time, it has been found that Voronoi diagrams can be implemented very well in a raster space. Compared with vector-based methods, raster methods are performed faster and more simply [2],[13],[15],[20]. However, these methods are implemented only in a uniformed grid space. Compared with a uniformed grid structure, a hierarchical structure of space like a quadtree often occupies less space and costs less execution time [23]. In fact, hierarchical methods like that of quadtree have been popular and proved efficient in many areas of GIs including spatial modeling, spatial query, spatial analysis, spatial reasoning and so on[7],[21]. But so far there are a few efforts related to the hierarchical implementation of Voronoi diagrams, despite the fact that the quadtree data structure was used very early in the implementation of Voronoi diagrams. For instance, in Ohya's method, quadtrees are actually only used as an index of buckets to reduce the time of the initial guess of inserted new generator points [17],[19]. However, this method works only for points in a vector space. In another research closely related to the computation of Voronoi diagrams based on framed-quadtrees, Chen et.al. (1997) adopted a modified quadtree structure to compute Euclidean shortest paths of robotics in raster (or image) space [4]. But the "quadtree structure" there is not the ordinary quadtree, it must be modified into the framed-quadtree whose each leaf node has a border array of square cells with the size of smallest cell before the method can work. In this paper, a hierarchical method is presented for implementing Voronoi diagrams, directly based on the ordinary quadtree structure as a very good structure representing multi-resolution data or for space saving. In this method, generalized Voronoi diagrams are performed with hierarchical distance transform and mathematical morphological operators in a raster space represented with the standard qudatrees. In the following Section 2, some related operators for the hierarchical computation of quadtrees are introduced. In Section 3 Voronoi diagrams are implemented hierarchically based on these operators, and three solutions are described and illustrated with experiments for different applications. In Section 4 the errors caused by this method are analyzed and a corresponding improvement solution is given by
1006 R. Zhao et al.
means of constructing the dynamical hierarchical distance structure elements. Conclusions are given in Section 5.
2 Quadtrees and Hierarchical Distance Computation As well known, quadtree is one of popular data structure representing spatial data in a variable multi-resolution way in many areas related to space such as GIs, image processing and computer graphics. Quadtree is a hierachical data structure and is based on the successive subdivision of space into four equal-sized quadrants[22], seen as Fig. 2. In such a quadtree, there are three types of nodes: black nodes, grey nodes and white nodes. They represent region of one spatial object, mixed regions of two or more spatial objects and free space respectively. Especially, if a quadtree only records its leaf nodes according to a certain location code, such quadtree is called linear quadtree, while the above quadtree is called regular quadtrees. Different kinds of quadtrees could be suitable for different applications. The reduction in storage and execution time is derived from the fact that for many images, the number of nodes in their quadtree representation is significantly less than the number of pixels. In terms of the definition of Voronoi diagrams, the formation of Voronoi diagrams is based on the distance. So the most important operation is the computation and propagation of distance in a hierarchical space. This operation can be accomplished with a local method or global method.
Fig. 2. A regular quadtree
The local method often uses small masks (for instance, the mask consisting of neighbouring pixels with 3 columns and 3 rows) to spread distance and obtain global distance, where the small masks are often associated with definitions of local distance such as city block and Chamfer distance. This is actually the distance transform using small neighbourhoods in image processing [2]. The high efficiency is one of main reasons why this method is popular in image processing. But it will be also costly if the propagation of distance operation covers all pixels of each level. The local method can be straightly generalized to linear quadtrees, shown in Fig. 3(a). The main difference is the distance transform is measured in multiples of the minimum sized quadrant.
A Hierarchical Raster Method for Computing Voronoi Diagrams 1007
The global method is different the local method. In the global method, the global distance can be directly calculated with an equation of distance in a coordinate system in which each pixel is assigned to a coordinate such as the number of row and column. For instance, the Euclidean distance D between two pixels Pl(il,jl) and P2(i2,j2)can be calculated with the following equation.
The global method can get the distance between any pixels in the whole space more accurately and flexibly than the local method, but the cost will be very high while all of distances between any pixels need. Therefore, in order to keep the balance between efficiency and precision, it is an alternative choice to integrate the local method and global method, i.e., the distance operation can be accomplished by combining the direct calculation of global distance and the distance transform with small masks. In this mixed method, one can firstly get the global distance between some connected pixels and a spatial object, and then make distance transform outside the spatial object at the beginning of these connected pixels with the global distance. (Seen in Fig. 3(b)) Derived from the above procedure, the efficiency and precision of this integrated method are between the local method and global method for distance calculation.
(4
\"I
Fig. 3. (a) The integration of global method and local method in a hierarchical space; (b)
Distance transform applied to linear quadtrees
The mixed method can play a better role in the distance calculation of the hierarchical space since it is more flexible than ordinary distance transform. As known, it is derived that some regions of a certain level do not have to continue to be processed in next levels in a hierarchical space. In this case, the flexibility of mixed method makes it possible to omit the distance calculation over the regions unnecessary to be processed. In Fig. 3(b), for instance, it is supposed that the white regions are unnecessary to be continued to process, others are to be continued. With the mixed method, only a part of pixels (with bold lines in Fig. 3(b)) are involved in the global distance
1008 R. Zhao et al.
calculation, the following distance propagation begins with these pixels with the global distance as seen in Fig. 3(b). In addition to the above approximate method for the distance calculation on linear quadtrees, such distance calculation between free pixels and spatial objects can be also obtained using the dilation operator in mathematical morphology. The dilation operator is one of two basic operators in mathematical morphology, represented as follows [24]:
Where A is the original space, and B is the structure element. The morphological dilation operator has been not only successfully applied to the distance calculation in a uniform grid space [13],[15], but also used for the distance calculation in a hierarchical structure such as the pyramid [16]. Here, it is attempted to implement a new method for the distance calculation using morphological dilation operator in regular quadtrees or linear quadtrees, called hierarchical morphological methods based on quadtrees. The key of a hierarchical dilation operator is substantially made up of the sequence of hierarchical structure elements in quadtrees corresponding to the ordinary structure element. Each hierarchical structure element belongs to a level of quadrees. Fig. 4 shows the structure elements at level 1 and 2 corresponding to 'city block'.
Fig. 4. The structure elements at level 1 and 2 corresponding to 'city block'
3 Hierarchically Implementing Voronoi Diagrams on Quadtrees In raster space, the formation of Voronoi diagrams is substantially viewed as an iterative procedure of subdivision of space. In each iterative step of current ordinary raster methods, each pixel will be determined to belong a certain spatial object according to the distance calculation values. However, it is unnecessary to determine all pixels because possible changed pixels are only those pixels located in boundaries between regions of different spatial objects or between free regions and regions of spatial objects. The application of quadtrees can avoid those redundant operations and finding those possible changed pixels efficiently. The quadtree based hierarchical implementation of Voronoi diagrams is actually also the procedure of continuous subdivision and reconstruction of the quadtree. Due to different ways of hierarchical processing, the hierarchical method for Voronoi diagrams is also implemented in
A Hierarchical Raster Method for Computing Voronoi Diagrams 1009
different routines. Different routine of the implementation will use different distance calculation.
3.1 Mixed Distance Calculation Routine For regular quadtrees, it is a good way to implement Voronoi diagrams from top to bottom and level by level. In this way, the distance calculation can be accomplished with the mixed method in Section 2. The hierarchical implementation can consist of the following steps.
Fig. 5. The hierarchical implementation for Voronoi diagrams using the mixed distance calculation based routine (Total pixels for distance transform 16+40+84=140,but total pixel will be 256 using non hierarchical methods)
Use global distance calculation method to compute the distance between white nodes and black nodes or grey nodes at the third highest level of the quadtree, then assign those white nodes to the corresponding black nodes or grey nodes Search for those nodes belonging to each black node or grey node including only one spatial object and not adjacent to other nodes belonging other spatial objects, label them as unused nodes because they are unnecessary to continue to be processed, and label those nodes adjacent to other nodes as boundary nodes; Subdivide the grey nodes and boundary nodes into less nodes at the next lower level, computing the global distance between less boundary nodes and black nodes or grey nodes including only one spatial objects; Perform the distance transform using the local method for distance calculation in the regions except nodes labeling 'unused', assign those white nodes to the corresponding black nodes or grey nodes; Repeat step (2)-(4) until the bottom level.
1010 R. Zhao et al.
Fig. 5 gives an example of the above routine.
3.2 Linear Quadtrees Routine While the raster space is organized into linear quadtrees, it will be very efficient to perform the hierarchical computation of Voronoi diagrams in a way of mixed levels. The feature of this routine is that local distance transform is straightly applied to all levels. Based on the technique, the hierarchical implementation for Voronoi diagrams is described as follows.
I
(b) Fig. 6. (a) The hierarchical implementation for Voronoi diagrams using linear quadtrees based routine, (b) Hierarchical morphological method for Voronoi diagrams Perform the distance transform on linear quadtrees, make each white nodes associated with a certain spatial object; (2) Search for those white nodes which belong to each black node and are not adjacent to other white nodes belonging other spatial objects, label them as unused nodes because they are unnecessary to continue to be processed, and label those nodes adjacent to other nodes as boundary nodes; (3) Subdivide the larger boundary nodes into less nodes at the next lower level, (4) Perform the distance transform in the regions except nodes labeling 'unused', make each white nodes associated with a certain spatial object; (5) Repeat (2)-(3) till no white node need subdivision. An example using this routine is shown in Fig. 6(a).
(1)
A Hierarchical Raster Method for Computing Voronoi Diagrams 1011
3.3 Morphological Routine When original raster data are represented in the uniform grid format, it is necessary to reorganize them into quadtrees firstly for hierarchically computing Voronoi diagrams. At the same time, the hierarchical distance calculation based on morphological method is actually perforrned at the bottom level of qudtree i.e., the original uniform grids. So in this case, it is suitable for implementing the Voronoi diagrams hierarchically using morphological dilation operator frorn the each higher level to the bottom level. It can be represented as follows. (1) Renew to organize the original raster data into a quadtree and construct a sequence of hierarchical structure elernents corresponding to the definition of distance in raster space; For each black node, perform the hierarchical dilation operation stated in the (2) previous section; (3) Make the distance transform only those pixels dilated by hierarchical operator, assign corresponding spatial object to each pixel of those; Performing the union operation of quadtrees, merger those dilated pixels and (4) update the current quadtree; ( 5 ) Repeat (2)-(4) till no distance value distance. Fig. 6(b) illustrates the result of this procedure.
4 Comparison and Improvement The hierarchical implementation of Voronoi diagrarns can be realized in three routines. The first routine based on mixed distance calculation method can be suitable for regular quadtrees. The second routine directly using distance transform on linear quadtrees is more suitable for raster data in the format of linear quadtrees. The third routine based on hierarchical morphological operators can be performed in uniformed grid space with the combination of linear quadtrees. More important, the three routines are also different in the efficiency and precision. From the viewpoint of efficiency, the best routine is the second routine because the distance transform is directly implemented on all levels and it cost less tirne to get an appropriate distance of all leaf nodes. The tirne cost tirne of the third routine is rnore than the others since it involves many the operation of union quadtrees in each iterative procedure of the irnplementation of Voronoi diagrams. In precision, it is derived that the third routine is the best because all operations are actually performed at the bottorn level and the result is the same as that of the implementation of non hierarchical methods. In the other two routines, the distance calculation is perforrned at various levels and causes rnore error than the third one. However, as pointed out in the literature [15], ordinary distance transforrns increase error varying with the growth of distance. This results in larger error of Voronoi diagrams via ordinary distance transforms. In order to irnprove the precision, an dynarnical distance transform hierarchical rnethod is presented here by constructing a set of hierarchical structure of elernents close to a circle.
1012 R. Zhao et al.
The key hierarchical structure of elements can be constructed with the method introduced in Section 2, but the difference i s that each structure elernent corresponding to a level here must consist of a set of elements whose combined effect of dilation is required to be a region very close a circle. Applying dynarnically these hierarchical dilation operators, the third routine is improved.
5 Conclusions A Voronoi diagram i s a geometric subdivision of space that identifies the regions closet to each of its generator sites. Voronoi diagrams have been regarded as an alternative data model and the researches on the applications of Voronoi diagrams have been attracted increasingly in GTS area[5],[8],[9],[11],[12],[15],[25]. The quadtree i s very good data structure representing a hierachical or rnultiscale spatial data. A quadtree based method for the hierarchical cornputation of Voronoi diagrams in raster space is described in this paper. It can be irnplernented in three differentroutines in order to adopt different conditions. Three approaches are different in efficiency and precision. The morphological routine is the best in the aspect of precision, and linear quadtree based routine i s rnore efficient that others in general cases. One can select a suitable routine in his practical needs. Acknowledgement. The work wab biibbtantially bupported by a grant from the Rebearch Grants Council of the Hong Kong Special Adminibtrative Region (Project No. PolyU 5048198E) and partially supported by a grant from the National Natural Science Foundation (No. 40025 10 1 ).
References I. Aurenhammer, Franz, 199 1, Voronoi diagram-A bunley of a fundamental geometric data structure. ACM Covlputirlg S~lr~ieys, 23 (3), 345.405. 2. Borgefors, G., 1986, Distance transfor~nationsin digital images. Computer Vision, Graphics and hnage Processing, 34, 344-371 3. Chakroun, H.; Benie, G.B; O'Neill, N. T., Deaileta, J., 2000, Spatial analyais weighting algorithm using Voronoi diagrams. International Journal of geographical Information Science. 14(4), 3 19-336 4. Chen D. Z., Szczerba R. J. and Uhran, J., 1997, A framed-quadtree approach for determining Euclidean shortest paths in a 2D environment. E E E Transaction on Robotics and Automation, vol. 13, pp668-68 1 5. Chen, Jim, Li, C., Li, Z. and Gold, C., 2001, A Voronoi-based 9-intersection model for spatial relations. International Journal of geographical Information Science, 15(3): 201220 6. Chen, Jim, Zhao, R.L., and Li., Zhi-Lin, 2000, Describing spatial relations with a Voronoi diagram-based Nine Intersection model. In: SDH'2000, Forer, P., Yeh, A. G. 0. and He, J. (eds.), pp4a.4-4a. 14
A Hierarchical Raster Method for Computing Voronoi Diagrams 1013 7. David, M, Lauzon, J . P. and Cebnan, J . A,, 1989, A review of qudtree-based ctrategiec for interfacirig coverage data with digital elevation models in grid form. Int. J . Geographical hforrnatiori Systemc, 3(1): 3-14. 8 .Edwards ,Geoffrey, 1993, The Vororioi Model and Cultural Space: Applications to the Social Scierices and Humanities In: Spatial information theory : a theoretical basis for GIS : European Conference, COSlT'93, Marciana Marina, Elba Ibland, Italy, September 19-22, Berlin ; New York : Springer-Verlag, Lecture Noteb in Cornpuler Science 7 16, pp202-214 9 .Gahegan, M. and Lee, I., 2000, Data btructiires and algorithrnb to bupport interactive bpatial analysis using dynamic Voronoi diagrarnb, Cornpiiterb, Environment and Urban Sybternb, 24: 509-537 I O.Gold, C. M., 1994a, a review of the potential applicationb for Voronoi methodb in Geomatics. In: the proceeding of the Canadian Conference of GIS, Ottawa, 1 647- 1 652 I I .Gold, C.M., 1994b, Advantageb of the Voronoi bpatial model. In: Frederikben, P. (ed.). Proceedings, Eurocarto XTI; Copenhagen, Denmark, 1994. pp I - 10. 12.Gold, C.M.; Rernrnele, P.R. and Roob, T., 1997, Voronoi methoda in GIs. In: Van Kreveld, M., Nievergeld, J., Roob, T. and Widmeyer, P. (edb.), "Algorithmic Foundations of GIs. Lecture Notes in Computer Science No. 1340", Springer-Verlag, Berlin, Germany, pp. 2 1 35. 13.Kotroplulos, C., Pitab, I. and Maglara, M., 1993, Voornoi tebbellation and Delauney triangulation using Euclidean disk growing in z'. TEEE, V29-V32. I4.Lee, D. T. and Drysdale, R.L., 1981, Generalization of Voronoi Diagram in the plane. SIAM Journal of Computing, 10, 73-87. 15.Li, C., Chen, J. and Li, Z. L., 1999, Raster-baaed methods or the generation of Voronoi diagrams for apatial entities. International Journal of Geographical Information Science, 13 ( 3 ) , 209-225. I 6.Liang, E. and Lin, S. 1998, A hierarchical approach to dibtance calculation wing the bpread f~inction,International Journal of Geographical Information Science, 12(6), 5 15-535 17.Marston, R.E.; Shih, J.C. 1995, Modified quaternary tree bucketing for Voronoi diagram:, with multi-scale generators Multirebolution Modelling and Analybk in 6nage Processing and Canputer Vision, TEE Colloquiuin on , 1995 Page(s): I 111 - 1 1 I6 1 8.Mioc, D.; Anton, F.; Gold, C.M. and Moulin, B., 1998, Spatio-tempera1 change representation and map updates in a dynamic Voronoi data btructure. In: Poiker, T.K. and Chrisman, N.R. (eds.). Proceedingb, 8th International Sympobiuin on Spatial Data Handling; Vancouver, BC, 1998. pp. 44 1-452. I (>.Ohpa,T., Iri, M. and Murota, 1984, A fabt Voronoi diagram with quadternay tree bucketing. Information Processing Letterb, Vol. 1 8, pp. 227-23 1 ZO.Okabe, A,, Boots, B. and Sugihara, K., 1992, Spatial Tessellations: Concept:, and Applications of Voronoi diagrams, Chichester, England, New York, Wiley & Sonb. 21 .Papadias, D. and Theodoridis Y., 1997, Spatial relatiom, minimum bounding rectangleb, and spatial data structures. International Journal of Geographical Information Science, 11(2), pp11 1-138 22.Samet, H., 1982, Neighbor finding techniques for images represented by quadtrees. Conputer Graphics and Image Proceaaing, Vol.18, pp.37-57. 23.Samet, H., 1989, Design and analyaia of apatial data structures: quadtrees, octreaa, and other hierarchical methods, Reading MA. 24.Serra, Jean Paul., 1982, hnage analysis and mathematical morphology London ; New York : Academic Press. 25.Wright, D.J., and Goodchild., M. F., 1997, Data from the deep: implications for GIs community. Iriterriational Journal of Geographical Iriforrnation Science, 11,523-528
The Dissection of Three-Dimensional Geographic Information Systems
'
Yong ~ u e ', Min sun' ,Yong zhang4 and RenLiang zhaoi 'LARSIS,Institute of Remote Sensing Applications, Chinese Academy of Sciences. Beijing, 100101, PR China '~choolof Infomialics and Multimedia Technology, University of North London, 166-220Holloway Road, London N7 8DB. UK {[email protected]}
'~nstituteof RS & GIS, Peking University, Beijing, 100871, PR Chim '~nstilute of Computer lechnology, Tsinghua University, Beijiiig, 100871. PR China 5~nstiluteof Surveying and Mapping, Central South Universily, Changsha, 410083. PR China
Abstract. In this paper we dissected the 3-dimensional Geographic Infom~ation
Systems (3D GISs) and disambiguated some points such as (1) 3D GIS should be a componail of GIs, il's a subsystem; (2) dala modelling is not the main obstacle to its development: (3) it's no necessary and also very difficult for 3D CiIS to replace GIs; (4) the main developing direction of 3D CilS should be special subsyslems, that is, 3D GIS research musl be based on relative application areas.
1.
Introduction
We are living in a true 3-dimensional (3D) world. The environment and almost all artificial objects have 3-dernensional sizes. However, as so far, it's difficult to establish a uniform data model to represent the environment and the objects. In CIS, objects are represented using its projections on 2D plane. Although this kind of representation could solve many problems, it is still very difficult to represent and process the geometry and attributes for third spatial dimension. For example, CIS can't represent real 3-dimensional nature and artificial phenomena in geosciences area, such as in geology, mining, petroleum, etc. Many research works about 3D CIS have been carried out in order to remedy such defects in solving true 3-dimensional problems. It has been found that it is almost impossible to establish a 3D CIS that has similar functions as those in current GISs, especially for the processing of complex 3dimensional spatial relationships. Data modelling of 3D objects is one of the most difficult problems. As data model is the fundament to the establishment of CIS, research works on 3D CIS have been mainly focused on data model. S o far, there isn't a perfect data model found. The great difficulties for 3D CIS research could be summarized in three points: P.M.A. Sloot et al. (Eds.): ICCS 2002, LNCS 2331, pp. 1014−1023, 2002. Springer-Verlag Berlin Heidelberg 2002
The Dissection of Three-Dimensional Geographic Information Systems 1015
I. Try to lind a completely new 3-dimensional data model to substitute CIS vector data model. Third dimensional value Z not only changes li-om attribute data to spatial data, i.e.: ( X , Y): Z 3 ( X , Y , Z ) but also a new volume element is added in addition to three elements such as point, line and face; 2. Tiy to represent various 3-dimensional objects in one data model; 3. Tiy to represent objects, especially complex real objects, in high precision. Most research works on data modelling have been to establish a common 3dimensional data model based on 3D CIS as a common platlbnn for real world complex objects, including geology, mining, petroleum, cataclysm, ocean, atmosphere, and urban environments. However, no signilicant progresses have been made so lar. In this paper, the concept and applications o l 3D CIS will be addressed lirst. Data modelling will be studied and the lurther development ol' 3D CIS will be discussed at the end.
2.
3D GIs CONCEPT AND APPLICATIONS
2.1 Applications of 3D GIs
We could find CIS applications almost in any areas that relate to spatial data. Object scales might valy li-om I :SO0 to I :1,000,000, even larger or smaller. That's in such scale areas, we could find actual objects represented by CIS data model, e.g.: in l:500 scale a polygon might represent a house, while in 1: 1,000,000 scale and a polygon might represent a city area. But in 3D CIS, the case changes a lot. We could analyse this question liom CIS data models. CIS is mainly established on vector data model. In this model, line is composed of points and face is composed of lines. Point element P(x, y, c) may represent any single object occupying one spatial point location, and also an object occupying 2-dimensional spatial areas, e.g.: streetlight, road crossing, station or even a whole city; Line element L(P,, P?, P,,. . ) may represent any object that occupies I dimensional or 2-dimensional space, e.g.: pipelines, roads and rivers: Face element S(L,, L2, L,, ... ) may represent any object that occupies 2dimensional space, from single position to sinall area or even a \ely large space, e.g.: house, square, city and whole country teriain. Therefore, CIS vector data model could represent random size objects in plane, and could be used in many spatial related applications. In 3D CIS, space is extended from 2D to 3D. Volume element has to be introduced in both vector model and raster model [14], [IS], [IG], [20]. If we consider volume element reprecented by componentc of point, line and face elements, i.e.: V( R(P, L, S )), then in theory, volume V could be used to reprecent any
1016 Y. Xue et al.
objects in Euclid space with random dimension ( d <= 3 ) and random size. However, such representation is limited in practice. For example,
City envimnment: in city environment, the main objects are artificial objects (including const~ucters,bridges, etc.). Their third dimensional sizes are usually in a certain range, e.g.: li-om I meter to 100 meters. If we use scale l:10,000 to represent one constructer with 100m in its width, length and height, its map size is only I cmx l cmx l cm. Such a size representation is much similar to its plane representation, and for many smaller objects, it already loses 3D representation significance. In addition, third infonnation significance of 3D city object would change with scale. For example, it has significance to represent one balcony in I: 1,000. But in I : 10, 000 scale, representation of whole building in detail would lose significance. Therefore, to city environment, 3D CIS spatial data modelling should be limited to large scale, e.g. no large than 1 : 10,000. Ocean t m l a t ~ m s p h ~e tr?~v i ~ ~ f ? ~ i e these t ? t s : two environments are reverse to city environment. Infonnation is captured usually in small scales, and artificial objects would lose significance. Infonnation represented usually only in small scales would have significance (e.g.: s~nallerthan I: 100,000), e.g. temperature and humidity distributing; clouds and ocean tlow excursion etc. At the same time, in such environments, medium distributing has continuity. It is difficult to model according individual object as in city environment. Therefore, in such environments, 3D G I s data modelling is not only limited to sinall scales, but also much different froin city environment. Data modelling can't be established on individual object. Mining, geology, ctztaclysm trncl pet?+oleunlen~i?+onnlent.r: in these environments, there are both nature objects and artificial objects. But if we review these environments, we could conclude two main cases. One related to resource exploring, which information capture is usually in large scales while with high precision. Another relates to macro geosciences research, such as: terra construction, cataclysm monitoring, etc., which infonnation capture is mainly limited to small scales, similar to atmosphere and ocean environments. In present research works on 3D GTS, most works are related to resource exploring. Therefore, we could get a rough conclusion that in such environments, 3D G I s data modelling are mainly limited to large scales, but have great difference to city, atmosphere and ocean environments. Now we can summarize a conclusion that 3D CIS data modelling not only has great differencer in different environments, but also har great different spatial scale requirements in different environmentr. It is vely difficult to lind a hind of volume element in data modelling to meet many different object representations in variour environments and various scaler.
2.2 3D GIs Concept and its research contents The current GTSs should be called 2D GTSs because they are developed from plane map. They could only disposal actual 2D spatial information and their data models are
The Dissection of Three-Dimensional Geographic Information Systems 1017 mainly based on 2D vector elements, such as point, line and face. Deunis et al. del'ined GIS as a soltware system that contains lunction to perlorm input, storage, editing, manipulation, analysis and display of geographically located data [4]. Here geographically located data are referred to 2D spatial data because in GTS data model, third dimensional data is disposed as attribute data. So far, there is no fonnal definition for 3D GTS. Different people have different considerations. Fritsch pointed out that main research work of 3D GTS is the integration of height information into 2D GTS [6]. Koller considered that one major difference between 3D GTS and 2D GTS is the amount of data to be processed, and the second fundamental difference is the user interface [I I]. Pigot pointed out that similar to 2D variants, 3D GTS should be capable to perform metric (distance, length, area, volume, etc), logic (intersection, union, difference), generalization, buffering, network (shortest way) and merging operations [19]. Pfound pointed out that today, most applications and data structures for 3D GTS are optiinised for visualization [I 81. They usually omit topological information in order to get a better pel-fonnance. Actually, these hints could be found in inany papers. These hints imply that 3D GTS is a new system that should be established on the base of 2D GTS. From there, we could define that 3D GTS is a software system that contains function to perform input, storage, editing, manipulation, analysis and visualization of three-dimensional geographical infonnation data in one certain application area.
2.3. Research content The majority of research works on 3D GTS are 3D data modelling and 3D data structure [6], [I 51, [I 61, [I 91, [20]. A few is on 3D visualization [I I] and others are on 3D web GTS [13][26]. From the scientific view, 3D GTS research content must be larger. On the other hand, i f we consider the above definition, 3D GTS research content may focus on following points:
3D data modelling base on actual applications; 3D data structure base on actual applications; 3D visualization technologies for actual applications; Huge multimedia data management; 3D technologies on web for actual applications. These contents have strong relations with application areas. The majority of 3D GTS research works are carried on with the other research works for actual application. Hence such considerations would meet the practice needs. While the mainstream of 3D GTS research works seldom consider situations in actual application, this is the main reason why we can't see one 3D GTS used in practice beside some special software used in certain areas like petroleum or city.
1018 Y. Xue et al.
3.
Data Modelling in 3D GIs
The final aim of data modelling is to truthfully represent the real world. However, because of limitation of theories and technologies, it is difficult to redivivus real world objects through data modelling in GIs. The data models could be classified into three levels or three types according to their representation precision: Syn7bol ~ioclel:in 2D GTS vector data model, point and line element could be considered as symbols on plane as a face is composed of lines. We can consider face is represented by line symbols because 2D GTS comes liom map, and objects representation manners are not changed from map to GIs. 2D GIs vector data model can be considered as a symbol model and this model can represent objects location and its rough shape. Symbols are suitable to represent symbolic logic. This is the reason that 2D GTS vector data model could be easily used to represent spatial topology. Basic nzodel: a few spatial data, and predefined elements, such as points, lines, faces and bodies can be used to represent various objects, and could basically retlect object's shape, spatial extension and higher precision 3-dimensional spatial location. 3D GIs data models are mainly used for the representation d basic inloimation ol' objects in real world. For example, tetrahedron models were used to represent irregular mine body [20] and CSG primitives were used to represent regular buildings [Id]. Tn city environment, third dimensional spatial information can be added to construct 3D urban G I s models easily [7][9]. Today, the majority of data models available are belong to basic model categoly. Refined nzotlel: a large number of spatial data are used in this kind of models. It is established mainly through interactive methods, and it is difference from above two kinds significantly. Many types of elements are used in this kind of data modelling, and these elements could be adjusted to meet modelling needs during modelling process. This kind of data model could rellect objects with high precision in spatial location, subtle spatial shape and vision characteristics, etc. Tt is mainly used in Virtual Reality G I s (VRGIS), especially in 3-dimensional landscape reconstruction based on GTS spatial data. As VR technology need very high precision representation of 3-dimensional objects (e.g.: construction, road bridges, trees, etc.), high precision are also requested in data capture and data modelling functions. Therefore, at the present time, this kind of data models could only be established using software packages like 3Dmax and AutoCAD. It is very difficult for GIs to perfonn this kind of data modelling. Obviously, refined model needs vely high precision representation. This may cause huge amount of data and modelling complexity. As different objects may have different characteristics, it is difficult to realize automatic data modelling in large quantities. Special nod el ling software packages must be used. Therelbre, refined model is not suitable in 3D GIs and it is more suitable for VRGIS. On the other hand, as symbol tnotlel can't be used in 3D GTS, basic nioclel.~are usually used in 3D GTS data nod el ling. Basic- niorld can still be classified into two types according its modelling purposes:
The Dissection of Three-Dimensional Geographic Information Systems 1019
I. C I I I - I - i~~ ~ o d~eJl :the aim o l this kind of inodels is to represent various objects in dillerent application areas in real world. The majority o l present data models are belong to this category. For example, Molenaar put forward a 3D FDS data structure [IS]. Pilouk et al. put fornard a tetrahedron based data inodel [20]. Pigot put forward a topological data model [19]. Li put forward a CSG based data model [14], and etc. Although these data models have different advantages, there are still some defects in different degrees as currency models to represent various complex 3-dimensional objects in real world. The realisation of truly 3-dimensional GTS solutions is not yet in view [3]. 2. Special n~oclel:it is established for certain application area. Early in 1989, Peter pointed out that elements in geosciences have a lot of communities [17]. It could be possible to establish one uniform geosciences data model. Raper discussed 3D data modelling for geosciences [21]. He pointed out that 3D data models have strong links with underground data samples and parameters. Turer pointed out that an exact model of the subsurface is not possible, simplified versions may be derived lioin geological and engineering data developed li-om exploration procedures, including borings, well tests, seismic exploration, and stratigraphic or sediinentological descriptions of depositional systems [XI. In city environment, Zlatanova used faces and nodes as main elements combining R-tree to establish a 3-dimensional topological data model on Web [26]. Gruen et al. developed a TOBAGO system using point sets to establish 3D data inodel by interactive method [8]. Teinplle put fo~warda 3D topological urban data model [24]. Koehl put fo~warda data inodelling method based on CSG level structure [lo] and Sun et al. put foiward a 3D spatial data model based on surface paitition [23]. Because special models disposal objects limited in certain application areas, data modelling is relatively simple. It is also easy to implement in practice. For example, using 3D suiface modelling method, ArcView 3D analyst module could realize modelling to construction [5].
4.
Is there a perfect spatial data model for 3D GIs?
4.1.
Perfect 3D GIs data model
Pfund pointed out in his research work that a perfect 3D GTS data model should be vector data structure, boundary representation and topology [I 81. Peter pointed out that boundary representation (B-rep) is a suitable modelling method for geosciences elements representation and operation when he did research works on one Geoscientfic Resource Management System [17]. Pigot had summarized 3D GIS modelling research works [19]. He pointed out that 3D GIS data modelling began with Molenaar's 3D FDS data structure and many others research works are developed on the base of it. Pigot himself put Ibrward a 3D spatial data model based on inanifold topology. This model is similar to 3D data inodels developed by Pilouk et al. [20].
1020 Y. Xue et al.
Actually, Molenaar's 3D FDS is an extension o l 2D GIs vector data model and it is a B-rep data model. B-rep data model is a kind ol' perlect data model in some degree. The reasons are: The basic elements are point, line, face and body. Point and line are the most basic elements, so it is easy to combine a 3D CIS with a 2D GTS; Amount of data is not large, and it is easy to establish data model and to manage; Easy to establish topology: Easy to realize visualization, and easy to do data operations; But B-rep method also has soine defects. It is difficult to verify representation validity, to perfonn Boolean operations, and to require complex algorithms for model generation [17]. However, i f we don't pay too attention to verification of representation validity, it doesn't el'lect today's 2D GIS so much. Boolean operations are not so urgent and complex algorithms become simple to today's computer hardware. So we think B-rep data model is a kind ol' perfect 3D GIs data model.
4.2.
Questions about spatial topology relationships
This question may raise many arguments, especially in 3D GTS. Topology representation is considered as the typical feature of CIS, but why do we have to represent topology relationships? Spatial topology could be established dynamically. The essence of this thinking is to calculate topology in real-time (e.g.: Voronoi diagram is a good structure). Actually, current 2D GISs could only represent a few topology relationships, and topology problems on 2D plane are still not solved vely well [I]. Topology relationships representation usually limits the establishment of data model, and it also increase the complexity of data updating and operation. The topology relationships in 3D CIS could be calculated in real-time. AutoCAD might give us a good example. There are no any topology concepts in AutoCAD, but it could be used vely well in industry, such as machine assemble. Bernhardsen pointed out that the essential difference between CIS and CAD becomes more and more blur [I]. By adding soine attribute data, inany CAD systein vendors could effectively compete GIS market. Koller used a CAD model and R-tree combined method in Vienna walking systein [I I]. Stoter also introduced a CAD model into CIS in order to solve the problem of 3D cadastral registration [22]. Koehl put forward CSG level model. His main purpose was to solve CAD models representation [lo]. In VRGIS, data modelling mainly comes li-oin 3Dmax or CAD.
4.3.
3D GIs data modelling for geoscience
Data models based on tetrahedron (or 3-simplex) could preferably be used to represent various elements in geosciences. In geosciences, most objects are natural objects not artificial ones. Lattuada and Raper put fo~ward that 3-dimensional tetrahedron network is suitable for modelling in these areas [12]. Recently, we also
The Dissection of Three-Dimensional Geographic Information Systems 1021 did some research works in term area, we reconstruct 3D geological body on the base ol' simulate artesian well data using tetrahedron network, and also reconstruct complex geological bodies on the constraint tetrahedron network. Tetrahedron model is not suitable to represent regular bodies, especially some bodies with arc shape, such as cylinder, cone and etc. Extra partition calculation and redundancy data will be introduced if tetrahedron is used to represent regular objects. In city environment, almost all objects are artificial objects. Besides, it is not necessary to represent inner attributes of city objects in GTS. As majority of city objects are regular ones, they could be represented by some primitives like rectangle, triangle, cuboids, cylinder, cone, etc. or their combinations [10][15][23]. Some CSG primitives could not be represented by R-rep, but they could be deal with parameter expressions.
5.
3D GIS Development Directions
Although the general 3D GIS system have not be progressed significantly because of the baftle in data modelling, the great progress have been got in Three-Dimensional City Model and Virtual Geographic Information System. These systems could provide very excellent visualization effects and interactive functions. 3D GIs should be designed as a subsystem of 2D GTS. Of course, it could run alone. The establishment of 3D GIS should be compatible with 2D GIs functions. It should have interface same as that for 2D GIs. And reducing 2D GIS functions would decrease much complexity in 3D GTS. On one hand, 3D GIS should have unique visualization interface as well as operation and analysis functions. Many systems are established on the base of GTS [9]. While in geology, mining, etc., although subsu1-face data is the principal, ground data is also important. For example, 2D data in GIS is usually the reference to underground works in resource mining process. Therefore, It is a crucial point to set up a connect between 3D GTS and 2D CIS, we have several views as ( I ) Establish special SDE for 3D spatial data management; (2) Extend present 2D G I s data manage functions and make them co~npatiblewith 3D spatial information management; (3) Develop 3D GTS modules in component methods; (4) Establish data change interface between 2D to 3D. Cambray et al. put lbnvard a multidimensional geographical data model, through add some hces with different heights to solve this problem [2]. Their data model could only solve regular and protruding enveloped objects. Arc view 3D analyst is a well known 3D G I s module developed based on component. Haala et al. established a 3-dimensional city model by adding some extra information and changing 2D data to 3D data [9]. 3D GTS has no great significance in many application areas, as in small scales space (e.g.: scales less then 1 : 10,000). There are great different spatial characteristics and requirements in city, geology, mining, ocean, atmosphere and other environments. It is i~npossibleto establish a currency data model to these areas in themy. In fact, li-om the review of GIs software packages (e.g. Arc/Tnlh, MapInfo, GeoMedia, AotoDesk), we would found that in these GIS software packages, 3D
1022 Y. Xue et al.
functions are limited to a few modules. For city environment, functions are limited to visualization and simple queries. We conclude that 3D GIs development should be based on special application. The research should focus on application areas information representation and analysis. 3D spatial data is only a carrier. Data modelling should focus on representation integrality to special information. It should not simply emphasize precision and topology consistency of spatial data representation. At the time being, it is not necessay, and also very difficult to establish a currency 3D GIs platform.
ACKNOWLEDGMENTS This publication is an output from the research project "Digital Earth funded by Chinese Academy of Sciences. Also, Dr. Yong Xue should like to express his gratitude to Chinese Academy of Sciences for the financial support under the "CAS Hundred Talents Program" and "Initial Phase of the Knowledge Innovation Program".
Reference 1. Bemhardsen T. "Geographic Information Systems An Introduction ". John Wiley &Sons Inc., New York, 1999. 2. Cambray B., Yeh T. S. "A Multidimensional (2D, 2SD, 3D) GeoGraphical Data Model". Proceedings of the International Conference on Management of Data (COMAD194),Bangalore, India, 1994, pp3 17-336. 3. Carosio A. "Three-Dimensional Synthetic Landscapes: Data Acquisition, Modelling and Visualisation". Photogrammetry week'99, 1999, pp293-302. 4. Deunis R. S., Arithur R. P., "Three-dimensional GIs for the earth sciences". In: Three dimensional applications in Geographical Information Systems, Taylor & Francis, Edit by Jonathan F. Raper, pp149-154, 1989. 5. Esri Company "ArcView 3D Analyst". http://www.esri.com 6. Fritsch D., "Three-Dimensional Geographic Information Systems -- Status and Prospects". Proceeding of Znternational Archives of Photogrammetry and Remote Sensing, Vol. 3 114, ISPRS, Vienna, Austria, 1996, pp215-221. 7. Gruber M. "The CyberCity Concept from 2D GIs to the Hypermedia Database". Proceedings of UM3'98, International Workshop on Urban Multi-Medid3D Mapping, 1998, pp47-54. 8. Gruen A. "Tobago-A semi-automated approach for the generation of 3D building models". ZSPRS Journal of Photogrammetry & Remote Sensing, 53, 108-118, 1998. 9. Haala N., Brenner C. and Anders K. H. "Generation of 3D City Models From Digital Surface Models and 2D GIs". ZSPRS, Vo1.32, Part 3-4W2, 3D Reconstruction and Modelling of Topographic Objects, Stuttgart, 1997, pp68-75.
The Dissection of Three-Dimensional Geographic Information Systems 1023
10.Koehl M. "The Modelling Of Urban Landscapes." Proc of International Archives of Photogrammetry and Remote Sensing. Vol. XXXI, Part B4. Vienna, 1996, pp460-464. 1l.Kofler M. "R-trees for Visualizing and Organizing Large 3D GIs Databases". PhD. thesis, Technischen Universitat Graz, Graz, 1998. 12.Lattuada R., Raper J. "Applications of 3D Delaunay triangulation algorithms in geoscientific modelling". httv://neon.bbk.ac.uk 13.Lee H.G., Kim K.H., Lee K., "Development of 3-Dimensional GIs Running on Internet". IEEE, pp1046-1049, 1998. 14.Li R.X., "Data Structure and Application Issues in 3-D Geographical Information Systems". Geomatics, 48(3): pp209-224, 1994. 15.Molenaar M., "A Formal Data Structure For Three Dimensional Vector Maps", Proc 4'h International Symposium on spatial data handling, Zurich, 1990, pp830843. 16.0osterom P.V., Reactive Data Structures for Geographic Information Systems, Oxford University Press, Oxford, 1993. 17.Peter R. G., Andrew J. B. "Three dimensional representation in a Geosicientfic Resource Management System for the minerals industsy." Three-dimensional applications in Geographical Information Systems, Taylor & Francis, Edit by Jonathan F. Raper, 1989, pp155-182. 18.Pfund M. "To~ologic Data Structure for a 3D GIs". httv://www.nis.eth 19.Pigot S. "A topolog~calmodel tor a 3-dimensional spatial information system". PhD. thesis, University of Tasmania, 1998. 20.Pilouk M., Tempfli K., Molenaar M., "A Tetrahedron-Based 3D Vector Data Model for Geoinfosmation." In: Advanced Geographic Data Modelling, Netherlands Geodetic Commission, Publications on Geodesy, (40): pp129-140, 1994. 21.Raper J. "The 3-dimensional geoscientific mapping and modelling system: a conceptual design." Three-dimensional applications in Geographical Information Systems, Taylor & Francis, Edit by Jonathan F. Raper, 1989, pp11-19. 22.Stoter J. "Needs possibilities and constraints to develop a 3D cadastral registration system". .,,,, ., ., , , clelling based on surface 23.Sun M., ,, partition". Acta Geodateica et Cartographica Sinica, Vo1.29, No.3, 257-265, 2000. 24.Tempfle K. "3D topographic mapping for urban GIs". ITC journal 1998-314, 181190, 1998. 25.Turer A. K. "The role of three-dimensional geographic information systems in subsurface characterization for hydrogeological applications". Three-dimensional applications in Geographical Information Systems, Taylor & Francis, Edit by Jonathan F. Raper, 1989, pp115-127. 26.Zlatanova S., Tempfli K. "Modelling for 3D GIs: Spatial Analysis and Visualisation through web". Proc of IAPRS, Vol. X X X I I I , Amsterdam, 2000, pp1257-1264. -l.uL,,
,.,,
..,
Genetic Cryptoanalysis of Two Rounds TEA Julio C6sar Hern5ndez1, Jos6 Maria Sierra1, Pedro Tsasi2 and Arturo ~ i b a ~ o r d a '
' Computer Security Group, Computer Science Dept., Carlob Ill Univerbity,
2891 1 kgan&,Madrid, Spain {jcesar, sierra, arturo}@inf.uc3m.es Artificial Intelligence Group, Cornputer Science Dept., Carlos III University, 2891 I Legan&, Madrid, Spain {isasi}@ia.uc3m.es
Abstract. Dibtinguishing the output of a cryptographic primiti~~e such as a block cipher or a habh function from the output of a random mapping seriously affects the credibility of the primitive security, and defeat:, it for many cryptographic applicationb. However, thi:, i:, ubually a quite difficult task. In a previous work [I], a new cryptoanalytic technique waa prebented and proved useful in distinguibhing a block cipher from a random permutation in a completely automatic way. This technique i:, babed in the belection of the worst input patterns for the block cipher with the aid of genetic algorithms. The objective i:, to find which input patters generate a significant deviation of the obsen~ed output from the output we would expect from a random permutation. In [I], this technique was applied to the case of the block cipher TEA with 1 round. The much harder problem of breaking TEA with 2 rounds is successfully solved in thia paper, where an efficient distinguiaher ia also preaented.
1 Introduction A cipher [2] is an algorithm characterised by a tuple
(II,r, K, El, A) that verities:
1. rI is the finite set of plaintexts 2. r is the finite set of ciphertexts 3. K is the finite set of possible keys 4. For every K in K there is a ciphering function E, in € and a deciphering function 6, in A that verifies 6,(&,(.n))=.n for every plaintext .n d e n. T h e filndarnental property here is nurnber 4, that irnplies that the original plaintext can be recovered if the ciphering key is known. But this knowledge of the ciphering key must not only be sufficient for recovering the text, but also necessary. Any robust cipher must not allow the recovery of plaintext without the knowledge of the key used. P.M.A. Sloot et al. (Eds.): ICCS 2002, LNCS 2331, pp. 1024−1031, 2002. Springer-Verlag Berlin Heidelberg 2002
Genetic Cryptoanalysis of Two Rounds TEA 1025
Block ciphers are a special class of ciphers, characterised for operating over bit blocks instead of over single bits as the stream ciphers do. This block length (typcally 64, 128 or 256 bits) and the key length used to cipher (typically 64 or 128 bits) are fixed and fundamental for the block cipher's strength. In an ideal block cipher, every key must define a random perrnutation of the input bits over the output bits. This is why the problem of deciding if a given block cipher is secure or not can be translated into the problem of deciding if the permutations it generates when fixing keys are or not random. Unfortunately, this problem is very hard. Randomness, although being a paramount matter in security, is quite hard to evaluate. However, and independently of the techniques used, if we are able of distinguishing a cryptographic primitive from a random mapping in an efficient and statistically significant way, then this cryptographic primitive must be discarded for any cryptographic application, and immediately removed and replaced by a rnore robust algorithm. This is why the techniques for evaluating randomness have a paramount irnportance in cryptography. This paper is organised as follows: Section I is a brief introduction to the concept of block ciphers and the importance of distinguishers in cryptoanalysis. Section I1 briefly introduces the block cipher TEA, which will be the objective of our attacks. Section I11 highlights the main ideas behind what we call genetic or poisoned cryptoanalysis. Section TV presents known results of this technique over TEA with 1 round, mostly for comparison porpoises, and new results over the harder TEA with 2 rounds. Section V presents some conclusions and Section VT finally rnentions sorne selected references.
2 The TEA Algorithm TEA stands for Tiny Encryption Algorithm. It is the narne of a block cipher invented by David Wheeler and Roger Needham, mernbers of the Computer Security Laboratory of Cambridge University. It was presented in the 1994 Fast Software Encryption Workshop [3]. TEA is a very fast block cipher that does not use predefined tables or s-boxes and does not need much initialisation time. It is a Feistel type algorithm, thus narned because it divides its input in two halves and operates over thern individually in each round and interchanges them at the end of every round. It works over 64 bit blocks and uses keys of 128 bits, which are large enough for today's security standards. As they authors say, it has a security (with 8, 16 or 32 rounds) at least comparable with the Data Encryption Standard (DES), and it is quite faster. The block cipher TEA has some additional advantages: it is very robust (only one published attack [4]
1026 J.C. Hernández et al.
in more than 7 years of existence, and it is one academic attack, that is, a theoretical attack with little or no practical implications) and it is very portable, sirnple and efficient as its compact code below shows: l o n g * w,
v o i d code (long* v,
l o n g * k) {
u n s i g n e d l o n g y=v [ 0 ] ,z = v [ 1 ] ,sum=O, d e l t a = O x 9 e 3 7 7 9 ! ~ 9 , n = 8 ; while
sum y z
+= +=
(n-->O)
+=
{
delta ;
( z < < 4 )+ k [ O ]
A
z+sum
A
( z > > 5 )+ k [ l ] ;
( y < < 4 )+ k [ 2 ]
A
y+sum
A
( y > > 5 )+ k [ 3 ] ;
}
w[O]=y ; w [ l l = z ; 1
3 Cryptoanalysis with genetic algorithms Some classic statistical tests [5] are based in observing the output of a block cipher algorithm while &ding it with inputs as randorn as possible. This is because the idea behind these test is to verify if the execution of the block cipher algorithm maintains the good randomness properties of the input, so this must be as random as possible. Our proposal is based in a complete different idea. Instead of using a random input, we use what can be called poisoned inputs. We try to check if the fixing of some of the input bits simplifies the process of discovering correlations between the input and the output bits. If significant correlations exist, we would be able to infer information about some bits of the plaintext just from the o u t p ~ bits ~ t of the ciphertext. These correlations could also be used to distinguish the block cipher algorithm from a random permutation in a simple and efficient way. It is important to mention that this fixing of some bits in the input must be reasonable in length to allow for enough different inputs to mathematically verify the effect it has over the output. If a statistically significant deviation from randomness is found, then we would have an easy method for distinguishing the primitive from a random mapping and the algorithm muct be discarded for any cryptographic use. But how to decide which bite to fix in the input and to which valuec is a very complex problem. It is, escentially, a search in a huge space with 2"fK elementc. For TEA this space bitmask hac 2"' poscible values, co an exhaustive search is infeacible. For finding the beet bitmasks in those huge spaces we propose the use of genetic algorithmc in which individuals will codify bitmasks.
Genetic Cryptoanalysis of Two Rounds TEA 1027
In our method, schematically represented in Figure 1, individuals codify bitmasks that will be used to perform a logical AND with every random input. In this way, by means of an AND with a given bitmask, we manage to fix some of the input bits to zero.
Figure 1: A schema of the procedure used for fixing some of the input bits to zero by performing an AND with a bitmask.
For every bitmask (individual) in a given population of the genetic algorithm, we will observe the deviation it produces in the observed output and we will decide if this deviation is or not statistically significant. Repeating this process, we will try to find the best individuals (bitmasks) that will we those which produce more significant deviations from the expected output. We have used the implementation of the genetic algorithm of William M. Spears, from the Navy Center for Applied Research. After a number of preliminary tests, we determined that a 0.95 probability of crossover and a 0.05 probability of mutation were adequate for our problem and we decide to fix them to these values. The fitness function we try to maximise is a chi-square statistic of the observed output. We decided to observe the distribution of the 10 least significant bits of the first output word of TEA because some authors, notably [6], have shown that block ciphers that use rotation as one of their round operations (as is the case of TEA) tend to show bad distributions in their lest significant bits. So we will measure the fitness of an individual by how the bitmask that it codifies affects the probability distribution of the 10 rightmost bits of the first output word of TEA. These 10 bits can be interpreted as the binary representation of the integers between 0 and 1023, and their distribution should uniformly take all these values. For checking if this is the case, we will perform a statistical chi-square test, which is one of the most extended test in cryptoanalysis due to its high sensibility to little deviations.
1028 J.C. Hernández et al.
The statistic distribution should correspond to a chi-square distribution with 10241=1023 degrees of freedorn. The values for different percentiles of this distribution are shown in Table 1: Table 1.Values of the Chi-square distribution with 1023 D.0.F. for different percentils
p-value
0.5
0.75
0.90
0.95
0.99
X' Statistic
1022.3
1053.1
1081.3
1098.5
1131.1
Our objective will be to find bitrnasks for the TEA input (both the input block of 64 bits and the key block of 128 bits) that produce a value in the chi-square statistic as high as possible, which will irnply a strong proof of nonuniforrnity in the output distribution and, hence, a strong proof of the block cipher nonrandornness.
4 Results Every rnask (codified as individuals in the genetic algorithrn population) was evaluated by performing an AND with 2'' randorn inputs, different for every individual and every generation. This makes convergence harder, but improves the generality of the results obtained because it makes overfitting nearly impossible. An example of a bitmask for TEAl obtained using this approach is:
This bitmask has a length of 192 bits (192=64+128) and a weight of 73. This implies that, choosing input vectors at random and applying this mask over them could give us 273different inputs to the block cipher TEAl. This is a huge number, so this bitmask is useful and applicable. This would not occur if the weight of the bitmask were very low. It is clear that if two masks provoke the same deviation in the output we should prefer the heavier one because more ones in the bitmask imply more input bits that do not affect the behaviour of the observed output. The chi-cquare ctatictic we are using ac a fittzess hnction can not increace indeijnitely, but hac a maximum, as we will chow. As we make 2'"ests for every individual and any of thern can return a value between 0 and 1023 that we expect will be uniformingly distributed, the number of occurrences of every value must be close to 21"211"=2"8. The ~naxirnu~n value for the chi-square statistic under these assumptions will occur when the observed distri-
Genetic Cryptoanalysis of Two Rounds TEA 1029
bution is as far as uniform as possible, that is, when only one of these 1024 possible values really occurs and all the rest do not occur. In this case we say the distribution has collapsed in a single output value and the,fitne.r.rwill be
This is exactly the case for the bitrnask show before. It produces a collapse of all the 10 rightmost bits of the h t output word of TEAl into a single value. For assuring the generality of this bitmask, it was tested with other sets of inputs not previously seen by the generic algorithrn and in every case it produced a collapse of the output. So the bitmask shows that the 10 bits we observe do not depend uniforrningly of every input bit (those positions that have a value of I in the bitrnask do not affect the output bits we are observing), which is a property that block ciphers rnust have. So, apart from showing that there are high correlations between the input and the output of TEAI, we can also use this result for constructing a distinguisher able of deciding if a given function is TEA1 or a randorn rnapping using as input a single randorn vector. This distinguisher algorithrn can be described as:
INPUT: F:
Z
I'
->Z
"
, a random mapplng or
TEAl
ALGORITHM: Generate a random vector v of Apply the mask m gettlng v 1 = v
Zbl'^ &
m that can take 2 '
possible values Compute F(v')=w[O]w[l] Compute r
=
w[O]
&
1023
OUTPUT : If r=441 then F is T E A l else F is not T E A l It is important to note that the algorithm for distinguishing TEAl from a random mapping only needs one plaintext (input vector). This is a very unusual fact in cryptoanalysis, because most attacks and distinguishers need a large number of texts (numbers like 2" are common) to work properly. This distinguisher is able of distinguishing TEAl frorn a randorn rnapping with an extrernely low probability of false positives (only 1/1024=0.000977, less than a 0.1%) and a zero probability of false negatives.
1030 J.C. Hernández et al.
The case of TEA with 2 rounds is quite harder. An additional round signiticantly improves the strength of the algorithm, so the same approach does not produce interesting results, even after many generations and different iterations of the genetic algorithm. Different fitness functions were tested for extending this method to TEA2, and most of then were discarded. We finally got results using as a titness function not the chi-square statistic used in TEAl, but its fourth power. This has the effect of amplifying differences in the chi square statistic that had little influence in the previous fitness function. It also makes the selection procedure of the genetic algorithm, that in our implementation is proportional to fitness, closer to a selection proportional to rank. Using this approximation, we managed to obtain the following bitrnask
that produces chi-square values that vary around 1900. As this statistic value is extremely high (see Table 1 ) and the weigh of the mask is 77, this bitrnask can be used to design a distinguisher for TEA2. The construction of the distinguisher, once we have the bitmask, is trivial. The algorithm will be: INPUT: F : z " - > Z
,
a random mapplng o r T E A 2
ALGORITHM:
G e n e r a t e 2'' random v e c t o r s vL o f Z.19^ Apply t h e mask m2 t o e v e r y v e c t o r vi, &
m2 t h a t c a n t a k e 2
g e t t i n g v i l = vi
posslble values
Compute F (vLr) =wi [ 0 ] wi,,, Compute ri
=
wi,,,
&
1023
Perform a chi-square t e s t f o r checking i f t h e observed d i s t r i b u t i o n of r i i s c o n s i s t e n t with t h e expected uniform d i s t r i @ u t + o n , , c a l c u l a t i n g t h e c o r r e s p o n d i n g c h i square s t a t l s t l c * OUTPUT: I f 02>1291.44 t h e n F i s T E A 2 e l s e F i s n o t T E A 2
This is a quite interesting distinguisher in the sense that it will produce a very low ratio of false positives (a value greater than 1291.44 will only occur in a chi-square statistic with 1023 degrees of freedom one in 10-"times) and also a very low probability of false negatives (the average of the statistics produced by this bit mask is around 1900 so a value of less than 1291.442306 is extremely unlikely).
Genetic Cryptoanalysis of Two Rounds TEA 1031
It is also worthy to mention that this distinguisher uses 2" input vectors, not at all a huge number but many more than the corresponding distinguisher does for TEA1 does. This is because we must perform the chi-square test and we need enough inputs for expecting at least 5 occurrences of all the 1024 possible outputs. Slightly increasof input vectors proing this minimum of 5 to 8 leads to the actual number 23+t0=t3 posed.
4 Conclusions In this work over the use of genetic algorithms in cryptoanalysis we have shown how a new technique which is usef~ilto perform an autornatic cryptoanalysis of certain cryptographic primitives (named poisoned or genetic cryptoanalysis) can be implemented with the aid of genetic algorithms. Although this fact was previously stated in [I], it was only proven to produce results over the very limited TEAI. By showing that this model is also able of finding strong correlations in variants of the block cipher TEA with more rounds (TEA2) we have finally provided enough evidence of the interest of this technique for cryptoanalysis.
References 1. HernGndez J.C., Isasi P., and Ribagorda A.: An application of genetic algorithms to the
cryptoanalysia of one round TEA. Proceedings of the 2002 Sytnpoaium on Artificial Intelligence and ita Application. (To appear) 2. Douglas R. Stinson, Cryptography, Theory and Practice, CRC Press, 1995
3. D. Wheeler, R. Needham: TEA, A Tiny Encryption Algorithm, Proceedings of the 1995 Fast Software Encryption Workshop. pp. 97-1 10 Springer-Verlag. 1995 4. John Kelsey, Bruce Schneier, David Wagner: Related-Key cryptoanalysis of 3-WAY, Biharn.DE, CAST, DES-X, NewDES, RC2 and TEA, Proceedings of the ICICS'97 Conference, pp. 233-246, Springer-Verlag, 1997. 5. Juan Soto et. al., NIST Randomness Testing for Round 1 AES Candidates Proceedings of the Round 1 AES Candidates Conference, 1999 6. John Kelsey, Bruce Schneier, David Wagner: Mod n cryptoanalysis with applications against RCSP and M6, Proceedings of the 1999 Fast Software Encryption Workshop, pp. 139-155 Springer-Verlag, 1999.
Genetic Commerce – Intelligent Share Trading Clive Vassell Harrow Business School, University of Westminster, London Email: [email protected], URL: users.wmin.ac.uk/~vasselc
Abstract. In time, it seems feasible that genetic algorithms will help to achieve similar levels of productivity gains in many service domains as has been achieved in line and, more recently, batch manufacture. And ecommerce will embody the standards and guidelines necessary to enable managers to make the most of the possibilities offered by this development. It will help them to both satisfy and retain the custom of their clients; and make it possible to operate in a highly efficient and effective manner. The paper discusses the nature of these changes and it assesses some of their implications for organisations and their management. It introduces the concept of intelligent share trading; a future manifestation of these developments. And it talks about the important role artificial intelligence, particularly genetic algorithms, will play in these systems.
Electronic Commerce The Internet looks set to become a key plank in the infrastructure needed to support a new era of business development. The promise of a network connecting all significant economic agents of the world (both human and software) and many of the devices on which they rely creates the possibility of a huge array of information services [1]. If the Internet is, in large measure, the infrastructure which facilitates this new era, e-commerce is the 'business protocol' which determines the standards and norms by which trade is conducted in this new context. It covers such issues as electronic data interchange (EDI) between the various members of the virtual supply chain, the payment systems which are to be used and/or permitted, and the maintenance of the levels of security necessary to reassure customers and potential customers. Just as importantly, it encapsulates the understanding gleamed about what works and what doesn't in the information age; the 'strategic protocol' which leads to satisfied and loyal customers as well as robust and profitable businesses. P.M.A. Sloot et al. (Eds.): ICCS 2002, LNCS 2331, pp. 1032−1041, 2002. Springer-Verlag Berlin Heidelberg 2002
Genetic Commerce - Intelligent Share Trading 1033
It is, as yet, early days. We still have a considerable amount to learn; and we still have many tried and trusted approaches carried over from the preceding era which will need to be unlearned as they no longer apply. A few indicators of the likely form of these protocols are beginning to emerge however. First of all, the customer is going to be better placed tomorrow than he or she is today [2]. The information age will make information available to the connected world in abundance; it will no longer be the preserve of the resourceful or powerful. Information on available products, prices, relative performance, cost of production, methods of production and distribution, environmental friendliness, suitability for various tasks and/or users, and much, much more will be available to all those interested to look for it [3]. And looking for such information will become progressively easier. We suffer from information overload now not so much because there is too much information out there but rather because the information searching and filtering tools we have at present are not sufficiently sophisticated. As these tools improve, so our ability to more effectively manage large amounts of information will increase as will our ability to be selective about the theme of the information presented, its format, and its level of detail. It will become successively more difficult, indeed counterproductive, for large suppliers to misrepresent their products or services. Should they do so, an army of empowered consumers may abandon their offerings and might well petition (via the Net) industry watchdogs, MPs, consumer groups or TV programs or web sites, or any others they feel appropriate to take action against the offending firm. Furthermore, for the foreseeable future there will remain a need for effective logistics [4]. Indeed e-commerce is likely to increase the need for this and yet it is an area that is often overlooked. Even information services require facilitating goods (equipment and consumables) and the more dispersed the service provision, the more carefully the supporting logistics will have to be planned and implemented. Recently, data mining has become an area of considerable interest to organisations. Many large firms have huge data warehouses full of transaction data, management information, information on the external environment, and information on much else of potential value. Data mining approaches, including artificial intelligence, can be very useful in making sense of this mountain of data and using that understanding to improve decision making [5], [6], [7]. Artificial intelligence is being used in a wide range of applications. It is being used to better facilitate manufacturing [8], and to make intelligent agents more effective [9], and in a host of other applications and domains in between. And, inevitably, it is being used in the financial arena. Neural networks and genetic algorithms are the preferred AI environments in this area. The vast amounts of data available and the complexity of the applications make them particularly well suited to the domain. And thus they are being used for a range of financial applications, including stock market prediction [10], [11].
1034 C. Vassell
These are but a few examples of the kind of insight of this new order which e-commerce will have to encapsulate. There will be many more which will emerge over the years ahead. Organisations will have to be sensitive to the effectiveness of the initiatives they introduce and be ready to respond rapidly where and when required. This will be no easy task but if it is done well enough and soon enough, the rewards could be a place among the great and the good of the new order.
Intelligent Share Trading So how might all this work in practice? Well, one important application of e-commerce is share trading. There are many organisations which provide a service that enables customers to buy and sell shares of their choice on one or more stock exchange(s). Typically the user is required to select the shares he or she wishes to buy or sell and carry out the transaction. The system makes it possible for the user to do so simply and quickly, and typically provides up-to-date prices and some relevant news. But it does not make the choices and it does not automatically conduct transactions. In principle, however, it would be perfectly feasible for such systems to be extended to include these two additional functions. The systems could use a selection strategy to choose which shares to buy and a deselection strategy to determine which ones to sell. It could then conduct the buy and sell transactions whenever the appropriate conditions applied. The selection and deselection strategies could be specified by the user so that the system behaves in the way that the user would but does so automatically and responds to changed circumstances almost instantaneously. Alternatively, the systems could use artificial intelligence to determine appropriate selection and deselection criteria. The latter is, in essence, an intelligent share trading system. It would automatically trade on its user's behalf according to criteria it has determined; the user simply needs to set the broad performance goals (such as strong growth or modest volatility), and periodically review the performance of the portfolio and the history of transactions.
The Constituents of an Intelligent Share Trading System The main constituents of an intelligent share trading system would be the data collection module, the share trading module, the share selection module and the strategy optimisation module. The data collection module would collect share price information and store it in a database ready for processing by the share selection and strategy optimisation mod-
Genetic Commerce - Intelligent Share Trading 1035
ules. The selection module would apply the current selection and deselection strategy to determine whether any shares ought to be bought or sold at present. If so it would request that the share trading module conduct a trade. The trading module would then buy or sell the requested quantity of the required share. The strategy optimisation module would run periodically and, using historical data, determine which investment/trading strategy would have given optimal results – in line with the broad performance objectives specified by the user. This would then become the strategy to be applied by the share selection module. Genetic algorithms would be used to effect the optimisation. They are arguably one of the best tools for finding optimal or near optimal solutions given an infinite or very large range of potential solutions. In fact the genetic algorithms would operate on two levels. On one level they would find the optimum strategy given a set of attributes; on the second level they would find the set of attributes which yield the best strategy given a larger universe of attributes. (This would help to ensure that the system comes up with both the optimal set of share attributes on which to base many of the better strategies, and the optimum strategy itself). The system would be designed to devise strategies which were robust or, in other words, optimised for say ten years (rather than 'over optimised' for a short period of time and/or a particular set of market circumstances). This should mean that the strategy selected would not need to change often unless the user changed the overall performance objectives.
Strategy Optimisation Module Share Selection Module Share Trading Module Data Collection Module Fig. 1. The main components of an intelligent trading system
Genetic Algorithms and Optimisation Genetic algorithms apply the principles of natural selection to finding optimum solutions to a problem. They operate on a population of potential solutions and apply the principles of selection, crossover and mutation to produce a new generation of candidate solutions.
1036 C. Vassell
The selection operation is used to choose the best candidates from the population by testing each solution against a fitness or target function. The crossover operator is used to produce a child solution from two parent solutions by combining elements of the chromosomes of one parent with elements of the chromosomes of the other. The mutation operator introduces an element of randomness into each population. The combination of these three operators leads to new generations of solutions which tend to improve in performance in relation to previous generations but which will not simply converge on sub-optimal solutions, in much the same way as they help living organisms to thrive.
The Nature of the Optimisation So what might the optimisation dialogue look like? It would have screen displays similar to some of the screens taken from one of the models I am currently using for my research in this area. I shall use this model to give an example of how this part of the intelligent share trading system might look and feel. An important component of the Strategy Optimisation Module of the intelligent share trading system would be the performance summary screen. An example of how it might look is shown below. The summary would indicate what the current share selection strategy was and how that strategy would have performed in the past. In fig. 2 no strategy has been selected. This is indicated by the fact that the chromosome values are all zero.
Fig. 2. The strategy performance summary screen
Genetic Commerce - Intelligent Share Trading 1037
Fig. 3. The strategy performance summary screen with a strategy selected
In fig. 3, the selection strategy is: Select those shares where the share price has risen by 10% or more in the previous year and the previous year's dividend has been 5% or more of the (prevailing) share price. It is important to understand that this selection strategy would have been applied at the start of each of the six years represented in the data set. And the size and performance of the resulting share portfolio for each of the six years for the subsequent three months, six months and one year are recorded at the top of the screen. The three ‘in sample’ rows provide summaries of the in sample data (in this case years one, three and five). The three ‘out of sample’ rows provide summaries of the out of sample data (years two, four and six) and the three overall summary rows provide summaries of all the data (years one to six). The target cell is also shown (the in sample minimum of the one year means). The target function is to maximise the value of this cell providing the associated count (the in sample minimum of the counts of the shares selected for each year) is at least ten. The target function is defined in the screen below:
1038 C. Vassell
Fig. 4. Defining the target function The Test attribute (or field) is used to determine which shares in the database meet the current share selection criteria. The contents of the Test attribute can perhaps best be explained if it is laid out in a logical format. This is done below. Here we can see how the genetic algorithm operates at the two levels spoken about earlier.
The Contents of the Test Attribute =And ( Or (Cppr3m=0, And (PriceRisePrev3m<>"", Or (Cppr6m=0, And (PriceRisePrev6m<>"", Or (Cppr1y=0, And (PriceRisePrev1y<>"", Or (Cpe=0, And (PERatio<>"", Or (Cyld=0, And (DividendYield<>"", Or (Ccap=0, And (EstCapital<>"", Or (Cepsgr=0, And (EPSgrowth<>"", Or (Cprofgr=0, And (Profitgrowth<>"", Or (Cdivgr=0, And (Dividendgrowth<>"", Or (Ctogr=0, And (Turnovergrowth<>"", )
PriceRisePrev3m>=Cppr3m) ), PriceRisePrev6m>=Cppr6m) ), PriceRisePrev1y>=Cppr1y) ), PERatio<=Cpe) ), DividendYield>=Cyld) ), EstCapital>=Ccap) ), EPSgrowth>=Cepsgr) ), Profitgrowth>=Cprofgr) ), Dividendgrowth>=Cdivgr) ), Turnovergrowth>=Ctogr) )
Where a chromosome has a value of zero, that chromosome plays no part in selecting shares. It is, in effect, switched off. Where it has a non-zero value, it is used in the selection process. Thus the genetic algorithm will, at any one time, be selecting which chromosomes are to be part of the selection criteria and which are not. Where a chromosome has a non-zero value, the Test attribute selects those shares which have both a non-blank value in the appropriate attribute and the value of the attribute meets the criteria level of the associated chromosome.
Genetic Commerce - Intelligent Share Trading 1039
In principle, this means we can have a large number of chromosomes but that, at any moment in time, the genetic algorithm will generally be using a few of them only. It means we should be able to test the effectiveness of a wide range of chromosome permutations while at the same time testing a range of selection strategies using each of those permutations. And we should be able to do this without having to build a whole suite of models. (I think this principle may well reduce the likelihood of 'overfitting' in a relatively straightforward manner. Whether this particular application of the principle is likely to work remains to be seen.)
Conclusions It should be noted that my main concern when trying to identify suitable strategies is robustness. In other words, I am interested in strategies which give good returns year in, year out rather than strategies which give high mean returns but are accompanied by high levels of volatility. This is why my target function is the minimum one year mean (rather than the mean of all years). I am particularly interested in strategies which provide positive returns in all test years (both in sample and out of sample). Perhaps in part because of the quite specific nature of my requirements, the results to date have not been entirely encouraging. I have not, so far, come across any strategy which meets these requirements (though there are one or two which show modest losses in one year only). However the work continues. I plan to extend the range of candidate chromosomes and possibly introduce both maximum value and minimum value chromosomes to see whether this improves the performance of the best strategies. While the results of the exercise are not yet terribly encouraging, there is some evidence to suggest that the search should prove fruitful in the end [12] and there is a prominent theoretical framework which is compatible with this kind of expectation [13], [14], [15].
Wider Implications We are entering a new era in the history of commercial activity. The understanding which helps to crystallise this new era will be the body of knowledge which e-commerce will comprise. It will be the content of the MBAs of the online business schools of tomorrow and will enable executives to profitably steer their firms in the decades ahead. The real winners however will probably not have learned much of what is critically important about e-commerce from business schools but rather by being the first to live them and learn from them in their own organisations [16]. And by being sufficiently capable as managers to capitalise extensively on the lead they so gain.
1040 C. Vassell
It is organisations like these that will fashion the new era; and it is those who study these organisations that will identify the e-commerce protocols associated with this novel form. Intelligent share trading is an example of this kind of development. It results from a fusion of electronic commerce, artificial intelligence and online trading. These systems are likely to prove profitable for their suppliers and their users alike. And the organisations who are the first to introduce robust and user friendly examples will in all probability do very well indeed. And these systems will inevitably make extensive use of artificial intelligence. The huge amounts of historical data which can be called upon to facilitate understanding and the complexity of the application areas will make the use of data mining techniques and tools very valuable. And neural networks and genetic algorithms will likely prove extremely pertinent. And genetic algorithms (and possibly hybrid approaches) will be of particular value in situations where it is necessary to both carry out appropriate actions on behalf of the users and explain to the users the underlying strategy behind those actions. So important might this kind of artificial intelligence become that perhaps genetic commerce will be the most appropriate way to describe systems of the type outlined in this paper. Indeed its popularity might well signal the next phase of the incessant rise of the machine.
References 1. Aldrich Douglas F, 'Mastering the Digital Marketplace: Practical strategies for competitiveness in the new economy', John Wiley, 1999 2. Hagel John, Armstrong Arthur G, 'Net Gain: Expanding markets through virtual communities', Harvard Business School Press, 1997 3. Evans Philip, Wurster Thomas S, 'Getting Real About Virtual Commerce', Harvard Business Review, November-December 1999 4. Jones Dennis H, 'The New Logistics: Shaping the new economy', Ed: Don Tapscott, Blueprint to the Digital Economy: Creating wealth in the era of e-business, McGraw Hill, 1998 5. Gargano Michael L, Raggad Bel G, 'Data Mining – A Powerful Information Creating Tool', OCLC Systems & Services, Volume 15, Number 2, 1999 6. Lee Sang Jun, Siau Keng, 'A Review of Data Mining Techniques', Industrial management and Data Systems, Volume 101, Number 1, 2001 7. Bose Indranil, Mahapatra Radha K, 'Business Data Mining – A Machine Learning Perspective', Information & Management, Volume 39, Issue 3, December 2001 8. Burns Roland, 'Intelligent Manufacturing', Aircraft Engineering and Aerospace Technology: An International Journal, Volume 69, Number 5, 1997 9. Saci Emilie A, Cherruault Yves, The genicAgent: A Hybrid Approach for Multi-Agent Problem Solving', Kybernetes: The International Journal of Systems & Cybernetics, Volume 30, Number 1, 2001
Genetic Commerce - Intelligent Share Trading 1041
10.Wittkemper Hans-Georg, Steiner Manfred, 'Using Neural Networks to Forecast the Systemic Risks of Stocks', European Journal of Operational Research, Volume 90, Issue 3, May 1996 11.Back Barbo, Laitinen Teija, Sere Kaisa, 'Neural Networks and Genetic Algorithms for Bankruptcy Predictions', Expert Systems With Applications, Volume 11, Issue 4, 1996 12.Bauer Richard J, ‘Genetic Algorithms and Investment Strategies’, John Wiley & Sons, February 1994 13.Fama Eugene F, French Kenneth R, ‘The Cross-Section of Expected Stock Returns’, The Journal of Finance, Volume 47, Number 2, June 1992 14.Fama Eugene F, French Kenneth R, ‘Size and Book-to-Market Factors in Earnings and Returns’, The Journal of Finance, Volume 50, Number 1, March 1995 15.Fama Eugene F, French Kenneth R, ‘Multifactor Explanations of Asset Pricing Anomalies’, The Journal of Finance, Volume 51, Number 1, March 1996 16.Senge Peter M, 'The Fifth Discipline: The art and practice of the learning organisation', Business (Century/Arrow), 1993
Efficient Memory Page Replacement on Web Server Clusters Ji Yung Chung and Sungsoo Kim Graduate School of Information and Communication Ajou University, Wonchun-Dong, Paldal-Gu Suwon, Kyunggi-Do, 442-749, Korea {abback, sskim}@madang.ajou.ac.kr
Abstract. The concept of network memory was introduced for the efficient exploitation of main memory in a cluster. Network memory can be used to speed up applications that frequently access large amount of disk data. In this paper, we present a memory management algorithm that does not require prior knowledge of access patterns and that is practical to implement under the web server cluster. In addition, our scheme has a good user response time for various access distributions of web documents. Through a detailed simulation, we evaluate the performance of our memory management algorithms.
1 Introduction With the growing popularity of the internet, services using the world wide web are increasing. However, the overall increase in traffic on the web causes a disproportionate increase in client requests to popular web sites. Performance and high availability are critical for web sites that receive large numbers of requests [1, 2, 3]. A cluster is a type of distributed processing system and consists of a collection of interconnected stand-alone computers working together. Cluster systems present not only a low cost but also a flexible alternative to fault tolerant computers for applications that require high throughput and high availability. Processing power was once a dominant factor in the performance of initial cluster systems. However, as successive generations of hardware appeared, the processor decreased its impact on the overall performance of the system [4]. Now, memory bandwidth has replaced the role of the processor as a performance bottleneck. The impact of networking has also decreased with the 100Mbps ethernet. Thus, efficient memory management is very important for overall cluster system performance. This work is supported in part by the Ministry of Information & Communication in Republic of Korea (“University Support Program<2001>” supervised by IITA). This work is supported in part by the Ministry of Education of Korea (Brain Korea 21 Project supervised by Korea Research Foundation). P.M.A. Sloot et al. (Eds.): ICCS 2002, LNCS 2331, pp. 1042−1050, 2002. Springer-Verlag Berlin Heidelberg 2002
Efficient Memory Page Replacement on Web Server Clusters 1043
The concept of network memory was introduced for the efficient exploitation of main memory in a cluster. Network memory is the aggregate main memory in the cluster and can be used to speed up applications that frequently access large amounts of disk data. This paper presents a memory management algorithm that always achieves a good user response time for various access distributions of web documents. The remainder of the paper is organized as follows. Section 2 presents related work on cluster memory management. In section 3, we explain the clustered web server architecture that is considered in this paper. Section 4 explains our memory management algorithm and section 5 presents simulation results. Finally, section 6 summarizes with concluding remarks.
2 Related Work Recently, some papers on the memory management of the cluster have studied the method of utilizing the idle client's memory [5, 6, 7]. The active client forwards cache entries that overflow its local cache directly to the idle node. The active client can then access this remote cache until the remote node becomes active. However, in these methods, a client's request must go to the disk if the requested block does not exist in the limited memory, even though another node has that block. The Greedy Forwarding algorithm deals with the memory of the cluster system as a global resource, but the algorithm does not attempt to coordinate the contents of this memory [8]. The main problem with this policy is that global memory is underutilized because of duplication. Duplicate Elimination [9] takes the other extreme approach. Since it is inexpensive to fetch a duplicate page from remote memory, compared to a disk input/output (I/O), every duplicate page is eliminated before a single page. Each node maintains two LRU (Least Recently Used) lists, one for single pages and the other for duplicate pages. The advantage of Duplicate Elimination is that it has a high global hit rate because of the global memory management. However, a main drawback of Duplicate Elimination is that the local hit rate of some nodes reduces because of the smaller cache size, even if the global hit rate increases. In order to adjust the duplication rate of the data page dynamically, the N-chance Forwarding algorithm forwards the last copy of the page from one server to a randomly chosen server, N times, before discarding it from global memory [8]. Also, the Hybrid algorithm dynamically controls the amount of duplication by comparing the expected cost of an LRU single page and an LRU duplicate page [9]. The expected cost is defined as the product of the latency to fetch the page back into memory and a weighting factor that gives a measure of the likelihood of the page being accessed next in memory. The Hybrid algorithm has a good response time on average, but it does not have a minimum response time for a special range of workload and node configuration. In this paper, we present a memory management algorithm that does not require prior knowledge of access patterns and that is practical to implement under the web
1044 J.Y. Chung and S. Kim
server cluster. Also, this method always has a good user response time for various access distributions of web documents.
3 Web Server Cluster Architecture In order to handle millions of accesses, a general approach adopted by popular web sites is to preserve one virtual URL (Uniform Resource Locator) interface and use a distributed server architecture that is hidden from the user. Thus, we consider the architecture of cluster system that consists of the load balancer and a set of document servers. Each of the document servers is a HTTP (Hyper Text Transfer Protocol) server. Figure 1 presents the web server cluster architecture. In this architecture, the load balancer has a single, virtual IP (Internet Protocol) address and request routing among servers is transparent. Every request from the clients is delivered into the load balancer over the internet. After that, the load balancer redirects the request to a document server in a round-robin manner.
y
j
p V p
y
y
o
~GG
sG
Fig. 1. Web server cluster architecture
Web documents that are serviced in the web server cluster are distributed on the disks of each node. Nodes that receive a user request are called the primary, and nodes that store the requested data on the disk are called the owner. This means that each node can be the primary or the owner as the case may be. The owner nodes maintain a directory in which they keep track of the copies of the data pages they own in global memory. The only times a node has to be informed about status changes to the pages are when a page becomes a duplicate after being the only copy in global memory and when the page becomes the last copy in the node's memory after being a duplicate.
Efficient Memory Page Replacement on Web Server Clusters 1045
4 Memory Management of Web Server Cluster Efficient memory management is the task of keeping useful data closer to the user in the memory hierarchy. Figure 2 shows the memory hierarchy that is considered in this paper. The user request is divided into page-sized units and is serviced by the primary node. If the requested page exists in the primary node, it is serviced immediately. If the page is absent, the primary node requests the owner node for the page. If the page presents in the owner node, it is forwarded to the primary node. Otherwise, the owner node forwards the request to the other node that has the requested page. The node receiving the forwarded request sends the data directly to the primary node. However, if no node contains the page, the request is satisfied by the owner's disk.
o
w G
X
jw| jw|
Y
j j
Z
t t
[
vG
t t t t
\ ]
yG
t t
k k
Fig. 2. Cluster memory hierarchy
When the owner node reads the page from disk, it retains a copy of the page in local memory. In order to improve global memory utilization, these pages are maintained on a FIFO (First In First Out) list and should be removed from the memory of the owner node as soon as possible. The page replacement policy is to minimize the number of possible page faults so that effective memory access time can be reduced. The LRU algorithm is a popular policy and often results in a high hit ratio in a single server. This policy assumes that the costs of page faults are the same. However, this assumption is not valid in the cluster system, since the latency to fetch a page from disk and the latency to fetch a page from the remote node are different. Therefore, the elimination rate of the duplicated page in the global memory has to be higher than that of a single page. Our proposed policy, DREAM (Duplication RatE Adjustment Method) for memory management of the web server cluster adjusts the duplication rate according to the various access distributions of web documents. The page replacement method of DREAM is as follows.
1046 J.Y. Chung and S. Kim
IF (Ws
Cs
<
Wd
Cd
)
Replace the LRU single page ELSE Replace the LRU duplicate page Where Ws and Wd are the inverse of the elapsed time since the time of last access for the LRU single page and LRU duplicate page, respectively. Cs is the latency to fetch a page back into memory from disk and Cd is the latency to fetch a page back into local memory from the remote memory. is a parameter for duplication rate adjustment. Ws, Wd, Cs and Cd are positive and Cd is lower than Cs. Thus Cs /Cd is always higher than 1. In order to observe the impact of the parameter , let's consider the behavior of DREAM with = 0. In this case, the ELSE statement is always performed since Ws Cs is positive. This means that DREAM is Duplicate Elimination when is equal to 0. Also, when is equal to 1, DREAM is the same as the Hybrid algorithm. If becomes Cs / Cd , Ws and Wd are just compared at the IF statement. This means that DREAM is a Greedy Forwarding algorithm that only considers time information. Thus, we can see that Duplicate Elimination, Greedy Forwarding and Hybrid are special cases of DREAM. When < 0, the right term of the IF statement is negative. This case is the same to of 0. When 0 < < 1, the elimination rate of the duplicated page is between Duplicate Elimination and Hybrid. When 1 < < Cs / Cd, the elimination rate of the duplicated page is between Hybrid and Greedy Forwarding. Finally, when > Cs / Cd, there is no meaning since the elimination rate of a single page becomes higher than that of a duplicated page. Therefore, the valid range of parameter is 0 Cs / Cd. Figure 3 shows the relation for the elimination rate of duplicated page and parameter .
lG XWWL k l
W
\WL n m
o ¥
X
¥
jVj
Fig. 3. The relation for the elimination rate of duplicated page and
In general, the access frequencies of web documents closely resemble a Zifpian distribution [10]. In this distribution, the shape of the curve is decided by the skew parameter. When skew is 0 the access frequencies of web documents are uniform, and when the skew parameter is 2 the access frequencies of web documents are skewed.
Efficient Memory Page Replacement on Web Server Clusters 1047
The algorithm that has the best response time among Duplicate Elimination, Greedy Forwarding and Hybrid is decided by the skew parameter. DREAM has the best response time by adjusting parameter according to the skew parameter. Also, it does not have an additional overhead for improving performance.
5 Performance Evaluation As indicated by the title, this section is devoted to a description of the simulation results which we obtained using our memory management for web server cluster. Table 1 presents the simulation parameters and the Simscript .5 process oriented simulation package is used to model the system.
Table 1. Simulation parameters
Parameter Number of nodes Memory per node Number of files File size Page size Message cost Network bandwidth Disk bandwidth
Value 3 64MB 384 512KB 8KB 1ms / page 15MB / sec 10MB / sec
Figure 4 shows the response time as a function of skew. At low skew, Greedy Forwarding is worse than the other algorithms, but the response time decreases drastically as skew increases. To the contrary, Duplicate Elimination has a good response time at low skew, but the improvement of response time is little even if skew increases. Hybrid has a response time that is close to the minimum at both low and high skews, but it is not the best solution at all skews. DREAM eliminates duplicate pages in a similar way to Duplicate Elimination at low skew and it duplicates the hot pages in a similar way to Greedy Forwarding at high skew. Thus, it has always the best response time even though skew increases. In Figure 5, the aggregate web document size that will be serviced is 2 times the size of global memory. In this case, Greedy Forwarding has the worst response time even though skew is high. Also, the data size of Figure 6 is 0.5 times the size of global memory. In this case, Duplicate Elimination has the worst response time at every skew because each node has enough local memory. When the local memory is enough, the duplication of hot pages improves performance by eliminating network overhead.
1048 J.Y. Chung and S. Kim
Fig. 4. Skew : Response time, (Number of files = 768)
Fig. 6. Skew : Response time, (Number of files = 192)
Fig. 7. Skew : Local hit rate
Fig. 5. Skew : Response time, (Number of files = 384)
Fig. 8. Skew : Global hit rate
Fig. 9. Skew : Disk I/O rate
Efficient Memory Page Replacement on Web Server Clusters 1049
Hybrid has a good response time on the average, but it does not have the minimum response time at all skews. On the contrary, we can see that DREAM has a best response time by adjusting according to the skew. Figure 7 and Figure 8 show the average local hit rate and global memory hit rate, respectively. In Figure 8, global hit rate of Greedy Forwarding is lower than other algorithms because disk I/O rate is high at the low skew and local hit rate is low at the high skew. In addition, Figure 9 shows the disk I/O rate. Greedy Forwarding performs frequent disk access at the low skew but has the best local memory hit rate at the high skew. Duplicate Elimination has the worst local memory hit rate but has a good global memory hit rate. DERAM has the best response time by optimizing the ratio of the local memory hit rate, global memory hit rate and disk access rate.
6 Conclusion Cluster systems are emerging as a viable architecture for building high performance and high availability servers in application areas such as web services or information systems. In initial cluster systems, processing power was a dominant factor of the performance, but memory bandwidth has replaced the role of the processor as a performance bottleneck. Thus, efficient memory management is very important for the overall cluster system performance. In this paper, we proposed an efficient memory management algorithm under the web server cluster. In addition, simulation results show that DREAM always achieves a good user response time for various skew parameters.
References 1. Du, X., Zhang, X.: Memory Hierarchy Considerations for Cost-effective Cluster Computing. IEEE Transactions on Computer (2000) 915-933 2. Cardellini, V., Colajanni, M., Yu, P.S.: Dynamic Load Balancing on Web-server Systems IEEE Internet Computing (1999) 28-39 3. Zhu, H., Yang, T., Zheng, Q., Watson, D., Ibarra, O.H., Smith, T.: Adaptive Load Sharing for Clustered Digital Library Servers. Proceedings of the Seventh IEEE International Symposium on High Performance Distributed Computing (1998) 28-31 4. Buyya, R.: High Performance Cluster Computing: Architectures and Systems. Prentice-Hall (1999) 5. Feeley, M. et. al.: Implementing Global Memory Management in a Workstation Cluster. In Proceedings of the 15th ACM SOSP (1995) 6. Venkataraman, S., Livny, M., Naughton, J.: Impact of Data Placement on Memory Management for Multi-Server OODBMS. In Proceedings of the 11th IEEE ICDE (1995) 7. Koussih, S., Acharya, A., Setia, S.: Dodo: A User-Level System for Exploiting Idle Memory in Workstation Clusters. 8th IEEE International Symposium on High Performance Distributed Computing (1999) 8. Dahlin, M., Wang, R., Anderson, T., Patterson. D.: Cooperative Caching: Using Remote Client Memory to Improve File System Performance. In Proceedings of the First Symposium on Operating Systems Design and Implementation (1994)
1050 J.Y. Chung and S. Kim 9. Venkataraman, S., Livny, M., Naughton, J.: Memory Management for Scalable Web Data Servers. 13th International Conference on Data Engineering (1997) 10. Zipf, G.: Human Behavior and the Principle of Least Effort. Addison-Wesley (1949) 11. Guchi, M., and Kitsuregawa, M.: Using Available Remote Memory Dynamically for Parallel Data Mining Application on ATM-Connected PC Cluster. 14th International Parallel and Distributed Processing Symposium (2000)
Interval Weighted Load Balancing Method for Multiple Application Gateway Firewalls B. K. WOO', D. S. ~ i n i ' S. , S. ~ o n g 'K. , H. ~ i n i ' and , T. M.
hung'
' ~ e a l - ~ i mSystems e Laboratory, School ot Electrical and Computer Engmeenng, SungKyunKwan University, Chon-chon dong 300, Chang-an gu, Suwon, Kyung-k~do, Republ~cof Korea {bkwoo, dsklm, sshong, byraven, tmchung}ertlab. skku.ac.kr
Abstract. Firewalls are inslalled a1 nelwork perimelers lo secure organiza~ion's network as alternatives to general gateways. Because of potential performance problems on the gateways, load balancing technique has been applied. However, co~npareclto general gateways, firewalls require more intelligent load balancing method to process massive network traffic because of their relatively complex operalions. In this paper, we analyze [he inherenl problems of existing simple load balancing methods for firewalls and propose the interval weighted load balancing (IWLB) to enhance the processing of massive network traffics. The IWLB deals with network traffics in consideration of the characteristics of application protocols to achieve more effective load balancing. We observed that the IWLB outperforms other simple load balancing methods during our simulation. Therefore, we expect that the IWLB is suitable to balancing loads for multiple firewalls at a network perimeter.
1 Introduction While the explosive growth of the Internet made it possible to exchange massive information, it caused some negative effects such as the increase of network traffics and security threats. Thus, organizations have a burden to deal with massive network traffic and protect their network froni any malicious security threats. The typical solution to counteract various security threats is deploying the firewall which applies policy-based access control to network traffic at a network perimeter [5, 6, 71. However, a single firewall cannot operate properly when massive network traffics are applied but also become the dangerous security hole of networks [S]. Furthermore, as shown froni DDoS attacks, such as the Trinoo, the Nimda worm, and so forth, using niassive packet deliveries or service requests, these attacks degrade network services and performance and create serious operational failures, as well [I 2, 131. To make firewall more robust and stable with massive network traffics, it is inevitable to install multiple firewalls to distribute network traffics by using various load P.M.A. Sloot et al. (Eds.): ICCS 2002, LNCS 2331, pp. 1051−1060, 2002. Springer-Verlag Berlin Heidelberg 2002
1052 B.K. Woo et al.
balancing methods. Thus, the performance and fairness of a load balancer is the critical factor to estimate the network performance in the environment above. Although there are many existing load balancing methods based on different principles such as Round robin, Hashing, and Randomization, most of them have flaws to be deployed for firewalls because they do not consider the characteristics of application protocols as the parameter for load balancing. In this paper, we propose the interval weighted load balancing (IWLB) method, which is designed to use the characteristics of application protocols as the parameter for load balancing decision, to overcome the critical problems of existing load balancing methods. Furthermore, the IWLB is able to enhance network performance and give more robustness in the environment that multiple firewalls are inevitable. With the analysis of simulation results, we observed that the IWLB outperfornls other load balancing methods when it is deployed as the principle for the load balancing to distribute traffics. This paper is organized as follows. In chapter 2, we issue the potential problems of existing load balancing methods when they are applied for multiple firewalls. In chapter 3, we introduce the principle and mechanism of the IWLB method. Our simulation model is described in chapter 4, and the analysis of simulation results and the coniparison to other load balancing methods are presented in chapter 5. At last, we close this paper with conclusions in chapter 6.
2 Simple Load Balancing Methods Under the environment that multiple firewalls are installed in a network for the load balancing purpose, it is desirable to build an appropriate load balancing principle to guarantee the performance of the individual firewall and network security. The most important factor for successful load balancing is to distribute service requests fairly to multiple firewalls. However, the fairness must be defined differently in this environment because a firewall could discard incoming service requests in terms of its access control policy. That is, even though the load balancer of multiple firewalls distributes incoming service requests fairly, the active session distribution on multiple firewalls can be distributed unevenly. The existing load balancing inethods show the limit to overcoine this problem because they do not regard the characteristics of application protocols as the parameter for fair load balancing. In the following sections, we issue the critical flaws of most widely used load balancing methods when they are deployed as the load balancer of multiple firewalls.
Round Robin. Round robin method distributes incoming service requests by simply allocating them to the next available firewall in the rotational manner. That is, this method does not consider the load or number of active sessions currently allocated to the individual firewall. Thus, it can cause the potential load imbalance and increase the skew when a specific application requires long service time [I].
Interval Weighted Load Balancing Method 1053
Hashing. Hashing method distributes incoming service requests by hashing the inforination of service request packets such as source address, source port number, and so on [I]. However, when considering that the flooding attacks generally occur with the identical source information, Hashing method can be vulnerable to those attacks. Moreover, when deployed Hashing inechanisni is exposed to attackers, the overall network can be plunged into the fatal situation. We strongly suggest not using Hashing nlethod for the load balancer of multiple firewalls. Randomization. Randomization method distributes requests to each node according to the value of pseudo random number [ I ] . In this method, the tine algorithni for random number generation is the key to the successful load balancing. Like other methods mentioned above, Randomization method does not consider the characteristics of application protocols or sessions. Therefore, it is hardly expected that this method is suitable for the load balancer of multiple firewalls.
3 The Interval Weighted Load Balancing As mentioned in chapter 2, sinlple load balancing nlethods are not suitable for the fair load balancing of multiple application gateway firewalls, because they do not take the clxiracteristics of application protocols in consideration. Thus, it is necessary to consider the characteristics of application protocols or sessions as the parameter in order to design more stable and efticient load balancer. The proposed load balancing method named the interval weighted load balancing (IWLB) makes use of the weight value allocated to each application protocol. In the IWLB, the weight of an application protocol is defined as the interval in the order of firewalls in order to decide which firewall will process a current incoming request. That is, the IWLB decides the firewall for the cursent service request by adding the weight of the application protocol to the order of the previously selected firewall for the previous service request that has the same protocol as the current one. Since the initial value of the weight, based on the former research about the traffic pattern of application protocols [14], is assigned to each application protocol, the IWLB keep track of the weight value by calculating it periodically in a statistical manner. To give a specific example how the IWLB obtains tlie weight value of each protocol, let's suppose the following situation. There are 8 firewalls and all of them can serve 4 different application protocols: HTTP, FTP, SMTP, and TELNET. And let's the weight of HTTP assigned by the IWLB is 3 at this moment. If 6 HTTP requests arrived at the IWLB load balancer sequentially, the order of firewalls to service these requests will be Fwl+ Fwd+ Fw,+ Fw2+Fw5+Fw8. That is, the IWLB decides the firewall to service these requests by adding tlie weight value of HTTP to the previously selected firewall in a rotational manner. Fig. 1 depicts this example and difference from the conventional round robin method.
1054 B.K. Woo et al.
Fw, 3 Fw, 3 Fw, 3 Fw, 3 .... Fw1
FW,
3 FW, 3 FW, 3 FW, 3 .... Fw1 I
I
(a) Round robin
(b) IWLB (weight = 3)
Fig. 1. The cornparison of [he seleclion sequence for Round robin and the IWLB
The IWLB uses the standard weight value to calculate the weight value of each application protocol, the standard weight value is set to 1 and assigned to the application protocol that has the largest average service time. After deciding tlie standard weight value, tlie weight values of other application protocols are decided as the ratio of their average service time to the average service time of the application protocol that has the standard weight value. Finally, the calculated ratio values must be rounded to a nearest integer value to be used as the weight values for application protocols. Table 1 explains the rule that the IWLB generates the weight values of application protocols. Tablc 1. The average service time and weight of application protocols
Application protocol
Average service time
Calculation rule
Weight
HTTP FTP SMTP TELNET
970 3090 790 430
Whrrp= 30901970 k 3.18 Wh,, = 1 (standard weight) W,,,,, = 3090/790 k 3.91 W,,(,,, = 3090/430 + 7.18
3 1 4 7
In Fig. 2, we depict how the IWLB balances the service requests with the example of Table 1. Load Distribution of each application protocol starts at the first firewall, i.e., Fw,. On the other hand, if the total number of firewalls is divisible by the weight value of a certain protocol, all of the incoming requests of the protocol would be assigned to the same firewall. To prevent this phenonlenon, the weight for each application protocol must be the prime to the total number of firewalls.
Interval Weighted Load Balancing Method 1055
Fig. 2. The distribution of incoming requests using the IWLB
4 Modeling and Simulation
I
Queuing system
I
1 : Arrival rate 1,: Arrival rate at each firewall 1, : Service rate at each firewall b, : Blocking rate A ( t ) : Number of arrivals up to time t R(t) : Number of rejections up to time t D f t ) : Number of departures up to time t Nit) : Number of connections at time t
..............................
Fig. 3. The modeled load balancer and multiple firewalls
It is widely known that TCP-based application protocols occupy the large portion of the Internet traffic. McCreary announced that some TCP-based application protocols such as HTTP, FTP, SMTP, and TELNET occupy 42.52%, 2.59, 1.70%, and 0.13%,
1056 B.K. Woo et al.
of the overall Internet traffic, respectively in his recent research [14]. From this fact, we can deduce easily that the performance of a proxy firewall is directly influenced by TCP-based application protocols. To build our simulation model, we chose 4 representative TCP-based application protocols: HTTP, FTP, SMTP, and TELNET. In our model, the load balancer distributes service requests to a group of firewalls modeled by the queuing system. To conipare the performance of existing load balancing methods to that of the TWLB, we include 3 other load balancing nietliods, Round robin, Hashing, Randonlization, in our simulation. Fig. 3 depicts the simulation model of the load balancer and multiple firewalls. Let R be a set of service requests, which is generated with the arrival rate A, arrived at the load balancer of multiple firewalls. We suppose that the inter-arrival time between two adjacent service requests and service durations of requests are exponentially distributed with the arrival rate 2 and the service rate p, respectively. Additionally, each element of R contains its application protocol for the TWLB. After service requests are distributed to a group of firewalls by various load balancing niethods, each firewall processes the allocated service requests. When we suppose that i n firewalls installed in our model, the summation of the service requests allocated to each firewall equals the R if there is no blocking of requests at the load balancer. Therefore, as shown in the equation ( I ) , the summation of the arrival rate at each firewall equals the arrival rate at the load balancer.
,=I
The service request allocated to a firewall can be blocked in terms of its access policy. If we assume that the blocking rate, b,, on service requests at Fw,denoted as the it11 firewall, then the rate of service requests processed by Fw,, ,Ii,would be defined as the equation (2)
If service requests are blocked by the access policy of a firewall unpredictably, the load of firewalls will be distributed unevenly irrelevant of the fair distribution of service requests by the load balancer. Since no simple load balancing methods are able to cope with this situation, the fluctuation of workload between firewalls is inevitable. In the TWLB, if a service request is blocked, the firewall signals to a load balancer. When the load balancer receives the signal from the firewall, it allocates the next service request to the firewall once more to prevent the fluctuation of the load between firewalls. We regarded each firewall as an M/M/c/c queuing system independently with the same capacity. The capacity of itli tirewall, C , means the maximum nuniber of active sessions that the firewall can handle with concurrently. If the nuniber of active sessions exceeds the C , then the allocated service request to firewall will be queued for a later service.
Interval Weighted Load Balancing Method 1057
5 Simulation Result and Analysis In the simulation, we assume that the load balancer interacts with 4 firewalls. We generated 200,000 service requests to monitor the fair distribution of service requests and applied 3 simple load balancing methods and the IWLB to the load balancer in our model. During the simulation, we supposed that the propagation delays from the load balancer to firewalls are ignorable. For the analysis of simulation results, we monitored the summation of service time and the waiting time of service requests in the queue at each firewall every second. If we analyze these values between firewalls, we would judge whether the applied load balancing methods distributed the service requests optimally to firewalls. Firstly, we compared the response time of the IWLB to that of other simple load balancing methods. Fig. 4 depicts the maximum response time of Hashing, Randomization, Round robin, and the IWLB, respectively. We sampled the response times of each firewall every 1 second and select the maximum response time among the sampled value. The graph shows that the maximum response time of the IWLB is remarkably lower than others. Additionally, the fluctuation of its curve is relatively narrower than that of other simple methods. Note that we put the results of each load balancing method together in a graph for comparison.
+Randomization
- Hashing
Roung robin
-
WLB
1
ti
Fig. 4. Maximum response time
For more sophisticated comparison, we calculated the mean response time of each load balancing method and the mean response time of each method is depicted in Fig. 5. While the mean values of Round robin, Randomization, Hashing methods are not much different each other, those of the IWLB are quite different from them. The IWLB shows very lower mean response time during our simulation. Now, from the two results of the IWLB, we can judge that the IWLB outperformed other simple load balancing methods.
1058 B.K. Woo et al.
20.00
1 1
& Rando~zation
-
41
21
Hashina
-BS-
61 Round robin
81 A
WLB
1
time
Fig. 5. Mean response time
In Fig. 6, we calculated the variance of the response time of each load balancing method. In this figure, we can see that the variance of the IWLB is lower than others, too. According this, we can conclude the response time of the IWLB is more stable than that of other simple methods.
IRandomization
IHashing
r Round robin
IIWLB
1
time
Fig. 6. The variance of response time
From the comparison, we can deduce two facts. One is that the distribution of service requests by simple load balancing methods causes the skewed load distribution among firewalls because they do not consider the blocking of requests in terms of the access policy of firewalls. The other is that the load balancing principle of the IWLB, considering the characteristics of application protocols, shows the positive effect for the fair traffic distribution. As Table 2 shows, we compared the mean and maximum buffer size of each firewall. In the case of the IWLB, the mean and maximum buffer sizes are greatly smaller
Interval Weighted Load Balancing Method 1059
than those of other simple methods. That is, the TWLB balances loads most fairly among simulated methods and can reduce the memory resource of firewalls. Note that the mean values of buffer size are calculated to the second decimal place. Tablc 2. The co~nyarisonof the buffer size of firewalls
Round robin Randomization Hashing TWLB
8.12 0.14
During the analysis of simulation results, we observed that the TWLB outperformed other simple load balancing methods in many aspects. Moreover, the simulation results explain that the TWLB proved that it is able to cope with massive traffic loads adequately. Consequently, it is strongly required to consider the weight of application protocols for fair load balancing to counteract massive service requests. Furthermore, when the load balancer interacts with firewalls, it should be able to manage the request blocking by the firewall to prevent the fluctuation of workloads between them. We expect that the TWLB meets to these requirements successfully.
6 Conclusion and Further Studies Although many organizations are deploying firewalls for the purpose of network security, it is doubtable that they can process massive network traffics without performance degradation. To make it worse, considering the trends of preferring an application gateway firewall that perfornis more sophisticated operations, it is obvious the performance degradation will be more serious for massive network traffic. Because the performance degradation or malfunction of firewalls implies the critical security flaws, it is strongly required to make the firewall more tolerate against massive network traffic. Several researches paid attention to deploying multiple firewalls and load balancers to counteract massive network traffic. It was unsuccessful to manage them because of the inherent drawbacks of existing simple load balancing methods for firewalls. In this paper, we proposed the enhanced load balancing method, the IWLB, to manage overloaded network traffic efficiently. Since the IWLB makes use of the weight values of application protocols, calculated by the statistical traffic pattern of application protocols, for the load distribution, it is optimized for the load balancing method for multiple application gateway firewalls. As shown in our simulation results, we observed that the IWLB outperformed other simple load balancing methods on the load distribution of massive TCP-based application service requests. From these results, we expect that the IWLB would be suitable for the load balancer for the network that deploying multiple application gateway firewalls is inevitable.
1060 B.K. Woo et al. At this moment, we are planning to extend our evaluation model to various application protocols and performance measurement to various aspects such as resource usage, packet loss, and s o on. Additionally, we will evaluate the scalability and tolerance of the TWLB when some firewalls are not functioning properly.
References 1. Rajkuinar, B.: High Performance Cluster Computing: Architecture and Systems, Volume 1, Prentice Hall PTR, (1 999) 2. Leon-Garcia, A,: Probability and Random Process for Electrical Engineering, 2nd Ed., Addison Wesley Publishing Company, Inc., (1 994) 3. Molloy, K.M.: Funclarnentals of Perforrnance Modeling, Macrnillan Publishing Cornpany, (1 989) 4. Law, M.A., Kelton, W.D.: Simulation Mocleling & Analysis 2nd eel., McGraw-Hill Book Co., (1991) 5. Cheswick , R.W., Bellovin, MS.: Firewalls and Internet Security : repelling the willy hacker, Aclclison Wesley, (1 994) 6. Chapman, D.B., Zwicky, D.E.: Building Internet Firewalls, 0 Reilly & Associations, Inc., ( 1 996) 7. Hare, C., Siyan, K.: Internet Firewalls anel Network Security - 2nd eel., New Readers, (1996) 8. Kostic, C., Mancuso, M.: Firewall Performance Analysis Reporl, Computer Sciences Corporation, Secure Systems Center - Network Security Department, (1 995) 9. Haeni, E. R.: Firewall Penelralion Tesling, The George Washington Universily, Cyberspace Policy Institute, (1 997) 10 Test Final Report - Firewall Shootout Networkcl+Interop, KeyLabs Inc., 28 May 1998. 11. Foundry ServerIron Firewall Load Balancing Guide, Foundry Nelworks, Inc., (2001) 12. Carnegie Mellon University, CCERT Advisory CA-2001-26 Nirnda Worm, CERTICC, h~tp:llu~ww.cert.ors/~1c1~~isorieslCA-2001-26.htd, (2001) 13. Carnegie Mellon Universily, CERT Incidenl Nole IN-99-07: Distribu~edDenial of Service Tools, CERTICC, http://www.cert.org/incic1ent~noteslIN-O-O7.ht1nl, (1 999) 14. McCreary, S., Claffy, K.: Trends in wide area IP traffic palterns - A view from Atnes Inlernet Exchange, Proceedings of 13th ITC Specialist Seminar on Internet Traffic Measurement and Modeling, Monlerey, CA. 18-20, (2000)
Modeling and Performance Evaluation of Multistage Interconnection Networks with Nonuniform Traffic Pattern* Youngsong Mun1 and Hyunseung Choo 2 1
School of Computing, Soongsil University, Seoul, KOREA mun@computing. ssu.ac.kr 2 School of Electrical and Computer Engineering Sungkyunkwan University, Suwon 440-746, KOREA [email protected]
Abstract Even though there have been a number of studies about modeling MINs, almost all of them are for studying the MINs under uniform traffic which cannot reflect the realistic traffic pattern. In this paper, we propose an analytical method to evaluate the performance of ATM switch based on MINs under nonuniform traffic. Simulation results show that the proposed model is effective for predicting the performance of ATM switch under realistic nonuniform traffic. Also it shows that the detrimental effect of hot spot traffic on the network performance turns out to get more significant as the switch size increases.
1 Introduction Since ATM has been adopted as a standard for broadband ISDN, many research efforts have been focused on the design of the next generation switching systems for ATM. The three main approaches employed for the design of an ATM switch are shared medium, shared memory, and space-division architecture [1]. In all these designs, the limitation on the switching size is the primary constraint in the implementation. To make a larger size ATM switch, thus, more than one system is interconnected in a multistage configuration [2]. Multistage interconnection networks (MINs) [3] constructed by connecting simple switching elements (SEs) in several stages have been recognized as an efficient interconnection structure for parallel computer systems and communication systems. There have been a number of studies investigating the performance of MINs in the literature [4-8]. However, almost all of these previous works are for studying the MINs under the uniform traffic pattern. Nonuniform traffic reflects the realistic traffic pattern of currently deployed integrated service network where a wide range of bandwidths needs to be accommodated. Therefore, the performance of the MINs under nonuniform traffic must be studied for obtaining efficient switch-based system. Even though
* This work was supported by Brain Korea 21 Project. P.M.A. Sloot et al. (Eds.): ICCS 2002, LNCS 2331, pp. 1061-1070, 2002. © Springer-Verlag Berlin Heidelberg 2002
1062 Y. Mun and H. Choo there have been some models considering nonuniform traffic patterns [5,7], they are not precise enough since the performance of the models has not been verified. In this paper, we propose an analytical method to evaluate the performance of ATM switch based on MINs under nonuniform traffic. It is mainly achieved by properly reflecting the nonuniform dispatch probability in modeling the operation of each switch element. To evaluate the accuracy of the proposed model, comprehensive computer simulation is performed for two performance measures - throughput and delay. MINs of 6 and 10 stages with buffer modules holding single or multiple cells are considered for evaluation. As nonuniform traffic pattern, hot spot traffic of 3.5% and 7% are investigated. Comparison of the simulation data with the data obtained from the analytical model shows that the proposed model is effective for predicting the performance of ATM switch under realistic nonuniform traffic. The detrimental effect of hot spot traffic on the network performance turns out to get more significant as the switch size increases. For example, the throughput is about 0.3 for 6-stage switch with 3.5% hot spot traffic, while it becomes only about 0.03 for 10-stage switch.
2 The Proposed Model
2.1 Assumptions, Buffer States, and Definitions In our models, 2x2 switching elements with the buffer modules of size m are used, and a network cycle consists of two phases. The sending buffer modules check the buffer space availability of the receiving buffer modules in the first phase. Based on the availability (and routing information) propagated backward from the last stage to the first stage, each buffer module sends a packet to its destination or enters into the blocked state in the second phase. In each network cycle packets at the head of each buffer module (head packets) in an SE contend with each other if the destinations of them are same. Based on the status of the head packet, the state of a buffer module can be defined as follows. Figure 1 shows the state transition diagram of a buffer module in SEs. • State- 0 : a buffer module is empty. • State- rik : a buffer module has k packets and the head packet moved into the current position in the previous network cycle. • State- bk : a buffer module has k packets and the head packet could not move forward due to the empty space of its destined buffer module in the previous network cycle. The following variables are defined to develop our analytical model. Here Q(ij) denotes the y'-th buffer module in Stage-z. And its conjugate buffer module is represented asQ(ijc). Also t/, represents the time instance when a network cycle begins, while ^represents the duration of a network cycle. • m : the number of buffers in a buffer module. • n : the number of switching stages. There are n = log2 N stages for NxN MINs.
Modeling and Performance Evaluation of Multistage Interconnection Networks 1063 • Pa{ij ,t) I P{ij ,k): the probability that Q(ij) is empty/not full at tj,. • Pn {ij, i): the probability that Q(y) is in State- nk at tb , where 1 < k < m . • Pb {ij, i): the probability that Q(ij) is in State- bk at tb , where 1 < k < m . m
m
• SPn {ij, t): £ P„k {ij, t)
• SPb {ij, t): £ \
k=\
(.it, 0
k=\
• Pbu{ij,t) IPlb{ij,i): the probability that a head packet in Q(ij) is a blocked one and destined to the upper/lower output port at tb .
0
0
Figure 1. The state transition diagram of the proposed model. • r{ij)lrx{ij,t) : the probability that a normal/blocked head packet in Q(y) is destined to the upper output port. • q{ij, t): the probability that a packet is ready to come to the buffer module Q(y). • rn{ij,t)Irb{ij,t): the probability that a normal/blocked packet at the head of Q(ij) is able to move forward during td . • r"{ij,t) Irln{ij,t): the probability that a normal packet at the head of Q(ij) can get to the upper/lower output port during td . • rb {ij,i) Irlb{ij,t): the probability that a blocked packet at the head of Q(ij) can get to the upper/lower output port during td . • r"n{ij,t) Irlnn{ij,t) : the probability that a normal packet at the head of Q(ij) can get to the upper/lower output port during td by considering Q(ijc) in either State- n or State- b . If Q(ijc) is in State - b , it is assumed that the blocked packet is destined to the lower/upper port (so no contention is necessary). r • "b(y^) IrnbQJ't) '• t n e probability that a normal packet at the head of Q(ij) is able to get to the upper/lower output port during td by winning the contention with a blocked packet at the head of Q(ijc).
1064 Y. Mun and H. Choo
•
r
bn (y»0 / rL (?/»0 : the probability that a blocked packet at the head of Q(ij) is able
to move forward to the upper/lower output port during td . Here it is assumed that Q(if) is empty or in the State- n . •
r
bb(v'•><) l'rlbb(i)!,t) : the probability that a blocked packet at the head of Q(ij) is
able to move forward to the upper/lower output port during td . Here it is assumed that Q(if) also has a blocked packet. •
Pna(ij,t) I Pba(ij,t) I Pbba(ij,t)
: the probability that a buffer space in Q(ij) is
avaible (ready to accept packets) during td , given that no blocked packet/only one blocked pakcet/two blocked packets in the previous stage is destined to that buffer. •
X"(ij,t)
IXln(ij,t)
: the probability that a normal packet destined to the up-
per/lower output port is blocked during td . •
Xb(ij,t)
IXlb(ij,t)
: the probability that a blocked packet destined to the up-
per/lower output port is blocked during td . • T(ij, i) : the probability that an input port of Q{ij) receives a packet.
2.2 Calculations of Required Measures
2.2.1 Obtaining
rn{ij,t)
A normal packet in an SE is always able to get to the desired output port when the other buffer module is empty or destined to a different port from it. When two normal packets compete, each packet has the equal probability to win the contention. The probability that a normal packet in Q{ij) does not compete with a blocked packet in the other buffer module is r(i/)\l-rx(i/c,t)}
+ \l-r(i/)}rx(i/c,t).
Therefore, the prob-
abilities r"„(ij,t) is as follows and rlnn{ij,t) is obtained similarly. C (H, 0 = r(ij)Pa {if, t) + [0.5r(u)r(if) + r(ij) {1 - r(ij c)} ]SP„ (if, t) +
(1)
r(ij){l-rx(jf,t)}SPb(jf,t)
r
"b(U't) a n d r «i(')»0 a r e the probabilities that a normal packet has the same destination as the blocked one in the other buffer module and wins the contention. Thus they are as follows: runb (ij, t) = 0.5r(ij)rx (if, t)SPb (if, t),
(2)
The probability that a buffer module is not full (P(ij, i)) is simply ~mT) = l-P
(ij,t)-PK
(ij,t).
(3)
Modeling and Performance Evaluation of Multistage Interconnection Networks If the originating buffer module of a packet is in State -bt (\
1065
then the des(2
<m).
If it has received a packet in the previous network cycle, it can be in State - rij (1 < j < m ) or State -bk(2
^P„k
+
{l-T(ij,t-l)}
Pb (ij ,t)rb(ij ,t) bm " .b" '. PbJvS)
(4)
m—\
(<J,t) + ^Pbk
Here A = -^
(Jj,t) + P„m (i/,t)rn(ij,t) + Pbm
(i/,t)xrb(i/,t)
^ 1-P0(ij,t)-Ph(ij,t)
The probabilities rn {ij, t) and rb {ij, t) will be discussed later in this section. Pna {ij, t) is obtained similarly. If the destined buffer module has not received a packet, it must be in any state except State- nm . Then Pna {ij, t) = T{ij, t -1) x A + {1 - T{ij, t-\)}xB m— 1
m—1
Po(i/,t) + ^Pnt Here B =
(5)
(<J,t) + ^Pbk
^
W,t) + Pbn
{ij,t)xrb{ij,t)
^
.
For a packet to move to the succeeding stage, it should be able to get to the desired output port and the destined buffer module should be available. Thus r"{ij,t) is as follows and r'n{ij,t) is obtained similarly. < (ij, t) = C (ij, t)Pna ({i + \),t) + runb (ij, t)Pba ((«+1), t)
(6)
So rn(ij,t) is r„(ij,t) = r%(ij,t) + rln(ij,t).
(7)
We can calculate rb(ij,t) using the similar method. 2.2.2 Obtaining Xun(ij,t), X^(ij,t)
X"b(ij,t),
Xln(ij,t),
X!b(ij,t),
and
rx(ij,t)
is the probability that a normal packet destined to the upper output port is
blocked. Xun(ij,t) = C(ij,t){l-Pna((i c
+ \),t)} + r^b(ij,t){l-Pba((i c
c
+ \),t)} c
+ 0.5r(ij)r(jj )SPn (ij , t) + 0.5r(ij )rx (ij , t)SPb (ij , t)
(8)
1066 Y. Mun and H. Choo The first two terms in the equation above are the probabilities that the destination has no available space. The last two terms are for the case of lost contention. Xb(if,t)
is
the probability that a blocked packet destined to the upper output port is blocked again. We can calculate this probability easily by the approach emplyed in X% (ij, t). Xub (ij, t) = C(if, t){1 - Pba ((«+1),0} + r^b(if, t){1 - Pbba{{i +1), t)} c
c
c
(9)
c
+ 0.5rx (ij)r(ij )SPn (ij , t) + 0.5rx {ij)rx (ij , t)SPb (ij , t) x'n(ij,t)
and x'b(if,t)
are obtained similarly.
Also rx{ij,t), which is the probability that a blocked head packet is destined to the upper output port, is calculated as follows. rx{ijJ)=
.
4(y ?
' ~1)
(i£(iM-l) + ^(i/,f-l)*0)
Pbh{ij,t-\) + Pbl{ij,t-\) Here Pb (if, t) and Plb (ij, t) are calculated as follows. Pb" (ij, t) = X"„ (ij, t)SP„ (ij, t) + X"b (ij, t)SPb (ij,t),
(11)
H (ij, t) = X'„ (ij, t)P„ (ij, t) + X{ (ij, t)Pb (ij, t).
(12)
2.2.3 Obtaining T(ij, t) and q(ij, t) Due to its inherent connection property of MINs, the two buffer modules in an SE are connected to either upper or lower output ports of the SE of the previous stage. On the contrary, the buffer modules below it are connected to the lower output ports. We denote T(ij, t) for the buffer modules connected to upper output ports as T(if,t) = SPn((i-l)g,tX((i-l)g,t) u
+ SPb((i-l)g,t)rb ((i-l)g,t)
+ SPn((i-l)gc,tX((i-l)gc,t) c
+ Pb((i-l)g ^((i-l^,t)
(13) '
The buffer modules which are connected to lower output ports of the previous stage are obtained similarly. T(ij, t) (\
(l4)
is obtained. T(if,t) p
p
(ti> 0 + nm W, t)r„ (if, t) + Pb (if, t)rb (if, t)
(15)
Modeling and Performance Evaluation of Multistage Interconnection Networks
1067
2.2.4 Calculating r{ij, t) r{ij) is calculated by using the transformation method proposed in [7]. It is a mapping scheme that transforms the given reference pattern into a set of riij) 's which reflect the steady state traffic flow in the network. For a steady state reference pattern, we represent it in terms of destination accessing probabilities Aj, the probability that a new packet generated by an inlet chooses the output port j as its destination. Then riij) can be represented as the conditional probability that the sum of Aj 's which are connected to the upper output port of Q{ij) given the sum of Aj 's of all possible destined output ports which are connected to the upper or lower output port of Q{ij). For example, riij) 's in three stage MIN are described as follows. For the last stage: r (31)
= r(32)=
A
- ™ _ -iyj-t; ™ _
r(35) = r(36) =
) = r(38)
+ A6
A i A 3 + A4
A A7
+A
For the second stage: r(21) = r(22) = r(25) = r(26) = r(23) = r(24) = r(26) = r(27):
A + A2 + A A2 + A3 + A4 ' A+A As+As+Aj+Az
For the first stage, all r{\j) (1 < j < N ) are same: r(ll) = r(12) =
= (rl8) = Al + A2+ A3+ A4 + A5 + A€+ A7 + A%
2.3 Throughput and Delay Normalized throughput of a MIN is defined to be the throughput of an output port of the last stage. If Port-j is the upper output port of an SE, the normalized throughput in this port is as follows. TNET{j, t) = SP„ (nj, t)run (nj, t) + SP„ (nf, t)run ( « / ,t)
(1 6 )
+ SPb {nj, t)r^ {nj, t) + SPb {njc, t)rg {njc, t) The delay occurred for a packet at the buffer module Q{ij) in the steady state is calculated by using Little's formula. m 2^k{Pnt{ij,t) + Pbt{ij,t)\ D{ij)= lim 1=1 T{ij,t)
(17)
1068 Y. Mun and H. Choo As delay at each output port are different, the weight of it should be considered for obtaining the mean delay. Hence the mean delay is "
(18)
7=1
Here w - the weight of Port-j for the mean delay - is obtained by the rate of the normalized throughput of that port as follows. w i = lim
„
TNET(j)
(19)
y\TNET{k,t) k=\
3 Verification of the Proposed Model Correctness of our model in terms of network throughput and delay is verified by comparing them with the data obtained from computer simulation for various buffer sizes and traffic conditions. For the simulation, 95% confidence interval is used and the following approaches are employed for the computer simulation. • Each inlet generates requests at the rate of the offered input traffic load. • The destination of each packet follows the given hot spot nonuniform traffic pattern. Here each inlet makes a fraction h of their requests to a hot spot port, while the remaining (l-hP"a(ij,t)) of their requests are distributed uniformly over all output ports including the hot spot port. • If there is a contention between the packets in an SE, it is resolved randomly. • The buffer operation is based on the FCFS principle. Figure 2 shows the mean throughput and delay comparison of a 6-stage single buffered MIN with 7% the hot spot traffic. The offered traffic load varies form 0.1 to 1, and simulation data are obtained by averaging 10 runs. In each run, 1,000,000 iterations are taken to collect reliable data. The variations in the last 100,000 iterations are less than 0.1%. Figure 3 shows the comparison of the throughput of the hot spot port and other ports between the analytical model and computer simulation in this case. It reveals that the throughput of the hot spot port is more than two times higher than that of other ports since the access probability to the hot spot port is higher than others. Also Figures 4 and 5 show the comparison results of multiple buffer MINs. In case of uniform traffic, more buffer entries can increase the performance of MINs 10% to 20%. As identified here, in case of the nonuniform traffic, the increase in the throughput is as small as about 2% even though more buffers are added since blocking among the packets is more likely due to the nonuniform traffic. Similar result are shown in case of the 3.5% hot spot traffic. The figures show that our models are effective for predicting the performance of MINs with realistic traffic. In case of the large sized MIN (1024x1024), the throughput of the hot spot port is always close to 1 since there always exists a packet to that
Modeling and Performance Evaluation of Multistage Interconnection Networks
1069
port coming from a large number of input ports. However, those of other ports are as low as less than 0.03 since blocking is so severe. Analytical Model
0
Simulation
Analytical Model
V
Simul
0.8 [ 0.6 V 0.4 V 0
0.2
0.4 0.6 Input Load
0.S
0.4
0.7 Input Load
Figure 2. Comparison of the throughput and mean delay for 6-stage, single-huffered MIN delay with 7% hot spot traffic. Analvtical Model
Ik-
1 0.8 a
0.6
1 li
0
0.2
0.4 0.6 Input Load
0.!
a) Hot spot port h) Other ports Figure 3. Comparison of the throughput of hot spot port and other ports with 7% hot spot traffic. •Analytical Model
0
Simulation
60 S
40
S
20
1 fcatfW
ol 0.2
0.4 0.6 nput Load
0.S
0.4
0.7 nput Load
1
Figure 4. Comparison of the throughput and mean delay for 6-stage 4-buffered MIN with 7% hot spot traffic. Analvtical Model
5 _g a
1 0.8 0.6
S
0.2
f
li
1 0.8 a
r
0'
0
0.2
0.6
0.4 0.6 Input Load
0.!
0
nlll W'"W'"W"1B—ill—ill—ip—ip—•I' 0.2
0.4 0.6 Input Load
0.8
a) Hot spot port b) Other ports Figure 5. Comparison of the throughput of hot spot port and other ports with 7% hot spot traffic.
1
1070 Y. Mun and H. Choo
4 Conclusion This paper has proposed an analytical modeling method for the performance evaluation of MINs under nonuniform traffic. The effectiveness of the proposed model was verified by computer simulation for various practical MINs; 6x6 and 10x10 switches, single and 4-buffered MIN with 3.5% and 7% hot spot traffic. According to the results, the proposed model is accurate in terms of throughput and delay. The detrimental effect of hot spot traffic on the network performance turns out to get more significant as the switch size increases. For example, the throughput is about 0.3 for 6-stage switch with 3.5% hot spot traffic, while it becomes only about 0.03 for 10-stage switch. Therefore hot spot traffic needs to be avoided as much as possible for especially relatively large size switches. Performance analysis of other structures such as gigabit ethernet switches and terabit routers, or MINs for optical switching networks under nonuniform traffic are underway.
References 1. Hyoimg-IL Lee, Seimg-Woo Seo and Hyuk-jae Jang. "A High performance ATM Switch Based on the Augmented Composite Banyan Network," IEEE International Conference on Communications, Vol.1, pp.309-313, June 1998. 2. Muh-rong Yang and GnoKou Ma, "BATMAN : A New Architectural Design of a Very Large Next Generation Gigabit Switch," IEEE International Conference on Communications, Vol.2/3, pp.740-744, May 1997. 3. K. Hwang, Advanced Computer Architecture: Parallelism, Scalability, Programmability. New York: McGraw-Hill, 1993. 4. Y. C. Jenq, "Performance analysis of a packet switch based on single buffered Banyan network," IEEE J. Select. Areas Commun, vol. SAC-3, pp. 1014-1021, Dec. 1983. 5. H. Kim and A. Leon-Garcia, "Performance of Buffered Banyan Networks Under Nonuniform Traffic Patterns," IEEE Transaction on Communicationis, Vol. 38, No. 5, May 1990. 6. Y. Mun and H.Y. Youn, "Performance Analysis of Finite Buffered Multistage Interconnection Networks," IEEE Transaction on Computers, pp. 153-162, Feb. 1994 7. T. Lin and L. Kleinrock, "Performance Analysis of Finite-Buffered Multistage Interconnection Networks with a General Traffic Pattern", ACM SIGMETRICS Conference on Measurement and Modeling of Computer Systems, San Diego, CA, pp. 68-78, May 21-24, 1991. 8. H.Y. Youn and H. Choo, "Performance Enhancement of Multistage Interconnection Networks with Unit Step Buffering," IEEE Trans, on Commun. Vol. 47, No. 4, April 1999.
Real-Time Performance Estimation for Dynamic, Distributed Real-Time Systems Eui-Nam Huh1, Lonnie R. Welch2, and Y. Mun3 1
Sahmyook University, Department of Computer Science Chongyang, P.O.Box 118, Seoul, Korea [email protected] 2 Ohio University, School of Electrical Engineering & Computer Science Athens, Ohio, USA [email protected] 3 Soongsil University, School of Computer Science Seoul, Korea [email protected]
Abstract. The main contribution of this paper is accurate analysis of real-time performance for dynamic real-time applications. A wrong system performance analysis can lead to a catastrophe in a dynamic real-time system. In addition, real-time performance guarantee combined with efficient resource utilization is observed by experiments, while the previous worst-case approaches primarily focused on performance guarantee but resulted in typically poor utilization. The subsequent contribution is schedulability analysis for a feasible allocation of resource management on the Solaris operating system. This is accomplished with a mathematical model and by accurate response time prediction for a periodic, dynamic distributed real-time application.
1 Introduction The rac-25 radiation machine killed cancer patients because of a software bug. Imagine that you have a car accident. After you have already been thrown into the windshield, the airbag inflates. This paper addresses the problem of certifying that software will respond to real-world events in a timely manner. Use of real-time systems is being spread rapidly to many areas. Real-time services need to react to dynamic user requests. A dynamic real-time system used to offer services to the dynamic user requests should consider variable execution times and/or arrival rates of tasks during run-time. One of the management problems of dynamic real-time systems that must be solved is to provide the quality of service (QoS) for a time-constrained task. The task commonly appears in systems such as air traffic control, robotics, automotive safety, mission control, and air defense. An example of this task is software that detects radar data, evaluates it and launches missiles if a hostile missile is detected. All of these dynamic real-time systems require consideration of variable execution times and/or arrival rates of tasks, as opposed to deterministic and stochastic real-time systems which have a priori known, fixed task execution times and arrival rates. P.M.A. Sloot et al. (Eds.): ICCS 2002, LNCS 2331, pp. 1071−1079, 2002. Springer-Verlag Berlin Heidelberg 2002
1072 E.-N. Huh, L.R. Welch, and Y. Mun
Resource allocation, mapping software (S/W) to hardware (H/W), for those systems is an essential component and needs to consider real-time constraints of the task and should analyze feasible resources. Thus, the feasible allocation of the available resources has to be carried out according to time-constrained requirements, failure of which might cause a catastrophe in hard real-time systems. To maintain the feasible allocation, shedulability analysis of the task is required. Prediction of the response time of the task compared to the time-constrained requirement is one of the schedulability analysis approaches. Furthermore, computing resources for those dynamic systems should be utilized efficiently. The distributed system is employed to provide scalable resources to the dynamic S/W systems that have various and unbounded execution times. Generally, the real-time system is designed to analyze that a task can meet its time constraint before it is executed. The Rate Monotonic Analysis (RMA) introduced by Liu and Layland in [1] is used primarily to determine schedulability of an application by using a priori Worst-Case Execution Time (WCET) and the priority of the application. The priority of the application to be applied to RMA is dependent upon arrival patterns and rates. The application, which has a higher arrival rate of task or needs to be executed more frequently, has a higher priority level than any other applications. However, as has been noted in [2], [3], [4], [5], and [6], resources are poorly utilized if the average case is significantly less than the worst case. Another drawback of RMA is that it cannot efficiently accommodate high-priority jobs that have relatively low rates. It must, however, be noted that RMA can be made to work in such cases, by transforming low-rate, high-priority jobs into high-rate jobs -- but this can be extremely wasteful in terms of resources. It is stated in [6], [7] and [13] that accurately measuring the WCET is often difficult, and is sometimes impossible. Puschner and Burns in [8] consider WCET analysis to be hardware dependent, making it an expensive operation on distributed heterogeneous clusters. The statistical RMA by Atlas and Bestavros in [9] considers tasks that have variable execution times and allocate resources to handle the expected case. The benefit of this approach is the efficient utilization of resources. However, there are shortcomings. Firstly, applications which have a wide variance in resource requirements cannot be characterized accurately by a time-invariant statistical distribution; and secondly, deadline violations occur when the expected case is less than the actual case. Similarly, real-time queuing theory by Lehoczky in [5] uses probabilistic event and data arrival rates for performing resource allocation analysis. On the average, this approach provides good utilization of resources. It must be noted, however, that applications which have a wide variance of resource requirements cannot be characterized accurately by a time-invariant statistical distribution. Called the dynamic real-time system by Welch and Masters in [10], there is a need for a new approach to the dynamic real-time system which would be more efficient than RMA and statistical RMA in assurance of the real-time QoS.
Real-Time Performance Estimation for Dynamic
1073
2 Problem Statement This section uses mathematical notation to describe the problem of certifying that a real-time application meets (or does not meet) its real-time requirement on a host that is running other applications. The notation will be used in subsequent sections to concisely define certification methods. For convenience, the Data Flow Diagram (DFD) is used for description of problems.
H, A Resource Allocator
Schedulability Analyzer SA-result
Fig. 1. Level 0 DFD of “Schedulabilty Analyzer System”
As shown in Fig.1, the Schedulability Analyzer certifies (or says that it cannot certify) that the application (‘A’) will meet its real-time requirement on the host (H). The application ‘A’ is represented as period, priority, execution time, segment and real-time requirement. The host ‘H’ is represented as its name, time quanta, and a list of applications. SA-result is returned to indicate the certification result, which consists of a boolean value and a predicted response time. A, H
Resource Manager
1 Predict Response Time
λreq λpred
SA-result
2 Check Schedulability
Fig. 2. Level 1 DFD of “Schedulability Analyzer”
1074 E.-N. Huh, L.R. Welch, and Y. Mun
The Schedulability Analyzer from Fig. 1 is depicted in Fig. 2. This paper presents a novel approach to schedulability analysis, which predicts response time (λpred) of ‘A’ and checks schedulability to certify that ‘A’ will meet its real-time requirement (λreq) on H before allocation. ‘Predict Response Time’ reads application list (‘A*’) on H and computes λpred of ‘A’. ‘Check Schedulability’ compares λpred with λreq and returns ‘true’ if λpred < λreq and returns ‘false’ otherwise; the predicted response time; λpred, is also returned. ‘A’, H and the elements of A* are defined as tuples below. Definition 2.1: An application tuple A = (T, p, C, U, S, λreq ) where T is period of A , p is priority of A, C is execution time of A, U is utilization of A , S is the number of time quanta used by A per period. (it is computed as C/TQ . ) λreq is the requirement of A. (note: we assume that λreq ≤ T) Definition 2.2: A host tuple H=(name, TQ, A*) where name is name of a host, TQ is a vector defining the time quantum for each priority on a host A* is a list of the applications:
Definition 2.3: Each element of A* is an application, represented as a tuple = (T, p, C, U, S ): where T is period of A , p is priority of A, C is execution time of A, U is utilization of A , and S is the number of time quanta used by A per period.
The convention used in this paper to denote an element ‘E’ of tuple ‘T’ is E(T). Thus, for example, the execution time of an application ‘A’ is denoted as C(A). The novel approach to schedulability analysis, prediction of response time, from Fig. 2 is described in Fig. 3 in detail. The prediction of response time technique considers estimation of queuing delay of ‘A’ due to contention with other applications (‘A*’) on H. An important consideration of prediction of response time is estimation of queuing delay due to the same priority (p) applications (Dpred1) as shown at 1.1 bubble in Fig.3. Queuing delay due to higher priority (p) applications (Dpred2) as shown as bubble 1.2 in Fig. 3 is also estimated as higher priority tasks always hold resources of H. Finally, Calculate response time of the application as shown as bubble
Real-Time Performance Estimation for Dynamic
1075
1.3 in Fig. 3 computes the estimated response time of ‘A’: λpred =C + Dpred1 + Dpred2, where Dpred1: delay experienced by a due to waiting for ai ∈ ‘A*’ such that p(ai)=p(A), and Dpred2: delay experienced by waiting for aj ∈ ‘A*’ such that p(aj) > p(A).
1.1 Calculate queuing delay: same priority
A, H
Dpred1
1.3 Calculate Response time of the application
Dpred2 1.2 Calculate queuing delay: higher priority
λpred
Fig. 3. Level 2 DFD of “Predict Response Time”
3 Real-Time Performance Analysis This section presents approaches for performing estimation of real-time performance of dynamic real-time applications introduced as bubbles, 1.1, 1.2, and 1.3 as shown in Fig. 3. These approaches based on probabilistic techniques are extended to be applicable to time-sharing round-robin scheduler, which considers queuing delay due to the same priority tasks. As mentioned in section 1, the worst-case response time analysis approaches, [11] and stochastic approaches, [5] and [9] are not appropriate for response time analysis in dynamic real-time systems. 3.1 A Execution Rate (ER) Technique This approach calculates queuing delay of ‘A’ due to the same priority tasks, and considers applications’ periods to find total contention among them. Least Common Multiple (LCM) of periods of ‘A’ and ‘A*’ on H is used in this approach. To extend
1076 E.-N. Huh, L.R. Welch, and Y. Mun
this approach to round-robin scheduler, the execution of ‘A’ is analyzed in the unit of time quantum (TQ) and segment (S). Following steps show how it is processed to compute Dpred1(A) due to the same priority tasks. Step 1. Compute the resource requirement of ‘A’ during the interval, [0,LCM]. Step 2. Compute the resource requirement of the other tasks A* during the interval, [0,LCM]. Step 3. Compute executive rate at which requests of ‘A’ are serviced. Step 4. Compute queuing delay Dpred1(A).
3.2 A Probabilistic Rate (PR) Technique for Same Priority Tasks An improved approach called PR calculates queuing delay of ‘A’ due to the same priority tasks for the bubble 1.1 as shown in Fig. 3 by considering the probability of the target task (‘A’) being blocked by any other tasks (‘A*’). The basic concept is the same as ER introduced in section 3.1. The improved probabilistic rate, pr(A), considers all possible probability that the target task could be executed including computation of progress rate within contention with other tasks as follows: (1) the probability of being with no usage of the resource, (2) the chance that only ‘A’ uses the resource, and (3) the chance that there is progress rate of ‘A’ within contention. This approach also implemented like the ER approach.
3.3 Response Time Prediction This section simply combines C(A), Dpred1(A), and Dpred2(A) to compute the predicted response time of the task, λpred(A) as shown as the bubble 1.3 in Fig. 3. That is, λpred(A) = C(A) + Dpred1(A) + Dpred2(A).
4. Experiments and Summary This section shows how the approaches accurately predict response time of an application using DynBench, path based s/w systems, introduced by Welch and Shirazi (see [12]). The ‘filter1’ application (which filters noise from sensed data) in the first sensing path (denoted as “path 1”) as the task is examined on the variable size of workload scenario. Thirty samples of the response time of ‘filter1’ are collected and averaged, when workloads of the path 2 are changed dynamically and monotonically. The workload of sensing path 3 is fixed by 800, and that of sensing path 2 is suddenly increased to 1300 by adding 800 workloads at 1 minute 20 seconds; it drops by 600 at 2 minutes 40 seconds; it adds 700 again at 4 minutes, and drops 500 at 5 minutes 20 seconds. See Fig. 4.
Real-Time Performance Estimation for Dynamic
1077
385
361
337
313
289
265
241
217
193
169
145
121
97
73
49
25
1600 1400 1200 1000 800 600 400 200 0 1
workload (tl)
Dynamically Changed Experiment File
tim e
Fig. 4. Dynamically changed workload scenario The response time, λpred=Dpred1+Dpred2+C(filter1), depicted in Fig. 5, is measured using the same approach for each contention case by ER and PR. And the worst-case analysis (denoted as WC) by Audsley in [11] is employed as well to see how the new technique is accurate.
Response Time Prediction of 'Filter1' on Dynamic Scenario 0.5
second
0.4
resp. time
0.3
WC
0.2
ER PR
0.1 0 500
1300
700
1400
900
workload
Fig. 5. Predicted Response time comparison
1078 E.-N. Huh, L.R. Welch, and Y. Mun
Table 1. Error comparison by each technique workload 500 1300 700 1400 900
WC
ER
PR
0.081 0.241 0.117 0.277 0.147
0.066 0.133 0.086 0.146 0.091
0.011 0.005 0.004 0.018 0.004
Table 1 shows the errors can be evaluated to find which technique is significant. The PR technique is significant. Overall error in terms of resource utilization is as follows: 17.3 percenatge for the worst-case (WC) , 10.4 percentage for ER, and 0.08 percentage for PR are observed. From the thirty times experiments, there is no measurement error observed under the 95% confidence interval. The PR, probabilistic contention analysis technique, can accurately predict response time of an application on a host, while the worst-case analysis poorly predicts response times. Therefore, these experiments for the dynamic real-time systems give strong analysis that the worst-case analysis poorly utilizes computational resources, and new approaches can predict response time accurately with the dynamic environment constraints using current utilization. The probabilistic response time prediction method for dynamic real-time systems rather than the worst-case is very useful in terms of resource utilization and certification of real-time performance.
5. Future Study This research opens the possibility for yet future interesting research. First, there is the topic of certification for an event-driven real-time application which is periodic as well as aperiodic. Second, a new research area derived from this study is the determination of confidence levels of certification, which will be analyzed by mathematical methods with an upper bound, or error.
Acknowledgements This work is supported in part by the Ministry of Information Communication of Korea, under the "Support Project of University Information Technology Research Center(ITRC)" supervised by KIPA".
Real-Time Performance Estimation for Dynamic
1079
References 1. Liu, C. L., and Layland , J.W.: Scheduling algorithms for multi-programming in a hard-realtime environment. Journal of the ACM, Vol. 20. (1973) 46-61 2. Ramamritham, J.A. Stankovic and Zhao, W.: Distributed scheduling of tasks with deadlines and resource requirements. IEEE Transactions on Computers, Vol. 38. (1989) 110-123 3. Haban, D. and Shin, K. G.: Applications of real-time monitoring for scheduling tasks with random execution times. IEEE Transactions on Software Engineering, Vol. 16. (1990) 1374-1389 4. Tia, T. S., Deng, Z., Shankar, M., Storch, M.,Sun, J., Wu, L.C., Liu, J.W.S.: Probabilistic performance guarantee for real-time tasks with varying computation times. Proceedings of the 1st IEEE Real-Time Technology and Applications Symposium. IEEE Computer Society Press (1995) 164-173 5. Lehoczky, J.P.: Real-Time Queueing Theory. Proceedings of IEEE Real-Time Systems Symposium. IEEE Computer Society Press (1996) 186-195 6. Abeni, L. and Buttazzo, G.: Integrating multimedia applications in hard real-time systems. Proceedings of the 19th IEEE Real-Time Systems Symposium. IEEE Computer Society Press (1998) 3-13 7. Stewart, D.B. and Khosla, P.K.: Mechanisms for detecting and handling timing errors. Communications of the ACM, Vol. 40. (1997) 87-93 8. Puschner, P., Burns, A.: A Review of Worst-Case Analysis. The International Journal of Time-Critical Computing Systems, Vol. 18. (2000) 115–128 9. Atlas, A., and Bestavros, A.: Statistical rate monotonic scheduling. Proceedings of the 19th IEEE Real-Time Systems Symposium. IEEE Computer Society Press (1998) 123-132 10. Welch, L.R and Masters, M.W.: Toward a Taxonomy for Real-Time Mission-Critical systems. Proceedings of the First International Workshop on Real-Time Mission-Critical Systems (1999) 11. Audsley, Neil C.: Deadline Monotonic Scheduling. Report YCS-90-146. Department of Computer Science, York University (1990) 12. Welch, L. R., Shirazi, B.: A Dynamic Real-Time Benchmark for Assessment of QoS and Resource Management Technology. IEEE Real-Time Application System (1999) 13. Burns, F., Koelmans, A., Yakovlev, A.: WCET Analysis of Superscalar Processors Using Simulation With Coloured Petrinets. The International Journal of Time-Critical Computing Systems, Vol. 18. (2000) 275–288
A Load Balancing Algorithm Using the Circulation of A Single Message Token Jun Hwang, Woong J. Lee, Byong G. Lee, and Young S. Kim School of Computer Science and Engineering, Seoul Women's University, 126 Nowon-Gu GongReung-Dong Seoul, Korea, 139-774 {hjun, wjlee, byongl, yskim) @swu.ac.kr
Abstract. We present an efficient load balancing algorithm which simplifies the system status information through the use of an information message (VISITOR) in distributed system. Using the proposed algorithm, information gathering for decision-making is possible with less number of messages than the one using existing load balancing algorithm. The proposed algorithm improves the performance of distributed systems over the existing algorithms, not only by exchanging fewer messages to gather the information for making decision on load balancing and migration, but also by automatically determining of node and when to migrate.
1 Introduction Load Balancing Strategy (LBS) is an activity of process sharing performed by distributed processes. The activity is called as a global or distributed scheduling [14, 111 in that process allocation is targeted to all distributed processes while typical process scheduling such as FIFO, LRU, RR are called as a local scheduling. LBS minimizes response time and maximizes the throughput by considering a system as a whole virtual process. The performance of load balancing mainly depends on its algorithm, which can be categorized as following; First, when a node has to inform its own load status to other nodes, or to select a node to migrate, load balancing algorithm takes necessary global information through message broadcasting and do process migration accordingly [16, 101. Second, a node utilizes the global information only for figuring out the load status of each node. For other decision-making, such as load balancing and migration, is done by heuristic information [16]. Third, each node does not need the global information from every node but from adjacent node only, and the load balancing is done by the judgment of a local node [16,7,4]. These load balancing algorithms may cause the inundation of broadcasting messages in picking up of global information and in selecting a node for migration [I]. In this paper, proposed is an algorithm that performs P.M.A. Sloot et al. (Eds.): ICCS 2002, LNCS 2331, pp. 1080−1089, 2002. Springer-Verlag Berlin Heidelberg 2002
A Load Balancing Algorithm
1081
the global information retrieval and load balancing through a single information message by simplifying the way of acquiring the global information and of making decision for migration. Chapter 2 describes a load balancing algorithm, e.g., VISITOR algorithm, and reviews its concept and characteristics. Chapter 3 introduces VISITOR (Y-)which enhances the VISITOR algorithm. In chapter 4, the simulation context and necessary parameters for VISITOR (11) are discussed, and the performance of the proposed algorithm is analyzed in comparison with VISITOR algorithm. Finally, Chapter 5 concludes the paper with a summary and future work.
2 Load Balancing Algorithm 2.1 The Concept of VISITOR Algorithm Provided with the load information, each node compares its own load state with those of entire system. In case that load imbalance is detected by exceeding a neutral zone , process migration occurs from an overloaded node to an under-loaded node. According to which node requests the migration, different approaches can be used, e.g., 'Receiver-Initiative' or 'SenderInitiative'. The VISITOR message contains the global information that represents the load status and migration information of each node. This message is circulated along the nodes at constant time interval and supplies necessary information to each node.
2.2 VISITOR Algorithm To find out whether the load status of each node is overload or under-load, an average load value must be used. The average load value influences so much to the performance and efficiency of load balancing algorithm [6, 151. The tracing function H for measurement of average load value is defined as follows;
where AVGk : average load value on VISITOR message after kth circulation AVGk+l : expected average load value after k+l circulation m : number of nodes in the distributed system LOCAL : measured value of local load (number of processes in local queue) CD : communication latency time
1082 J. Hwang et al.
The function H is derived from the facts that the load value at a node influences to the global information at the ratio of l/m. Except from the information that is used in process migration, a control message is also necessary to run VISITOR algorithm as following. VISITOR:
circulates along all the nodes in the connection during load balancing and provides the load balancing information. ACCEPT: permits migration of overloaded node to under-loaded node. RECOMM: does not participate in the migration, but informs the overloaded nodes that other nodes are under-loaded states. Figure 1 depicts the structure of VISITOR message. VISITOR message fills with the information such as average load value, the least loaded node, and the most loaded node, and circulates along logical ring consecutively. As the VISITOR message passes through nodes, each node keeps its local load value and figures out load status of its own as to the whole system.
VAL
Fig. 1. The VISITOR message format
3 VISITOR(Y-) Algorithm Because VISITOR is complicated algorithm, it is a burden for a distributed system to take responsibility of load balancing. To lessen the burden, VISITOR ( Y - ) algorithm is suggested by modifying original VISITOR algorithm. The difference between VISITOR and VISITOR (Y-) algorithm is that VISITOR ( Y - ) adopts 'Receiver-Initiative' method (VISITOR uses SenderReceiver-Initiative' method). It is known that if the communication latency is short and the node is not overloaded, 'Receiver-Initiative' method performs better than 'SenderJReceiver-Initiative' method.
A Load Balancing Algorithm
1083
3.1 Detailed Description of VISITOR (Y-)Algorithm In section 2.2, the heuristic function H was defined with equation (1). The function H is to calculate a new average load value with local load value and system-wise load value. However, with the function H, average load value is amplified without any control as the number of nodes and communication latency increases, thus resulting in high variance of average value. Thus, a function F that is complimentary of function H, is introduced.
where, LOCAL : number of process in local queue m: number of nodes in distributed system ACE: Adaptive Control Element (ACE) to reduce the error in tracing new average C: communication Delay Factor*Constant The ACF module influences the result of average load value. The ACF prevents the load value from influencing other nodes' values without any controlling over the value. That is, if the local load value in H is extremely high or low, its load value influences average load-tracing value of other nodes at the ratio of l/m. This displacement occurs periodically and reacts sensitively if the state of entire distributed system changes. Therefore, a settlement is necessary in calculating the average load value of node. Adding a simple function called as ACF solves this problem. The influences of ACF on average load value should keep minimal, even as the difference increases [5]. In addition, VISITOR (11) includes message transmission information. The message format of VISITOR (Y-)is shown in figure 2.
ID
MODE
LOAD VAL
Fig. 2. VISITOR(y-) message format
1084 J. Hwang et al.
3.2 VISITOR(Y-) Algorithm VISITOR(I1) algorithm for load balancing follows; CASE 1: In case of a VISITOR(Y-) message arriving, - Calculate a new average load value through ACF and decide its own load status by averaging the load value. - Execute the routine steps of load balancing. CASE 2: In case of overload state, - If there exist MIN information and its load value is less than average load value, choose a process from local queue. - Migrate selected process to the under-loaded node. - Eliminate the field information from MIN-INFO on VISITOR message. - Pass the VISITOR(Y-) message to the next node. CASE 3: In case of under-loaded state, - If there exist MIN information and its load value is greater than average load value, insert its own node ID and load value information into MIN-INFO field. - Send the VISITOR(Y-) message to the next node. CASE 4: In case of optimal loaded state, - Send the VISITOR(Y-) message to the next node. CASE 5: In case of process arriving, - Increase its own local load and put it into the local queue
4 The Simulation and Analysis 4.1 The Environment of Simulation The simulation uses 5* (M/M/l) model with following assumptions. (1) Fully-connected form of network topology is assumed. (2)The node number of distributed system is 5 and VISITOR message circulates in a numerical order (clockwise or counterclockwise). (3) Suppose that the load status of random node, average arrival rate and average service time is given from table 1. (4) Suppose the communication capacity of network system be 0.1 Mbps basically. (5) Physical or logical loss of message or any related error is ignored. (6) The message transmission is done in the unit of packet, and the length of packet is 1,000bits. (7) Transmission delay that takes for transmitting one packet is l0msec. (8) Length of process (work) is 12Kbits. (9) Size of message is 4Kbits after remote execution of process. (10) Length of control message is lKbits and duration time that stays for transmission and collection of information in a node is l0msec. (11) Length of VISITOR message is lKbits and duration time for staying and collecting of information is l0msec.
A Load Balancing Algorithm
1085
Table 1. The load status of each node (unit: sec) EU
UL
NL
OL
EO
Average arrival time
11.76
11.11
10.10
9.09
8.33
Average service time
10.0
10.0
10.0
10.0
10.0
Load status
EU (Extremely Under-loaded State), UL (Under-loaded State) NL (Loaded Normally State), OL (Overloaded State) EO (Extremely Overloaded State) Communication Delay = (PACKET SIZEIO.1Mbps) * (PROCESS SIZEl1000)---(2)
The equation (2) calculates communication delay time independently with the simulation environment [12]. The total communication time is achieved by dividing the size of process with the length of packet (lOOObits), and by multiplying it with communication latency time, which takes for transmitting one packet. Although above equation appears to be simple and insufficient, it still fully represents the characteristics of the communication system. The communication latency time is artificially calculated by equation (2) after accepting a message from a random node. After intended latency time is elapsed, the message is sent to a target node. This model is simulated in a fully connected network environment. Otherwise, more refinement should be done to the model.
4.2 The Performance of Average Load Value Tracing Function The real load value of the entire system is calculated by dividing the total number of processes in waiting queue at random time with the number of nodes. The trace value of VISITOR message is calculated by function F and H (Figure 3). The status of each node is assumed as shown in table 2.
Fig. 3. The relation of real average load value (COMM.DELAY = 20msec) and tracing average load value by function F and H respectively
1086 J. Hwang et al. Table 2. The load status of each node Node number
1
2
3
4
5
Load status
EU
OL
UL
EO
NL
To figure out the deviation between the real average load value and average tracing load value, the variation of standard deviation is used as depicted on figure 4. To get a accurate sample set, a sampling is done at a stable time of the system and the number of sample is about 3000. To get average deviation value, the difference of real value and traced value is calculated [3]. avelage bad
real avemge avemge b a d by H avemge b a d by F 1 2
3
4
5
6
7
8
9 10 11 12 13 14 15 16 17 18 19 20 tine
Fig. 4. The change of tracing value due to VISITOR passing delay
The actual average load value and sample mean of each trace function, F grows accordingly as~communicationlatenEy increases. This means that the communication latency is proportional to error rate. Function F, compared to H, is appeared to be low in speed of performance decrease due to the increase of communication costs. The performance degradation is stable with function F while function H showed sharp performance degradation. Figure 5 represents the degree of average load imbalance by communication latency. About 7,000 samples are used at the sampling rate of 50,000msec 100,000msec. Communication delay is multiples of 10m sec. From the figure, load imbalance ranges from 3 to 35 as the latency value changes, but is very high around 81 when no load balancing is applied. In addition, in case of communication time being small, VISITOR (Y-)algorithm performs better than VISITOR algorithm. The performance of VISITOR algorithm is improved as the communication latency increases.
-
A Load Balancing Algorithm
1087
4-3 Neutral Zone Neutral zone is a handler for controlling load state o f random node and regulating frequency o f process migration. Therefore, suitable adjustment o f Neutral zone is a key element for influencing the performance o f load balancing algorithm. Neutral zone as shown in figure 6 has two neutral widths approximately and these widths represent optimal load state. If this zone gets narrower, the load difference between two nodes decreases picking up many load migration, and thus increases communication costs o f load balancing. As the zone width gets larger, the frequency o f migration becomes smaller and thus may allocate more time on process execution.
Fig. 5. The change of load imbalance due to communication latency
1 TRACED
OVERLOADED STATE
- - -.I - -NEUTRAL - - - ZONE ------------
LOAD
NEUTRAL WIDTH
v UNDERLOADED
Fig. 6. The relation of neutral zone and load status
4.4 The Average Response Time
It is important to measure response time o f random process for analyzing the system performance. The simulation uses equation ( 3 ) for measurement o f average response time that represents the total length o f visit from which the process arrives in waiting queue and until it leaves, including remote processing
1088 J. Hwang et al.
R(t) represents the response time of random process and p is the number of process. Defined in equation (4) is the average response time of total system.
Rho
Fig. 7. The average arrival rate versus average service time
Figure 7 indicates that the result of average response time is better than the one when no load balancing algorithm is applied. Especially, when Y Z is 0.7-0.8, the average response time increases sharply while VISITOR incurs a rather smooth increases. Before approaches to 0.7, the proposed algorithm shows the worse performance due to various overheads, such as status information exchange, communication latency, message loss, etc.. After Y Z passes 0.7, however, VISIOTOR (11) shows stable increases in average response time.
5 Conclusion and Future Work In this paper, different methods of load balancing are presented to remove load imbalance, which occurs in typical distributed system. For this, two algorithms are introduced and analyzed with their performance comparison. With VISIOR (11) algorithm, the global information is achieved with less communication costs, compared with the one in broadcasting mechanism, and the accuracy of the information is proved. The performance, in terms of average response time, is improved about 50 80% when Y Z is 0.9 0.99 than when no load balancing algorithm is applied. The communication cost for updating the status information is reduced remarkably because the global
-
-
A Load Balancing Algorithm
1089
information is managed with simple message circulation and tracking functions. "Victim Problem" [I] in VISITOR algorithm, however, can cause poor system performance as decision making at each nodes is combined. For future studies, cooperation and coordination mechanism to reduce the frequency of decision-making will be examined. Also, the recovery and rehabilitation mechanism for reducing the possibility of message loss in communication line will also be studied.
References Alon, M., Barak, A., Manber, U., "On Dissemination Reliably without Broadcasting," IEEE The 7th International Conf. On DistributedIComputing Systems, pp. 74-81, Sep. 1987. Casavant,T. and Singhal, M., Reading in Distributed Computing Systems, IEEE Computer Society Press, 1994. Chao, L. L., Statistics Methods and Analysis, McGraw-Hill, 2ndEdition, pp. 188-192, 1974. Eager, D. L., et al., "Adaptive Load Sharing in Homogeneous Distributed Systems," IEEE Trans. On Software, Vol. SE-12, No. 5, pp.662-675, May 1986. Kness, C., Matkowsky B. J., Schuss, Z., Tier, C., "Two Parallel Queue with Dynamic Routing," IEEE Trans. on Comm., Vol. COM-34, No. 12, pp. 11701175, Dec. 1986. Krueger, P., Livny, M., "A comparison of Preemptive and Non-Preemptive Load Distributing," The gthInternational Conf. On DCS, pp. 123-130, 1988 Lin ,F.C.H., Keller, R. M., "The Gradient Model Load Balancing Method," IEEE Trans. on SE, Vol. SE-13, No. 1, Jan. 1987. Ni , L. M., et al., "A Distributed Drafting Algorithm for Load Balancing," IEEE Trans. on Software Engineering, Vol. SE-11, No. 10, Oct. 1985. Smith, R. G., "The Contract Net Protocol: High Level Communication and Control in a Distributed Problem Solver," IEEE Trans. on COMPUTERS, Vol. C-29, No. 12, Dec. 1980. 10. Stankovic, J., "A Perspective on Distributed Computer Systems," IEEE Trans. on Comp. Vol. C-33, No. 12, pp. 1102-1115, Dec. 1984. 11. Stankovic, J., "An Application of Bayesian Decision Theory to Decentralized Control of Job Scheduling," IEEE Trans. on Comp., Vol. C-34, No. 2, pp. 117130, Feb. 1985. 12. Tay, Y. C. and Pang, HweeHwa, "Load Sharing in Distributed Multimedia-onDemand Systems", IEEE Transactions on Knowledge and Data Engineering, Bol. 12, No. 3, MayIJune 2000. 13. Tonogai, D., "A1 in Operating Systems: An Expert Scheduler," PROGRESS REPROT No. 88,12, Dec. 1988. 14. Zhou, S., Ferrari, D., "A Measurement Study of Load Balancing Performance," IEEE The 7thConf. on DCS, pp. 490-497, Sep. 1987. 15. Zhou, S., Ferrai, D., "An Experimental Study of Load Balancing Performance," REPORT No. UCBlCSDl871336, Berkeley, California, Jan. 1987.
A Collaborative Filtering System of Information on the Internet DongSeop Lee1, HyungIl Choi2 Soongsil University, 1-1 Sando-5Dong, DongJak-Gu, Seoul, Korea 1 [email protected], 2 [email protected]
Abstract. In this paper we describe a collaborative filtering system for automatically recommending high-quality information to users with similar interests on arbitrarily narrow information domains. It asks a user to rate a gauge set of items. It then evaluates the user's ratings and suggests a recommendation set of items. We interpret the process of evaluation as an inference mechanism that maps a gauge set to a recommendation set. We accomplish the mapping with FAM (Fuzzy Associative Memory). We implemented the suggested system in a Web server and tested its performance in the domain of retrieval of technical papers, especially in the field of information technologies. The experimental results show that it may provide reliable recommendations.
1 Introduction In this paper we describe a collaborative filtering system for automatically recommending high-quality information to users with similar interests on arbitrarily narrow information domains. Our system follows the same operational principle as the Eigentaste system [4]. It asks a user to rate a gauge set of items. It then evaluates the user's ratings and suggests a recommendation set of items. We interpret the process of evaluation as an inference mechanism that maps a gauge set to a recommendation set. We accomplish the mapping with FAM (Fuzzy Associative Memory). FAM provides a framework that maps one family of fuzzy sets to another family of fuzzy sets [3]. This mapping can be viewed as a set of fuzzy rules that associate input fuzzy sets (gauge sets) with output fuzzy sets (recommendation sets). FAM also provides a Hebbian-style learning method that establishes the degree of association between an input and output [2]. This learning method is very simple and takes very little computation time. Another aspect of collaborative filtering is how to form groups of users with similar tastes, so that the known preference of a group of users may be exploited to predict the unknown preference of a new user. This is a typical problem of clustering [6]. However, our approach does not require this type of explicit clustering, since the clustering is embedded in connection weights of FAM. In fact, FAM generates fuzzy rules that classify data into groups of classes. This grouping is supervised at the stage P.M.A. Sloot et al. (Eds.): ICCS 2002, LNCS 2331, pp. 1090−1099, 2002. Springer-Verlag Berlin Heidelberg 2002
A Collaborative Filtering System of Information on the Internet 1091
of learning the connection weights. The details will be discussed in the next section[1]. Our collaborative filtering system consists of two main parts, a learning part and inferring part. The learning part operates off-line. It analyzes training data made up of input and output pairs in order to form fuzzy sets, where input data correspond to a set of rates on a gauge set of items and output data correspond to a set of rates on a recommendation set of items. Our system asks users to rate their preference on a continuous rating scale. To rate items, we may use a horizontal "rating bar" as in [4], where a user is supposed to click a mouse. The learning module then generates a correlation matrix that shows the degree of association between input and output fuzzy sets. The details will be discussed in section 3. The inferring part operates online. It presents a gauge set to a new user. It then processes the rates on the gauge set and draws a recommendation with the fuzzy rules built up in the learning part. Our system makes a conclusion in the form of induced preference rates on a recommendation set. The details will be discussed in section 2. ô
ÿ
þ
ý
ó
ò
ü
û
ñ
ú
þ
ù
ø
÷
û
ý
õ
ð
ü
ü
ø
ï
ú
þ
ù
ð
ø
î
ÿ
ø
þ
ý
û
ö
ú
õ
ý
ú
û
ï
ÿ
ø
ö
ý
û
ý
ø
õ
ý
ø
ù
ø
÷
ý
ê
ì
ú
ë
ý
ý
î
ï
ë
ý
ý
û
ü
ö
ø
é
ð
ù
ö
é
ú
ø
û
ø
ù
ø
÷
ú
í
þ
ì
ú
ú
ë
ý
ý
ý
ë
ï
õ ë
ð
ý
ø
û
ü
ø
ù
ý
ø
ý
õ
þ
ù
ð
ø
è
ö
ë
é
ç
ç
!
ý
þ
ý
ù
÷
ì
þ
#
%
&
(
Fig. 1. System organization
2 Inferring Model With FAM FAM can be viewed as the fusion of associative memory and fuzzy logic. It associates a family of fuzzy sets with another family of fuzzy sets. Figure 2 shows a basic structure of FAM. The antecedent term Ai in fuzzy association
( Ai , B j ) denotes an input associant, the consequent term B j denotes an output associant, and the synaptic weight wij denotes the degree of association between input and output associants. In our collaborative filtering system, the input associant corresponds to a fuzzy set of a gauge items, the output associant corresponds to a fuzzy set of recommendation items, and the synaptic weight corresponds to the degree of correlation between these two sets. As an example, one may consider fuzzy association ("high preference of computer", "high preference of internet") with the association degree of 0.8. There are several ways of interpreting the synaptic weight
1092 D. Lee and H. Choi
wij . One popular interpretation considers it as a fuzzy Hebbian-style correlation coefficient in which the weight is encoded as the minimum of input and output associant values [3]. The details of this interpretation will be discussed in the third section. If we now somehow encoded the set of synaptic weights, then FAM carries out forward recalling through max-min composition relation [3]. Suppose that a fuzzy set Ai is defined on the domain of an item xi and ai Ai ( xi ) is the fit value of xi to a *
+
membership function y j and b j
,
-
Bj
*
Ai
, and a fuzzy set B j is defined on the domain of an item
( y j ) is the fit value of y j to a membership function *
Bj
. Then
FAM exhibits forward recalling as in (1) A W
B
.
-
(1)
where denotes a max-min composition operator. That is, the recalled component b j is computed by taking an inner product of fit values ai 's with the jth column of /
W.
bj ?
1
0
-
(2)
max{min(ai , wij )} i
1
7
8 8
5
9
?
2
0
2
?
3
0
3
:
1
5
5
<
?
4
0
=
@
1
;
2
6
@
2
@
6
>
4
Fig. 2. Fuzzy associative memory We suggest a fuzzy inference system that employs FAM for implementing fuzzy rules. These rules embody the task of grouping users with similar tastes in the form of association between gauge items and recommendation items. That is, we interpret a gauge items as an antecedent part of a fuzzy rule, a recommendation items as a consequent part, and a synaptic weight as the degree of reliability of the rule. When an antecedent part of a rule has several fuzzy terms, the corresponding input associant forms a conjunction of such fuzzy terms. Figure 3 shows the structure of our inferring model that consists of four layers. For the purpose of illustration, we assume that there are n gauge items xi (i=1, ..., n) and m recommendation items y j (j=1, ..., m), each furnishes pi fuzzy sets, and each is represented by a fuzzy set B j .
A Collaborative Filtering System of Information on the Internet 1093
The input layer of Figure 3 just accepts input values that correspond to rates that a user assigns to a set of gauge items. Thus, the number of nodes in the input layer becomes n. The fuzzification layer contains membership functions of input items. Since there are n input items and each input item xi produces pi fuzzy sets, the total n A
number of nodes in this layer becomes
pi . The output of this layer then becomes
i 1 C
the fit values of input rates to associated membership functions. The antecedent layer contains antecedent parts of fuzzy rules, which have the form of logical AND of n B
individual fuzzy terms. This layer requires N
pi nodes, since we allow every
D
i 1 C
possible combination of fuzzy sets drawn one from each group of pi fuzzy sets. Thus, each node in the antecedent layer has n incoming links. Each incoming link has a weight that represents the degree of usefulness of an associated fuzzy set. If links from some node of the fuzzification layer have a high value of weight, it means that the fuzzy set contained in the node is very useful in inferring a desired recommendation. r
s
w
x
t
y
u
v
z
{
u
}
}
~
|
w
~
x
|
x
y
z
v
~
ÿ
s
x
E
F
G
z
s
Z
[
\
F
G
H
E
J
x
G
y
z
z
z
s
v
ÿ
{
s
w
E
H
U
G
z
x
y
u
z
z
s
v
{
H
]
H
^ T
v
w
Y E
s
{
_
`
a
b
F
O
P
P
c
d
e
f
g
E
F
G
H
E
J
G
H
E
U
G
I Q
E
F
G
i
kj
G
L
l
E
J
S
K
h
E
R
I
F
G
H
E
J
G
E
I
U
G
H
H
K
T
F
E
E
J
G
U
G
G
H
J
E
G
E
I
U
G
I
I
E
E
F
G
I
E
J
G
E
H
U
G
H
H
K
T
M
J
U
V
E
U
G
N
W
X
I
m
n
o
p
q
E
F
G
I
E
J
G
I
E
U
G
I
E
Fig. 3. Inference model with FAM The details of how to determine the weights will be discussed in section 3. Each node of this layer just compares incoming weighted values and takes the minimum of
1094 D. Lee and H. Choi
them, since the truth value of an antecedent part of a rule is determined by taking a logical AND of individual truth values of participating fuzzy terms. The consequent layer contains consequent parts of fuzzy rules. This layer contains m membership functions, each of which is to determine the preference of an individual recommendation item. We allow full connections between the antecedent layer and the consequent layer. But, each connection may have a different value of weight, which represents the degree of credibility of each connection. We basically follow the max-min compositional rule of inference. Thus, when N antecedent nodes A1, , AN are connected to the jth consequent node B j with weights wij 's, the
output of the jth consequent node is a deffuzified value determined as in (3). y* j where
Bj
1
Bj
(max1
(3)
i N {min( wij , output ( Ai ))})
( y ) is a membership function contained in the jth consequent node, and
output ( Ai ) is an output of the ith antecedent node. The output of each consequent node is a final crisp conclusion in the form of the degree of recommendation on the corresponding item. Here, we use an inverse mapping for deffuzification, which maps the max-min composed membership to the argument of a corresponding function.
(x) ¦
§
preference
( y)
¢ ¢
¤
¥
¡
¢
¢
Fig. 4. Basic structures of membership functions
3 Learning Model Based on Observation Our inferring model can work properly only when membership functions as well as synaptic weights are determined in advance. In this section, we propose a learning method that derives the necessary information from input-output training data that correspond to rates on gauge-recommendation items obtained form users. The first issue is how to determine the number of fuzzy sets for each item and corresponding membership functions. There must be some systematic way to divide the range of each input and output item into subranges and associate each subrange with a proper membership function. For this purpose, one may consider a tuning approach that
A Collaborative Filtering System of Information on the Internet 1095
exploits the distributions of data [2]. Here we take a simple approach and supplement it with the usefulness measure to be described later. Figure 4 shows the basic structures of our membership functions. A user is supposed to rate his preference on an item by a degree between 0 and 1. A large value denotes a high preference and a small value denotes a low preference. For a gauge items, we divide the entire range into three parts and assign them with low, medium, high fuzzy sets. Their membership functions are defined as in (4). For a recommendation items, we take an entire range for a fuzzy set "preference" and define its membership function to be monotonically increasing, y . This is to reduce the number of nodes in the consequent layer of preference ( y ) ©
¨
our inferring model. But, one may introduce several fuzzy sets also for a recommendation items, if one wants to be more specific.
1 2 x, 0,
0 x 0.5 0.5 x 1
®
low ( x ) ²
« ¬
°
ª
(10 x 2) / 3, ( 10 x 8) / 3,
0.2 x 0.5 0.5 x 0.8
®
medium ( x ) ²
²
high ( x )
« ¬
°
ª
±
®
«
¬
°
ª
¯
(4)
¯
0 x 0.5 0.5 x 1
0, 2 x 1,
®
¯
To determine the usefulness of an input fuzzy set, we evaluate how the fuzzy set discriminates recommendation items. If an input fuzzy set shows such a characteristic that the training data that fall on the fuzzy set have their corresponding rates on recommendation items evenly distributed, we claim that the fuzzy set is not useful in terms of recommending some specific items. However, if their rates on recommendation items show a spiky type of distribution, we claim that the fuzzy set is useful in inferring recommendation. This criterion is equivalent to emphasizing a distinctive fuzzy set. ³
D Al °
1
M
PAl ( j ) ln( PAl ( j ))
±
·
j 1 µ
³
K ²
PAl ( j )
i 1 M K
Al
( xi ) rB j ( xi ) ·
(5)
µ
°
³
³
²
j 1 i 1 µ
Al
( xi ) rB j ( xi ) ·
µ
To evaluate the discriminating power, we use the measure of entropy. The entropy is a statistical measure of uncertainty, and it is good at revealing the dispersion of data. We compute degrees of usefulness of input fuzzy sets as in (5). In (5), xi denotes ith input-output training data, A1 ( xi ) denotes the fit value of xi to an input ©
fuzzy set Al , rB j ( xi ) denotes the rate of xi on a recommendation item B j , M is the number of recommendation items, and K is the total number of training data. Thus,
1096 D. Lee and H. Choi
PAl ( j ) denotes the probability density of rates on the jth recommendation item,
which are associated with an input fuzzy set Al . Our inference model requires a predetermined correlation matrix that represents the degrees of associations between input and output fuzzy sets. We take a Hebbian-style learning approach to build up the correlation matrix. The Hebbian learning is an unsupervised learning model whose basic idea is that "the synaptic weight is increased if both an input and output are activated." In this way, the phenomena of habit and learning through repetition are often explained. Many artificial neural networks which take a Hebbian learning approach increase network weights according to the product of excitation levels of an input and output. In our fuzzy associative memory, input and output values are fit values to membership functions. Thus, we replace a product operation with a minimum operation and an addition operations with maximum operation. That is, when ai (n) is an input associant for the nth learning datum and b j (n) is an output associant for the nth learning datum, the change of weight is carried out as in (6). wii (n) ¼
(1
¹
) wii (n 1)
¹
¸
»
¸
»
º
(ai (n) AND b j (n))
(6)
is a positive learning rate that is not greater than 1. This learning rate In (6), controls the average size of weight changes. Our inference model allows full connection between the antecedent layer and the consequent layer. This full connection may present too many rules, causing too much computation time. The connection weight may be considered as the confidence measure of a related fuzzy rule, since our learning approach is based on Hebbian learning. We may then consider the rules with small connection weights to be unimportant in inferring a useful conclusion. Thus in practical situations, we may prune some of the connections whose weights are less than some predetermined threshold and reduce the number of rules. ½
4 5.
Experimental Results and Conclusion Experimental Results and Conclusions
To confirm the effectiveness of our suggested model, we chose the domain of retrieval of technical papers, especially in the field of information technologies. We implemented our learning and inferring module using C++ in a Web server. We also developed Web interface module using Java that allows internet users to rate papers. The interface module presents papers and collects ratings as users click on a rating bar. In the learning phase, all the papers in both the gauge set and recommendation set are presented to users one by one. After a user rates each paper, another is presented. The collected ratings are then used to build up membership functions and connection weights of our FAM. In the testing phase, our interface module presents to a user the papers in the gauge set and asks for ratings on the papers. After all papers in the gauge set are rated, our system recommends to the user the papers in the
A Collaborative Filtering System of Information on the Internet 1097
recommendation set with induced preferences. Our system also collects user's ratings on each recommended paper, so that it may compare recommended ratings against user's ratings. To evaluate the performance of our system, we performed three types of experiments: we first examined the usefulness measure of a fuzzy set in terms of its effect on accuracy of induced preferences. The second experiment was to examine the effect of the number of connections of FAM on the accuracy of induced preferences. We also compared the performance of our system against those of other systems. In our experiments, the learning rate in (6) was set to 1.0 and the initial weights were set to 0. 100 users were involved in the learning phase, and also 100 different users were involved in the testing phase. The gauge set contains 10 different papers and the recommendation set contains 20 different papers, so the numbers of nodes in input layer and consequent layer are 10 and 20, respectively. The table I lists the degrees of usefulness of input fuzzy sets. These values reflect how input fuzzy sets discriminate recommendation items. For illustration, A21 has the high value of 0.87, since the training data that fall on this fuzzy set have the most of their recommendation rates only in 4 items. In contrast, A62 has the small value of 0.17, since the training data that fall on this fuzzy set have their recommendation rates spreaded among 17 items. ¾
Table I. Usefulness of each input fuzzy set Fuzzy set A01 A11 A21 A31 A41 A51 A61 A71 A81 A91
Usefulness 0.78 0.76 0.87 0.05 0.84 0.65 0.86 0.92 0.57 0.87
Fuzzy set A02 A12 A22 A32 A42 A52 A62 A72 A82 A92
OMAE
1 M Ã
MAE ( j ) Ã
usefulness 0.85 0.44 0.92 0.46 0.65 0.14 0.17 0.82 0.65 0.88
¿
M
Â
1 N
Usefulness 0.32 0.15 0.86 0.65 0.82 0.09 0.60 0.14 0.79 0.88
MAE ( j )
j 0
(7)
À
N ¿
Fuzzy set A03 A13 A23 A33 A43 A53 A63 A73 A83 A93
Â
rij Á
pij
i 0 À
To evaluate the accuracy of the performance, we use the Overall Mean Absolute Error metric that is often used in the literature. In (7), pij is the prediction for how user i will rate item j and rij is the actual rate given by user i for item j, M is the number of items user i has rated, and N is the total number of users involved in the
1098 D. Lee and H. Choi
test. In (7), MAE(j) represents the mean error for the jth recommendation item over the whole users, and OMAE represents the mean error over the whole items and users. Table II shows the performance of our system with and without the usefulness measure of input fuzzy set. As can be noted, the performance is improved by 0.07 on average with the help of the usefulness measure. Table II. MAE of each recommendation item Recommendation Item Paper01 Paper02 Paper03 Paper04 Paper05 Paper06 Paper07 Paper08 Paper09 Paper10
MAE without usefulness measure 0.234 0.164 0.209 0.186 0.160 0.207 0.242 0.219 0.182 0.170
MAE with Usefulness measure
Recommendation item
0.215 0.156 0.201 0.178 0.149 0.203 0.234 0.216 0.179 0.168
MAE without usefulness measure
Paper11 Paper12 Paper13 Paper14 Paper15 Paper16 Paper17 Paper18 Paper19 Paper20
0.154 0.189 0.220 0.237 0.205 0.156 0.151 0.214 0.178 0.203
MAE with usefulness measure 0.142 0.185 0.215 0.231 0.196 0.154 0.147 0.206 0.171 0.194
We examined the effect of the number of synaptic connections on the performance of the system. We pruned out synaptic connections whose weights are less than some threshold Th, so that the number of connections is reduced and the system is simplified. For illustration, we have 25% reduced connections for Th=0.1, and 35% reduced connections for Th=0.2. We also compared the performance of our systems with those of other two systems: POP [5] and Eigentaste [4]. Table III shows the results. As can be noted, our system with Th=0.1 surpasses the others. It is interesting to note that the performance of our system does not drop drastically with increasing the threshold Th. Table III. Comparison of performance System Our system with th = 0.0 Our system with th = 0.1 Our system with th = 0.2 POP[9] Eigentaste[6]
OMAE 0.187 0.192 0.245 0.302 0.284
Acknowledgement This work was supported by the Korea Science and Engineering Foundation (KOSEF) through the Advanced Information Technology Research Center(AITrc).
A Collaborative Filtering System of Information on the Internet 1099
Reference 1. Dae-Sik Jang, Hyung-Il Choi, Fuzzy Inference system based on fuzzy associative memory, journal of Intelligent and fuzzy systems, Vol.5, 271-284. 1997. 2. Hideyuki T, Isao H, NN-Driven fuzzy reasoning, Int. J. Approximate Reasoning, 191-212. 1991. 3. Kosko B, Neural Networks and Fuzzy Systems," Prentice-Hall International. 1994. 4. Zimmermann HJ, Fuzzy Set Theory and Its Applications, KALA. 1987. 5.
Ken Goldberg and Theresa Roeder and Dhruv Gupta and Chris Perkins, Eigentaste: A constant time Collaborative Filtering Algorithm, University of California, Berkeley, Electronics Research Laboratory Technical Report M00/41. 2000.
6. Jonathan Herlocker, Joseph Konstan, Al Borchers, and John Riedl. An algorithmic framework for performing collaborative filtering, In Proceedings of the SIGIR. ACM, August 1999.
Hierarchical Shot Clustering for Video Summarization YoungSik Choi, Sun Jeong Kim, and Sangyoun Lee Multimedia Technology Research Laboratory, Korea Telecom, Seocho-Gu Woomyeon-dong 17 Seoul, Korea, {choimail, sunjkim, leesy}@kt.co.kr
Abstract. Digital video is rapidly becoming a communication medium for education, entertainment, and a variety of multimedia applications. With the size of the video collections growing to thousnads of hours, efficient searching, browsing, and managing video information have become of increasing importance. In this paper, we propose a novel hierarchical shot clustering method for video summarization which can efficiently generate a set of representative shots and provide a quick and efficient access to a large volume of video content. The proposed method is based on the compatibility measure that can represent correlations among shots in a video sequence. Experimental results on real life video sequences show that the resulting summary can retain the essential content of the original video.
1 Introduction With the recent advances in compression and communication technologies, vast amounts of video information are created, stored, and transmitted over networks for education, entertainment, and a host of multimedia application. Therefore, efficient searching, browsing, and managing video information have become of increasing importance. The MPEG group has recently begun a new standardization phase for efficient searching and managing multimedia content (MPEG-7). The MPEG-7 will specify the ways to represent multimedia information by means of descriptors and description schemes. The question of how to obtain these descriptors automatically is becoming a highly important research topic. Particularly, automatic video summarization is gaining the attention as a way to condense a large volume of video into smaller and comprehensible units, and allows quick and easy access to video content. There are a few approaches to video summarization: (1) selecting and concatenating the “most representative” images or shots [3], [6]. (2) creating a “skim” video which represents a short synopsis [7]. In this paper, we address the problems with selecting the “representative” shots and propose a novel hierarchical shot clustering for video summarization. Shot clustering has been frequently used for video summarization and segmentation. In P.M.A. Sloot et al. (Eds.): ICCS 2002, LNCS 2331, pp. 1100−1107, 2002. Springer-Verlag Berlin Heidelberg 2002
Hierarchical Shot Clustering for Video Summarization 1101
this approach, a video sequence is first segmented into shots and then shot clustering is applied to select the representative shots [3], [6] or to group the shots into scenes (story units) [4], [5]. Agglomerative hierarchical clustering with time constraint has been used for shot clustering [1], [2]. This approach can produce the tree-structured representation that is useful for video summarization. However, it is expensive, requiring insertions and deletions of the cluster dissimilarity matrix, and requiring a search for each closest cluster pairing. Window-based shot grouping has also been used [4], [5]. In this paradigm, incoming shots are compared with the shots in a given window in order to check if incoming shots may be included in the current segment. This method is computationally less expensive to group shots into scenes than agglomerative clustering. It is, however, difficult to have a compact hierarchical representation for video summarization. To overcome the limitations in both approaches, we propose a novel shot clustering algorithm that can efficiently produce the representative shots and extract the hierarchical structure from a video sequence. The proposed clustering algorithm is based on the assumptions that the shots from a scene are more likely to be compatible with each other than those from other scenes, and that the shot highly compatible with other shots is more likely to be the representative shot of a scene. We define the compatibility measure to be the average value of the degrees to which a shot is similar to its neighboring shots within a given window. Using the compatibility measure, we develop an efficient clustering algorithm for video summarization.
2 Hierarchical Shot Clustering
2.1 Compatibility Measure We define the compatibility measure as follows. Let S = {s1, s2, …, sk, …, sN} denote a video sequence, where si is i-th shot and N is the total number of shots. The compatibility measure C(si ) of si within a given window is defined as: C(si ) = (1/ N(i)) åj∈N(i) µ ji,
(1)
where N(i) is the set of the neighbors of si and µji is a fuzzy membership function which determines the value of the degree to which si is similar to sj. N(i) may be {si-2, si-1, si+1, si+2} if the window size is 4. The membership function can be defined as a monotonically decreasing function of the dissimilarity between si and sj. In this Letter, we propose to use the following bell-shaped membership function:
µ ji = exp(-D(si, sj)/β j).
(2)
1102 Y. Choi, S.J. Kim, and S. Lee
Note that in general, µ ij is not equal to µ ji . The D(si, sj) denote the dissimilarity between si and sj. The β j is a scaling factor and is determined for each sj to consider the local context as follows.
β j = (1/|N(j)|)åi∈N(j)D(si, sj),
(3)
where |N(j)| is the number of neighbors of sj. The dissimilarity D(si , sj) between si and sj may be defined as a function of shot feature vectors. In this Letter, the shot feature vector is computed as the average of all frame feature vectors within a shot.
Fig. 1. Illustration of grouping and handling local minima
Pi-1
qi-1
pi-1k*
Pi+1
Pi
qi
pik*
qi +1
2.2 Clustering Algorithm: The proposed compatibility measure can represent the correlations among shots in a video sequence. If the value of C(si ) is higher than the values of its neighbors, si is more compatible with its neighbors and is, therefore, more likely to be the representative of its neighbors. If the value of C(si) is lower than its neighbors, si is less compatible with its neighbors and, therefore a scene boundary more likely exists around si . Taking these into accounts, we propose the following hierarchical clustering method. (Step 1) Initialization: Let S = {s1, s2, …, sk, …, sN} be a video sequence and {h(s1), h(s2), …, h(sk), …, h(sN)} be the corresponding shot feature vectors, where h(sk) is the feature vector of shot sk. We initialize S as the set of initial clusters, P. (Step 2) Grouping and selecting a key shot: For each cluster in P, we obtain the compatibility using equations (1), (2), and (3). We find the clusters with local minimum compatibility values. Let {q1, q2, …, qn} denote the set of these clusters, where n is the total number of local minima (See Figure 1). We group the clusters between clusters qi and qi+1 into a new cluster Pi . Let Pi = {pi1, …, pik*, …, pin}, where in is the total number of clusters between qi and qi+1, and pik* is the cluster with the local maximum compatibility value between qi and qi+1. Note that only one pik* exists between qi-1 and qi. We select pik* as the representative cluster of Pi .
Hierarchical Shot Clustering for Video Summarization 1103
If this is the first iteration, we select pik* as the key shot of Pi . Otherwise, we select pik*‘s key shot as the key shot of Pi. Note that the representative clusters and the key shots are the same in the first iteration. (Step 3) Handling Local Minima: We can consider {q1, q2, …, qn} as outliers or boundary clusters. Therefore, we first check whether qi is an outlier or a boundary cluster according to C(qi ). If qi is an outlier, qi becomes a new cluster between Pi-1 and Pi. Otherwise, we merge qi into the closer cluster Pi-1 or Pi with respect to the dissimilarity function defined in equation (4). The following describes this process. For each local minimum cluster qi starting from q1, do the following. (Step 3-a) If C(qi) < TC, then make qi as a new cluster between Pi-1 and Pi . TC is the threshold value. (Step 3-b) If C(qi) ≥ TC, do the following. If A(qi , Pi-1) < A(qi, Pi ), then add qi into cluster Pi-1. Otherwise, add qi into cluster Pi . A(qi , Pi ) is the dissimilarity between clusters qi and Pi, and defined as A(qi , Pi ) = min D(qi, pij), where pij ∈ Pi.
(4)
(Step 4) Time constraint and terminating condition: For each new cluster Pi , we check the time constraint with threshold value TT. If the duration of cluster exceeds TT, then we ungroup the cluster and make the clusters in Pi as new clusters between Pi-1 and Pi+1. Note that this ungrouping is similar to clustering with time-constrained distance as in [1]. We terminate the clustering process if there is no change in clusters Pi . (Step 5) Update cluster feature vector: We update the feature vectors of new clusters Pi such that the feature vectors of more compatible clusters may gain more weights than those of less compatible clusters. That is, we update the feature vector h(Pi ) for each cluster Pi as the weighted average of feature vectors h(pij) of clusters pij in Pi . h(Pi ) = åC(pij)h(pij)/åC(pij), where pij ∈ Pi. Set P = {P1, …, Pm} and m is the number of new clusters Pi . Go to Step 2.
3 Video Summarization The proposed clustering algorithm results in a hierarchical structure of a video sequence where each node corresponds to Pi. Each node has the representative cluster pik* and the key shot (See Step 2 in 2.2). The set of clusters Pi in the highest level can be considered as a partition of a video sequence (a video segmentation) and the set of the representative clusters pik* and key shots as a abridged version (a video summarization). Taking these into accounts, we present the following scheme for video summarization.
1104 Y. Choi, S.J. Kim, and S. Lee
Suppose that our clustering method produces M clusters in the top level and we select the set of key shots in each cluster as a summary. Then, the length TC of the summary becomes TC = åi=1.. M T(Si ), where T(Si) is the length of key shot Si from cluster Pi. Conversely, if a user requests a summary of length Treq, we need to determine the number of clusters in the top level. Assuming that T(Si ) be the average shot length Tavg in a video sequence, the number of segments M becomes M = Treq/Tavg. Now, we need to adjust the threshold value TT to produce M clusters in the top level. In the proposed clustering algorithm, the number of clusters M has an inverse relation to the threshold value TT in Step 4 in Section 3. Therefore, we propose to set TT as M 5 where Torg is the length of the original video and k is a constant greater than 1. Using equation (5), we can generate a summary as a user requests in terms of Treq, the length of the summary. Let M ′ be the actual number of clusters in the top level produced by using the threshold value in (5). Then, we can simply generate a summary by selecting the set of key shots from M ′ clusters.
4 Experimental Results To test the proposed clustering and summarization method, we used a 55 minutes and a 45 minutes TV dramas. We segmented the 55 minutes drama into 399 shots and the 45 minutes drama into 291 shots using traditional histogram difference. We used mean color histogram for a shot feature vector as in [5]. We set the threshold values TC and the window size for the compatibility computation as 0.1 and 4, respectively. With these values and TT of 250 seconds, our clustering converged within 3-4 iterations. Figure 2 shows the compatibilities of the TV drama with window size 4. In this Figure, the peaks correspond to the representative shots and the valleys correspond to the possible scene boundary shots.
Hierarchical Shot Clustering for Video Summarization 1105
Fig. 2. Compatibilities versus shot numbers: (a), (b), and (c) compatibilities after 2, 1, and 0 iterations, respectively.
In order to test the effectiveness of equation (5), we varied Treq and compared the desired number of clusters with the actually generated number of clusters. Table 1 shows the results obtained from the two TV dramas with k = 2 in equation (5). The results show that the number of generated clusters is close to the desired number of clusters. Table 1. Comparison of the desired and generated number of clusters with respect to Treq Treq (Requ est length in minut e) 2
TV Drama I (399 shots: 55 minutes)
TV Drama II (291 shots: 45 minutes)
C (Desired number of clusters)
TT (Threshold value in second)
C′ (Generated number of clusters)
C (Desired number of clusters)
TT (Threshol d value in second)
C′ (Generate d number of clusters)
14
480
14
12
450
9
3
21
320
21
19
284
22
4
28
240
31
25
216
33
5
35
192
31
32
169
36
10
71
95
72
64
84
71
1106 Y. Choi, S.J. Kim, and S. Lee
We defined the number of key shots coming from the different scenes as the performance measure of the proposed summarization method. That is, the summary is better if it represents more scenes in the original video. For this experiment, we obtained the ground truth scene boundaries by manual segmentation. Table 2 shows the results. In Figure 3, we show the summary result of TV Drama I. The images shown in this Figure are the first frames of the selected key shots. Table 2. Performance measure Treq (Reque st length in minute ) 2 3 5
TV Drama I (399 shots: 30 scenes) C′ (Generated number of clusters)
Number of scenes that the key shot represents
14 21 31
12 15 23
Fig. 3 Summary Result
TV Drama II (291 shots: 27 scenes) Number of C′ scenes that (Generated number the key shot of clusters) represents 9 22 33
9 17 24
Hierarchical Shot Clustering for Video Summarization 1107
5 Conclusions In this paper, we presented a hierarchical clustering method based on the compatibility measure that can represent shot correlations. The proposed method can efficiently generate a set of representative shots and also extract the hierarchical structure of a video sequence. Experimental results show that our proposed summarization abridged the original video where compaction is up to 25:1 and still kept most of important scenes. This result is accredited to the clustering capability of extracting the hierarchical structure of a video sequence.
References 1. M Yeung and Boon-Lock Yeo, and Bede Liu “Extracting Story Units from Long Programs for Video Browsing and Navigation”, Proceedings of IEEE International Conference on Multimedia Computing and Systems1996, pp. 296-305. 2. E. Venequ and et al, “From Video Shot Clustering to Sequence Segmentation”, Proceedings of IEEE International Conference on Pattern Recognition 2000, pp. 254-257. 3. Shingo Uchihashi and, et al, “Video Magna: Generating Semantically Meaningful Video Summaries”, Proceedings of ACM International Conference on Multimedia 1999, pp. 383391 4. A. Hanjalic and et al, “Automated High-Level Movie Segmentation for Advanced videoRetrival Systems”, IEEE Transactions On Circuits and Systems for Video Technology, Vol. 9, No. 4, June 1999, pp. 580-588. 5. Ong Lin and Hong-Jiang Zhang, “Automatic Video Scene Extraction by Shot Grouping”, Proceedings of IEEE International Conference on Pattern Recognition 2000, pp. 39-42. 6. Nikolaos D. Doulamis and et al, “Efficient Summarization of Stereoscopic Video Sequences”, IEEE Transactions on Circuits and Systems for Video Technology, Vol. 10, No. 4, June 2000, pp. 501-517. 7. Michael A. Smith and Takeo Kanade, “Video Skimming and Characterization through the Combination of Image and Language Understanding Techniques”, Proceeding of IEEE, pp.775-781. 1997
On Detecting Unsteady Demand in Mobile Networking Environment V.V. Shakhov, H. Choo, and H.Y. Youn School of Electrical and Computer Engineering Sungkyunkwan University, Suwon 440-746, KOREA {vova, choo, youn}@ece.skku.ac.kr
Abstract. One of the key issues in mobile communication system is how to predict the number of calls per each cell. It is an important parameter and usually assumed as random Poisson value. For effective management of cellular network, the average number of calls should be traced and the changes in the numbers need to be promptly detected. In this paper we propose an algorithm detecting the changes in the behavior of the users using the technique proposed for point-of-change problem based only on the number of call arrivals. Computer simulation reveals that the proposed method can effectively detect the discord, and the developed model is very accurate as showing mostly less than 1% differences.
1
Introduction
Recently information technology has been evolving into the direction leading to the convenience of users. Mobile and wireless communication have a significant role in this IT networking era due to the request of users. One of main elements of mobile networking, especially for network design, is the prediction mechanism for the number of calls per unit time in each cell [1]. The number is related to the number of frequency channels, the call blocking probability, and many other key issues of resource management in mobile computing. With steady-state behavior of mobile users, several techniques [1] can be employed for effective mobility management. If the number of mobile users is simply modeled as a constant with a certain level of blocking probability, the required number of channels can be easily obtained. However, as we know, behavior of mobile users and their movement change dynamically. More specifically, the average and maximum number of calls can widely fluctuate. If they increase, the call blocking probability also increases and thus the quality of service drops. Meanwhile, if they decrease, less amount of resource will be required for maintaining the same quality of service. Hence, the number of calls is an important parameter in mobile network. For effective
This work was supported in part by Brain Korea 21 and grant No. 2000-2-30300-0043 from the Basic Research Program of Korea Science and Engineering Foundation. Dr. Choo is the corresponding author.
P.M.A. Sloot et al. (Eds.): ICCS 2002, LNCS 2331, pp. 1108−1117, 2002. Springer-Verlag Berlin Heidelberg 2002
On Detecting Unsteady Demand in Mobile Networking Environment 1109
management of cellular network, the average number of calls should be traced and the changes in the numbers need to be promptly detected. Sometimes it is enough to use a simple deterministic model for calculating the number of calls, where the parameters of the model are known. Usually a flow of call arrivals is represented by a stochastic model, and the number of calls is a random value decided depending on the behavior of users. Several models for analyzing the movement of users were proposed in [2, 3]. The models require some additional information such as the speed of mobile users, the direction of movements, travel path, and so on. Since a lot of factors affect the system performance, an approximation technique is employed for simplifying the modeling. For example, assumption of uniform distribution for the speed of mobile users, a cellular network as a set of identical hexagons, etc. Note, however, that movement of users is irregular, and the assumptions employed for the sake of simplicity can cause unrealistic modeling for the average number of calls. In this paper, therefore, it is offered to detect the change in the behavior of the users using the point-of-change problem solution technique based only on the number of call arrivals. The rest of the paper is organized as follows. Section 2 provides the basic notation and previous works. The main algorithm of discord detection under the Poisson distribution and an admissible lag is presented in Section 3. In Section 4, we present the results of numerous simulation experiments with the proposed algorithm. Section 5 is a brief conclusion.
2
Preliminaries
To calculate the required number of frequency channels or traffic load in a cell, we have to know the number of calls per unit time which is a random number. Let N c be the required number of calls, N be the total number of users in a cell, L be the probability of a user to have a mobile terminal, and k be the probability of mobile users making network access. Then N c = N Lk. Here, N depends on the average distance between mobile users and k depends on the speed of mobile users [1]. Actually, the larger the sojourn time for a mobile user is, the higher the network access probability is. If traffic congestion occurs, the average distance between adjacent mobile terminals decreases, and thus the sojourn time increases. Hence, both N and N c increase. In this situation, more channels are required for customers to have a certain level of quality of service. Refer to Figure 1. In this figure, Ni , ki and Nic are the parameters discussed above before (i = 1) and after (i = 2) traffic congestion. Here we assume that each user carries a mobile terminal (L = 1) and two channels are assigned to a cell. The number of calls can be predicted by mathematical modeling. In this section some models proposed for modeling the behavior of the users are discussed. Their approaches are usually based on a certain regularity of user movement. The models proposed in [2, 4] assume that the average speed of users and their directions are same in each cell. However, movements of customers are usually
1110 V.V. Shakhov, H. Choo, and H.Y. Youn
different in practice, and even the information on the direction and speed of the users are unavailable. In [5], the authors consider irregular movements of mobile users. For predicting the future location of mobile users a Markov chain model and database of the movement patterns are used. This scheme does not work if preliminary information such as pattern database is either absent or obsolete. On the other hand, an approach for predicting the behavior of the users based only on current location information is considered in [6]. This scheme is costly because no previous information is used while highly accurate data observation is necessary. A mobility prediction scheme based on neuro-fuzzy theory is offered in [3]. Here a unique reason can affect the behavior of a user, and thus all previous information and regulations can be out-of-date.
Fig. 1. The losses before and after traffic congestion.
We know that the movement of mobile users may be irregular, and thus prediction of their behavior is very difficult. However, we can observe and analyze the number of calls in a cell, which is considered as a random value. If the demand is steady we have a sequence of random values under one distribution. Otherwise, the distribution of the number of calls is changeable. Usually the following assumption is acceptable, where the number of calls ξ in a unit time is independent and identically distributed Poisson random value [7]. P (ξ = k) =
λk −λ e , k = 0, 1, . . . . k!
(1)
The number of calls before traffic congestion has the Poisson distribution with parameter λ1 , and it has the same distribution with rate λ2 after traffic congestion occurs. It is necessary to detect the demand change, in the strict sense, the moment when the distribution change occurs should be found. The detection problem is known as a point-of-change problem or a discord problem. It is a very popular problem [8, 9], and minimization of the time between the moment of discord and the moment of alarm is an important issue. As far as the channel resource is available, the quality of service is satisfied and the delay for the discord detection can be tolerated. In this paper, we focus on a point-
On Detecting Unsteady Demand in Mobile Networking Environment 1111
of-change problem with acceptable delay, and propose a technique based on a point-of-change problem with admissible lag.
3 3.1
The Point-of-Change Problem with Poisson Distribution Problem Statement and Discord Detection Algorithm
Let a sequence of independent random values ξi , i = 1, 2, . . ., be under consideration. ξi denotes the number of calls in time unit i. Before discord, the distribution of ξi is modeled by the Poisson distribution of Eq. (1) with parameter λ1 . In unknown point of time tD the intensity of flow of calls is changed. It means that the distribution parameter of ξi becomes different from λ1 . We denote the distribution parameter after the discord as λ2 . Without loss of generality, we assume that λ1 ≤ λ2 . The other case can be easily modeled with simple modification. It is desirable and sometimes required to alarm system managers if a discord takes place in advanced systems. Let us designate the moment of alarm as tA and T be an admissible lag. Then we know that a discord is detected if tD < tA ≤ tD + T . If tA < tD , it must be a false alarm. Now we propose the following algorithm for discord detection. Discord Detection Algorithm T Step 1. Calculation of the sum ST = i=1 ξi . Let us denote the variate ST has Poisson distribution [10]. Step 2. If Sn > h, then the system raises an alarm and go to Step 5. Otherwise, proceed to the next step. Here n takes the value T, T + 1, . . .. Note that h is a threshold value calculated prior to the beginning of the algorithm. It is considered in detail below. Step 3. The following sum is computed Sn = Sn−1 + ξn − ξn−T , where n = T + 1, T + 2, . . . It is obvious that the variate Sn also has the Poisson distribution because Sn is a sum of independent Poisson random values. Hence, if the distribution parameter of ξi is λ, the distribution parameter of Sn becomes T λ. Step 4. Go to Step 2. Step 5. End. The value of h decides the tradeoff between the rate of discord detection and false alarm. For example, if h = 0, discord will be unambiguously detected, but a false alarm will be generated on each observation. If h value is high, false alarm will be seldom declared, but the probability of discord detection will be very low. Let us discuss the basic concept of the proposed method. The random values ξn−T , . . . , ξn are under consideration. Assume that the probability of a false alarm is fixed if the following restriction is true.
1112 V.V. Shakhov, H. Choo, and H.Y. Youn
P (Sn > h, n < tD ) ≤ α.
(2)
Here α is a constant representing the degree of false alarm. Before applying the method, it needs to be decided according to how often false alarm can be tolerated. The second issue is the number of discord omissions. A discord is lost if Sn < h and n ≥ tD + T . Hence, the probability of omission is P (Sn < h, n = tD + T ).
(3)
For the quality of discord detection algorithm, it needs to minimize the probability of Eq. (3), while taking into account Eq. (2). It is achieved with an optimal choice of threshold value hopt . Theorem 1. For the proposed method the optimal choice of threshold is hopt = min{h :
h (λ1 T )k > eλ1 T (1 − α)}. k!
(4)
k:=0
The probability of omission is then hopt
(λ2 T )k e−λ2 T . k!
k:=0
Proof: Let us consider Step 2 of the algorithm. The decision rule is true, but the discord is not found (n < tD ). Hence, a false alarm takes place and Sn has the Poisson distribution with parameter T λ1 . It is clear that Sn is a positive integer. Hence, it can be assumed that h is a positive integer. We have
P (Sn > h, n < tD ) = 1 − P (Sn < h, n < tD ) = 1 −
h k (λ1 T ) k=0
k!
e−λ1 T ≤ α.
Thus, if the condition of Eq. (2) is true, then the threshold value belongs to a set h k (λ1 T ) ≥ (1 − α)eλ1 T }. {h : (5) k! k=0
Without loss of generality we assume that tD = 0. An omission of pointof-change occurs if Sn < h and n ≥ T . The aim of the proposed method is to minimize the probability of Eq. (3) under the condition of Eq. (2). Hence, it needs to minimize the probability P (ST < h). h can take a value from the set of Eq. (5). Obviously, the discord omission probability decreases if the threshold decreases, but the probability of false alarm increases. So, an optimal value of threshold is the least possible value of the set of Eq. (5). Hence, we have Eq. (4).
On Detecting Unsteady Demand in Mobile Networking Environment 1113
At the same time hopt
P (ST ≤ hopt ) =
hopt
P (ST = k) =
k:=0
(λ2 T )k e−λ2 T . k!
k:=0
This contradiction proves the theorem.
✷
The following theorem is needed for the sequel. Theorem 2. hopt does not depend on λ2 . That is, the distribution after discord does not affect the optimal threshold value. The proof directly follows Eq. (4). 3.2
Varying Parameter after Discord
The rate of distribution after discord is a constant λ2 as defined earlier. Sometimes it is possible to have unsteady rate such that the random value ξtD +1 has a Possion distribution with the rate λ2 − δ1 , ξtD +2 with the rate λ2 − δ2 , . . ., ξT with the rate λ2 − δT . Here, we have δi < λ2 . It is possible that δj = 0, δj+1 = 0, . . . δT = 0; 0 ≤ j ≤ T . In this case the quality of the proposed method is described by the following theorem. Theorem 3. The probability of discord omission is hopt
P (ST < hopt , tD = 0) =
T (λ2 T − ∆)k e−(λ2 T −∆) , ∆ = δi . k! i=1
k:=0
Here hopt is calculated using Eq. (3). Proof: By Theorem 2, hopt is obtained by Eq. (3). Since ξ1 , . . . , ξT are independent Poisson random values, the sum of those values, ST , has the Poisson distribution too. The parameter of the distribution of ST is equal to T λ2 − ∆. So, hopt
P (ST < hopt , tD = 0) =
(λ2 T − ∆)k e−(λ2 T −∆) . k!
✷
k:=0
4
Performance Evaluation
In this section we analyze the quality of the proposed method through computer simulation. For it, a pseudorandom generator for Poisson distribution is necessary, and the choice of best generator is discussed. Also, we compare the analytical data and simulation data, and verify that the degree of false alarm is limited. The proposed algorithm is tested for both the constant and varying parameters of the distribution. For each simulation experiment, 106 runs are averaged to show the sensitiveness of the algorithm to the lag and parameters of distributions.
1114 V.V. Shakhov, H. Choo, and H.Y. Youn
4.1
Pseudorandom Generator for Poisson Distribution
For the performance evaluation of mobility management such as mobile location management and hand off schemes, simulation is frequently employed. Usually the ranges of parameter values of wireless network are quite large, and it results in complex calculation and large simulation time. Therefore, it is important to choose a fast pseudorandom number generator. In this paper we use a sequential random values of Poisson distribution for modeling the number of calls per unit time. There exist many algorithms generating Poisson pseudorandom numbers [10–12]. They are compared taking into account the specific property of the target problem. The methods in [10, 11] use the idea that if the time between arrivals is exponentially distributed, then the number of arrivals have the Poisson distribution. We also need to use a uniform pseudorandom number generator. The method [12] uses only one uniformly distributed numbers. In Table I below the result of the generators tested is given. It shows that the algorithm from [12] is the best for average and large values of the Poisson distribution parameter. Method (A) of Molloy and Knuth[10, 11] have an advantage for small value parameters.
Table I. The time for generating 1,000,000 pseudorandom numbers (nanosecond). Distribution rate 0.0001
Ermakov 38
Knuth; Molloy (A) 33
Knuth; Molloy (B) 44
0.01
38
33
44
0.1
39
38
44
0.2
39
44
44
0.3
38
50
44
0.5
43
47
50
1
47
61
58
2
49
99
71
3
66
137
82
5
82
198
109
10
143
363
176
100
1027
3279
1373
300
3010
9694
4010
500
5141
16323
6712
On Detecting Unsteady Demand in Mobile Networking Environment 1115
4.2
Verification of Limited Degree of False Alarm
According to Eq. (2), the proposed method gives a small number of false alarm. In each test the method employs a random value T , and a counter is increased if a false alarm takes place. Let the degree of false alarm be equal to or less than 5% throughout the tests. An admissible lag is selected among different values. Table II. The degree of false alarm. λ2 0.01
T 20
hopt 5
False alarm(%) 2.0
1
10
15
5.0
2
10
28
3.4
3
10
39
4.8
3
20
73
4.3
3
50
170
4.9
3
5
22
3.3
3
3
14
4.0
3
1
6
3.5
10
10
117
4.2
50
10
537
4.8
Each row in Table II is a result of simulation experiment for different λ2 values. At the beginning of each experiment, the optimal threshold is calculated based on Theorem 1 and the value is used for all the runs. The number of false alarms is in the column for false alarm which is given in percentage. We see that the degree of false alarm is always lower than 5%. 4.3
The Efficiency of the Proposed Algorithm
Let us find out the quality of the proposed discord detection algorithm. Without loss of generality we assume that tD = 0 for discord detection. So, we have a distribution with a discord at the beginning. The simulation method is the same as above and the degree of false alarm is equal to 5%. The discord detection rates from the simulation and analytical model are shown in Table III. Notice that they are very close, which verifies the effectiveness of the proposed model. The proposed discord detection algorithm always detects a demand change for relatively large distribution parameter values even though λ1 and λ2 are close and the admissible lag is not large. Let us now consider the case of non-constant rate after the point-of-change. Let a random value ξi , i = 1, . . . , T , after the discord have the parameter of distribution λ1 + δi, if λ1 + δi < λ2 . Otherwise, the rate is equal to λ2 . In other words, the rate of distribution after the discord changes gradually with a constant step δ up to λ2 . The results for δ = 0.1 are given in Table IV.
1116 V.V. Shakhov, H. Choo, and H.Y. Youn Table III. The efficiency of the algorithm. λ1 1
λ2 1.2
T 10
hopt 15
Simulation 15.7
Model 15.6
1
1.2
100
117
58.2
58.5
1
1.2
300
329
94.9
94.8
1
3
10
15
100
100
2
3
20
28
87.1
86.5
3
5
15
56
98.8
98.7
2
4
30
73
100
100
5
10
20
117
100
100
10
11
100
346
100
100
100
120
10
346
100
100
Table IV. Non-constant rate after discord. λ1 1
λ2 2
T 10
Simulation 48.5
Model 48.3
1
2
20
91.1
88.3
10
11
10
12.1
12.2
10
11
20
29.7
26.8
1
3
10
88.4
88.9
100
120
10
100
100
We see that the quality of the proposed method in the case of varying distribution parameter is worse than that of the constant case. However, it is not true if T is relatively large. In Figure 2 the relationship between the quality and T is demonstrated for the case of λ1 = 2, λ2 = 3, and δ = 0.1.
5
Conclusion
It is important to know the number of calls for moving users for effective management of cellular networks. The number depends on the behavior of customers, which is usually irregular and very difficult to model. The method for detecting the demand change in terms of the numbers has been presented in this paper. The point-of-change problem under admissible lag has been stated and analyzed by computer simulation with Poisson random value. The proposed algorithm for detecting the demand change shows high accuracy under heavy load, and the developed model can accurately predict the performance of the algorithm.
On Detecting Unsteady Demand in Mobile Networking Environment 1117
Fig. 2. The relationship between the detection rate(%) and admissible lag; λ1 = 2, λ2 = 3.
References 1. W. Lee, Mobile Cellular Telecommunications: analog and digital systems, Second Edition, New York: McGraw-Hill, 1995. 2. R. Thomas, H. Gilbert, and G. Mazziotto, “Influence of the Movement of Mobile Station on the Performance of the Radio Cellular Network,” Proceeding of the 3rd Nordic Seminar, Copenhagen, September 1988. 3. C.Y. Park, Y.H. Han, C.S. Hwang, and Y.S. Jeong, “Simulation of a Mobility Prediction Scheme Based on Neuro-Fuzzy Theory in Mobile Computing,” Simulation, v. 75, No 1, p. 6-17, 2000. 4. D. Hong and S. Rappaport, “Traffic Model and Performance Analysis for Cellular Mobile Radio Telephone Systems with Prioritized and Non-Prioritized Handoff Procedures,” IEEE Transactions on Vehicular Technology, v. 35, No 3, pp. 77-91, August, 1986. 5. G. Liu and Jr. Maguire, “ Class of Mobile Motion Prediction Algorithms for Wireless Mobile Computing and Communications,” Mobile Networks and Application, v. 1, pp.113-121, 1996. 6. T. Lui, P. Bahl, and I. Chlamtac, “An Optimal Self-Learning Estimation for Predicting Inter-cell User Trajectory in Wireless Radio Networks,” IEEE 6th International Conference on Universal Personal Communications, v. 2, pp.438-442, 1997. 7. Y. Fang, I. Chlamtac, and Y.-B. Lin, “Call completion probability for a PCS network”, IEEE 6th International Conference on Universal Personal Communications, v. 2, pp. 567-571, 1997. 8. E.S. Page, Continuous inspection schemes. Biometrika, 1954, v.41, N2, p.100-114. 9. B. Yakir, “Dynamic sampling policy for detecting a change in distribution, with a probability bound on false alarm,” Ann. Statist, v. 24, pp. 2199-2214, 1996. 10. M. Molloy, Fundamentals of Performance Modeling, New York : Macmillan Publishing Company, 1988. 11. D.E. Knuth, The art of computer programming, 3rd ed., v.2, pp. 137, AddisonWesley, 1998. 12. S.M. Ermakov and G.A. Michaelov, Statistics modelling, oscow: Nauka, 1982 (in Russian).
Performance Modeling of Location Management Using Multicasting HLR with Forward Pointer in Mobile Networks Dong Chun Lee1 Sung-Kook Han2 , and Young Song Mun3 1
Dept. of Computer Science Howon Univ., Korea e-mail:[email protected] 2 Dept. of Computer Eng. Wonkwang Univ., Korea 3 Dept. of Computer Science Soongsil Univ., Korea
Abstract. We propose a new location management called Multicasting HLR with Forward Pointer (MHFP) which exploits receiver side call locality in Mobile Networks (MN). When a call is established, Multicasting HLR (MH) records the caller’s VLR ID according to the callee. Periodically, MH ranks the VLRs and determines which VLRs frequently make calls to the callee. During a location registration process, MH sends the terminal’s location information to the determined VLRs. And also, when a mobile terminal frequently moves between two Registration Areas (RAs) or within a small area, the terminal’s location information does not register to HLR but to VLR using the Forwarding Pointer (FP) with unit length and link both sides.
1
Introduction
For mobility management scheme in the MN, the standard commonly used in North America is the EIA / TIA Interim Standard 41 (IS-41), and in Europe the GSM[2]. And whenever a terminal crosses a RA or a call originates, the HLR should be updated or queried. Frequent DB accesses and message transfers may be cause the HLR bottleneck problem and then degrade the system performance. A number of related works have been reported to reduce overhead traffic of the HLR. In [7], [8], a Location Forwarding Strategy is proposed to reduce the signaling costs for location registration. A Local Anchoring Scheme is introduced in [1], [4]. Under these schemes, signaling traffic due to location registration is reduced by eliminating the need to report location changes to the HLR. Hierarchical database system architecture is introduced in [3]. These schemes can reduce both signaling traffics due to location registration and call tracking using the properties of call locality and local mobility. We propose a new location management scheme to reduce the location overhead traffic of HLR.
2
IS-41 Standard Scheme
The whole MN coverage area is divided into cells. Each mobile terminal within a cell communicates with the network through a Base Station (BS) which is P.M.A. Sloot et al. (Eds.): ICCS 2002, LNCS 2331, pp. 1118−1127, 2002. Springer-Verlag Berlin Heidelberg 2002
Performance Modeling of Location Management
1119
installed inside the cell. These cells are grouped together to form larger areas called RAs. All BSs, belonging to a given RA, are wired to a Mobile Switching Center (MSC). In this paper, we assume that the VLR is co-located with the MSC and a single HLR in the network [4]. In order to locate a terminal effectively when a call arrives, each terminal is required to report its location whenever it enters a new RA. We call this reporting process location registration. In order to track a call to the proper terminal, the HLR and the VLRs are queried to find the current RA, and all cells within the RA are paged to find it. Within the call tracking, we call the queries to the HLR and the VLRs as the Search. According to the IS-41 location strategy, the HLR always knows exactly the ID of the serving VLR of a mobile terminal. We outline the major steps of the IS-41 location registration. (For the details, refer to [4], [9].) 1. The mobile terminal sends a Registration Request (REGREQ) message to the new VLR. 2. The new VLR checks whether the terminal is already registered. If not, it sends a Registration Notification (REGNOT) message to the HLR. 3. The HLR sends a Registration Cancellation (REGCANC) message to the old VLR. The old VLR deletes the information of the terminal. and the IS-41 call tracking is outlined as follows: 1. The VLR of caller is queried for the information of callee. If the callee is registered to the VLR, the SEARCH process is over and the call is established. If not, the VLR sends a Location Request (LOCREQ) message to the HLR. 2. The HLR finds out to which VLR the callee is registered, and sends a Routing Request (ROUTREQ) message to the VLR serving the callee. The VLR finds out the location information of the callee. 3. The serving MSC assigns a Temporary Local Directory Numbers (TLDN) and returns the digits to the VLR which sends it to the HLR. 4. The HLR sends the TLDN to the MSC of the caller. 5. The MSC of the caller establishes a call by using the TLDN to the MSC of the callee. Among the above 5 steps, the call tracking process is composed of step 1 and step 2.
3
MHFP Scheme
When a call is established, MH records the caller’s VLR ID according to the callee. Periodically, MH ranks the VLRs and determine which VLRs frequently make calls to the callee. During a location registration process, MH sends the terminal’s location information to the determined VLRs. And also, when a mobile terminal frequently moves between two Registration Areas (RAs) or within a small area, the terminal’s location information does not register to HLR but to VLR using the Forwarding Pointer (FP) with unit length and with both side links. Fig. 1 shows one case of location registration when lenth of FP is longer
1120 D.C. Lee, S.-K. Han, and Y.S. Mun
than 1, and the VLR in the starting point of FP is not multicasting. And Fig. 2 also shows one case of call tracking when the receiving terminal has multicasting data, and the VLR has FP of the receiving terminal. HLR
HLR
(4) REGMULT
(4) REGCANC
(3) REGNOT
(4) TLDN
(4) REGMULT
(2) REGNOT
VLR
VLR Old FP
Multicasted Pointer(MP)
(3) ROUTREQ
(2) ROUTREQ VLR
VLR
VLR
VLR
VLR
VLR
FP
New FP (2) REGREQ MP
(1) CALLREQ Caller
Fig. 1. Location registration(m=2).
MP
Caller
Fig. 2. Call tracking.
Here REGMULT represents the registration multicasting and m represents the number of VLR to multicast the location information. We outline the major steps of location registration as follows (see Fig. 1): · The REGREG message from a terminal is transferred to the VLR that messages a new RA. The VLR transfers the REGNOT message to previous VLR. Previous VLR performs the query by terminal number. · If the FP of RT exists in the end point in query result, · The previous VLR of terminal transfers REGNOT message including FP information to the HLR. The previous VLR of terminal make new FP to current VLR of terminal. The HLR performs the query by number of received terminal. · If length of FP is longer than 1, and the VLR in the starting point of FP is not multicasting , · The REGCANC message is transferred to the VLR which has previous FP of terminal. The REGMULT message is transferred to the multicasting VLRs. · If length of FP is longer than 1, and the VLR in the starting point of FP is multicasting , · The REGMULT message is transferred to the multicasting VLRs. · Else if length of FP is less than 1, · Previous VLR of terminal makes new FP to current VLR of terminal. And call tracking is outline as follows (see Fig.3): · The sending terminal (ST) requests a call to the VLR. The VLR performs the query by number of RT. · If the RT exists in query result, · The TLDN is assigned to the RT via the MSC. · Else if the FP of RT exists in query result, · The ROUTREQ message is transferred to the VLR directed by the FP. The VLR directed by the FP assigns TLDN via the MSC and transfers to the VLR of ST.
Performance Modeling of Location Management
1121
· Else if the multicasting information of the RT exists in query result, · The ROUTREQ message is transferred to the VLR directed by multicasting information. The VLR directed by multicasting information performs the query by number of RT. · If RT has multicasting information and the VLR has in the RT in query result, · The TLDN is assigned to the RT via the MSC and is transferred to the VLR of sending terminal. · Else the RT has multicasting information, and the VLR has the FP of RT, · The ROUTREQ message is transferred to the VLR directed by the FP. The VLR directed by the FP assigns TLDN via the MSC and transfers to the VLR of ST. · Else if the multicasting information of the RT doesn’t exists in query result, · The VLR of ST transfers LOCREQ of message to the HLR. The HLR find the VLR of RT by query and transfers to ROUTREQ. The VLR which receive ROUTREQ message performs the query by number of RT. · If the RT must search to the HLR and the VLR directed by the HLR has the RT, · The TLDN is assigned to the RT via the MSC and is transferred to the HLR. The HLR transfer the TLDN to the VLR of ST. · Else the RT must search to the HLR and VLR directed by HLR has the FP of RT, · The ROUTREQ message is transferred to the VLR directed by the FP. The VLR directed by the FP assigns TLDN via the MSC and transfers to the VLR of ST. The HLR transfer The TLDN to the VLR of ST. In selection of multicasting objects, m, Fig. 3 shows the conceptual structure of the CL field and RL field. VLR ID 1
Counter
VLR ID 2
Counter
. .
(a) Caller list field. VLR ID 1
VLR ID 2
VLR ID 3
. .
(b) Rank list field. Fig. 3. Conceptual structure of the CL field and RL field.
Fig. 3 (a), the CL field consists of the pairs of the ID of frequently calling VLR and number of calls from it. Every established call adds a new pair or increases the number of existing calls. The CL fields are distributed over the HLR and multicast VLR recording the established calls, and merged into HLR periodically to construct the RL field. In Fig.3 (b), each element of RL field is an ID of VLR and the ID of VLR calling more frequently course before. The RL field exists only in the user profile of HLR while the CL field is dispersed over VLR and HLR.
1122 D.C. Lee, S.-K. Han, and Y.S. Mun
4
Analytical Model
Form the user’s point of view, the end-to-end service delay (location registration or call tracking) will be an important performance metric. To evaluate this end-to-end delay, we treat the database of MN as Jackson’s network. The service time of each database operation is assumed to be a major delay, and we do not consider a link cost [5], [6]. We assume that there are n VLRs and one HLR in the system. The HLR is assumed to have an infinite buffer and single exponential server with the average service time µ1h . Likewise, the VLR is assumed to have an infinite buffer and single exponential server with the average service time µ1v . We assume that within a RA, the location registration occurs in a Poisson process with rate λu and the call origination occurs in a Poisson process with rate λc . With these assumptions, the MN using IS-41standard as a mobility management method becomes Jackson’s network [10] as shown in Fig.4. λlr and λtt represent the average arrival rate of REGCANC message and the average arrival rate of ROUTREQ message, respectively. λh represents the average arrival rate of messages to the HLR from other VLRs, and by the Burke’s theorem [10] it is the same as the average departure rate of messages from the HLR. Pvo is the probability the departure message from the VLR leaves the system. Pvh is the probability the departure message from the VR enters the HLR. Phv is the probability the departure message from the HLR enters one of n VLRs. From the definition of λlr and λtt , we know that these messages get out of the system after going through the VLR.
Phv = lnh
Pvh =
Pvo =
lh
n- 1 n l c +l u l lr + l tt + l c + l u
1 n l c + l lr + l tt l lr + l tt + l c + l u
l lr l tt
l cl u
Fig. 4. Jackson’s network modeling of the IS-41standard scheme.
λu enters the VLR in the form of REGREQ message, and after receiving services from the VLR it is delivered to the HLR in the form of REGNOT message. After receiving services from the VLR, n1 λc get out of the system because the probability the callee is in the same RA with the caller is n1 . So, we have 1
λ +λ +λ
c tt lr Pvo = λlrn +λ , tt +λc +λtt
(1)
Performance Modeling of Location Management n−1
λ +λ
1123
c tt n Pvh = λlr +λ , tt +λc +λtt
(2)
Phv = λnh .
(3)
By the property of Jackson’s network, we have λh =n×( n−1 n λc + λtt ). From
λh n =λlr
(4)
+ λtt , we have
λlr =λtt ,
(5)
λtt = nn1 λc .
(6)
Though Eqs. (5) and (6) could be directly inferred from the definitions of λc and λu , we lead these equations through Eqs. (1)-(4) to help understand how λc and λu are delivered to HLR and VLRs and how many messages arrive at these DBs on average using the property of Jackson’s network. We will repeat these steps in the following Jackson’s network modeling the proposed scheme. Now, let Wv and Wh represent the average system time (queue plus service) in the VLR and the average system time in the HLR, respectively. By the Little’s Rule [10], Wv and Wh becomes Wv = µv −(λlr +λ1tt +λc +λtt ) = µ
v −(
1
2n−1 n λc +λtt )
,
Wh =µh − n( n−1 n λc + λtt ).
(7)
(8)
Likewise, Fig.6 shows Jackson’s network model of MHFP scheme. The focus of message flow aims at the quantity. Phv =
pvv =
n- 2 nm + n - m (1 - q)l u + 2 (1 - p)l c n n
2 (n - 2) p n- 2 lu+ lc+ 5l c + (1 - p)l c n n 2 n l lr + l tt + l u + l c n- 2 (1 - p)l c n l lr + l tt + l u + l c
(1 - q)l u + 2 pvh =
(q + ( Pvo =
nm + n - m (1 - q )) l u + l n l lr + l tt + l u + l c
c
l lr l tt l u + l
c
Fig. 5. Jackson’s network modeling of the MHFP scheme.
1124 D.C. Lee, S.-K. Han, and Y.S. Mun
The wide difference between Fig.4 and Fig. 5 is that the number of message out coming HLR. When call managed in HLR, passing one more device dealing with multicasting can solve this problem. Another difference is what communicates call between VLR passing HLR. It is because of obtaining location information of the receiving terminal by multicasting information, m. We need to calculate the probability that the departure message from the VLR is delivered to another VLR. We denote this probability by Pvv . Considering the definition of λlr , λtt , λc , λu as follows: Pvo =
(q+ nm+n−m (1−q))λu +λc n . λu +λlr +λc +λtt
Pvh =
(1−q)λu +2 n−1 n (1−p)λc . λu +λlr +λc +λtt
Pvv =
n−2 2 λu + n λc + n p 2 5λc + n (1−p)λc . λu +λlr +λc +λtt
(11)
Phv = nm+n−m (1 − q)λu + 2 n−1 n n (1 − p)λc .
(12)
(9)
(10)
(n−2)
P represents probability that multicasting message search to VLR of receiving terminal ,and q is probability that the FP has less than 1. By the definition of λtt and λlr , we obtain λtt =2 n−1 n (1 − p)λc ,
(13)
λlr = nm+n−m (1 − q)λu . n
(14)
From the property of Jackson’s network, we have λh =n(2 n−1 n (1 − p)λc +
nm+n−m (1 n
− q)λu ).
(15)
Let Wv and Wh denote the average system time in the VLR and in the HLR, respectively, From the Little’s Rule, Wv and Wh are: Wv = µv −(λtt +λ1lr +λc +λu ) 1 = µ −(2 n−1 (1−p)λ + nm+n−m (1−q)λ v
n
Wh = µ
v −(2
c
n
1
u +λc +λu )
n−1 nm+n−m (1−q)λu ) n (1−p)λc + n
.
.
(16)
(17)
Table 1 shows the average number of arrival messages to the VLR, and the average system time in the VLR when the proposed scheme is used as
Performance Modeling of Location Management
1125
Table 1. VLR comparisons between the MHFP and IS-41scheme
IS-41
Average number of arrival messages 2n−1 λc +2λu n
MHFP
(2 n−1 (1 − p)λc n
Average system time in VLR
+ nm+n−m (1 − q)λu + λc + λu ) n
1 µv −( 2n−1 λc +2λu ) n 1 µv −(2 n−1 (1−p)λc + nm+n−m (1−q)λu +λc +λu ) n n
Table 2. HLR comparisons between the MHFP and IS-41scheme
Average number of arrival messages IS-41 n( n−1 n λc + λu ) MHFP
n(2 n−1 n (1 − p)λc + nm+n−m (1 − q)) n
Average system time in VLR 1 µh −n( n−1 λc +λu ) n 1 n−1 nm+n−m µh −n(2 n (1−p)λc + (1−q)λu ) n
compared to those of IS-41scheme. And table 2 shows the average number of arrival messages to the HLR, and the average system time in the HLR. As shown in table 1 and table 2, the proposed scheme distributes messages from the HLR to the VLRs, and it also reduces the average system time in the HLR with the small increase of the average system time in the VLRs. Based on the delay times of the HLR and the VLR, we can calculate the mobility management costs for IS-41 standard and the proposed scheme as follows. • The IS-41scheme for mobility management cost. (1) WIS−41L (Location registration cost) = Wv +Wh +Wv . (2) WIS−41C (Call tracking cost) = n1 Wv + n−1 n (Wv +Wh +Wv ). (3) WIS−41M (Mobility management cost) u c = λcλ+λ ×WIS−41L + λcλ+λ ×WIS−41C . u u • The proposed scheme for mobility management cost. (1) WP roposedL (Location registration cost) =q × 2Wv + (1 − q)(2Wv + W h). (2) WP roposedC (Call tracking cost) p n−2 (1−p) =( n1 ) × 4Wv + n−2 n × 2 × 7Wv + n 2 (7wv + 4wh ). (3) WP roposedM (Mobility management cost) u c × WP roposedL + λcλ+λ × WP roposedC . = λcλ+λ u u
5
Numerical Results
To get numerical results, we use the same value of system parameters as those in [2], n= 128, for example. From these parameters, the average occurrence rate of location registration in an RA, λu , is calculated as λu =
1126 D.C. Lee, S.-K. Han, and Y.S. Mun
5.85/s. And the average call origination rate in an RA, λc , is calculated as λc = 8.70/s. We assume that the average service rates of HLR and VLR are µh = 2000/s,µv =1000/s, and Probability p is 0.5. In figures, X-coordinate shows probability q that the FP has less than 1 and Y-coordinate shows the ratio dividing the result value in IS-41scheme into the result value in proposed scheme. If the ratio value is 1, the IS-41scheme performance is equal to proposed scheme performance. If it has greater than 1, the performance of proposed scheme is superior to the performance of IS-41scheme. And if it has less than 1, the performance of IS-41scheme is superior to the performance of proposed scheme. The example of the lower graph can show m of multicasting factor. It is determined whether we multicast the location information of terminal into several VLRs. In Fig.6, we can know that the results of performance are superior to IS41scheme. As multicasting the location registration, this result becomes a matter of course. If m increases gradually, the level of performance is lower. In Fig.7, regardless of m, the graph is determined by the probability q, but m has no effect upon the value. Otherwise, a relation of between m and probability q is actually represented by an expression but also, it is facts that m has no effect upon the value. When m is 3, we may assume that probability q is 0.5. If probability q is 0, the number of the call tracking message is equal to the number of the location registration message. Fig.8 shows message ratio in mobility management cost. If probability q is 0.5, and m is 3, the value is equal to 0.87 in short. It is fact that proposed scheme is 1.15 times as many messages as IS-41scheme. 0 .8
1.8
0 .7
1.6 1.4
IS - 41/MHFP
IS - 41/MHFP
0 .6 0 .5 0 .4 0 .3 0 .2
1 0.8 0.6 0.4
m =1
0 .1 1 .0
1.2
0 .9
0 .8
m=2
0 .7
0 .6
m=3
0 .5
0 .4
m=4
0 .3
0 .2
m=5
0 .1
m =1
0.2 1.0
0 .0
0.9
0.8
m=2
0.7
0.6
m=3
0.5
0.4
m=4
0.3
0.2
m=5
0.1 0.0
q
q
Fig. 6. Message number ratio in registration cost.
Fig. 7. Message number ratio in call tracking cost.
1 .4 1.2
IS - 41/MHFP
1 0.8 0.6 0.4 m =1
0.2
1 .0
0 .9
0 .8
m=2
0 .7
0 .6
m=3
0 .5
0 .4
m=4
0 .3
m=5
0 .2
0 .1 0 .0
q
Fig. 8. Message number ratio in mobility management cost.
Performance Modeling of Location Management
6
1127
Conclusions
We propose MHFP scheme to reduce the location traffic of HLR. In numerical results, the proposed scheme has lower value in the number of call tracking message and the management delay time of HLR than the IS-41scheme. Especially, the management delay time of HLR has more performance than the IS-41scheme. The proposed scheme can be expected to have the prominent performance advancing for mobile management in the 3G mobile networks.
Acknowledgments This work is supported in part by the ministry information Communication Communication of Korea, under the ”Support Project of University information Technology Research Center(ITRC)” supervised by KIPA”.
References 1. I. F. Akyildiz, J, McNair, J, Ho, H. Uzunalioglu, W. Wang, ”Mobility Management in Current and Future Communication Networks,” IEEE Network, Jul./Aug., 1998, pp. 39-49 2. R. Jain, Y. B. Lin, C. N. Lo., and S. Mohan, ”A Caching Strategy to Reduce Network Impacts of PCS,” IEEE Jour. on Selected Areas in Com., Vol. 12, No. 8, 1994, pp. 1434-1445 3. C. L. I, G. P. Plooini and R. D. Gitlin, ”PCS Mobility Management using the Reverse Virtual All Setup Algorithm,” IEEE/ACM Trans. Vehicle. Tech, 1994, pp. 1006-1010 4. T. Russel, Signaling System 7, McGraw-Hill, 1995 5. J. Z. Wang, ”A Fully Distributed Location Registration Strategy for Universal Personal Communication Systems,” IEEE Personal Com., First Quarter 1994, pp. 42-50995. 6. C. Eynard, M. Lenti, A. LOmbardo, O. Marengo, S. Palazzo, ”Performance of Data Querying Operations in Universal Mobile Telecommunication System (UMTS).”Proceeding of IEEE INFOCOM ’95, Apr. 1995, pp. 473-480 7. T. H. Cormen, C. E. Leiserson, R. L. Rivest, Introduction to Algorithms, MIT Press and McGraw-Hill, 1990 8. N. Shivakumar, J. Widom, ”User Profile Replication for Faster Location Lookup in Mobile Environment,” Proceeding of ACM MOBICOM, Nov. 1995, pp. 161-169 9. Y. -B. Lin, ”Reducing Location Update Cost in a PCS Network,” IEEE/ACM Trans. Networking Vol. 5, Feb. 1997, pp. 25-33 10. L. Kleinrock, ” Queuing Systems”, Vol. 1, Wiley-Interscience, 1975
Using Predictive Prefetching to Improve Location Awareness of Mobile Information Service Gihwan Cho Division of Electronic and Information Enginccring, Chonbuk Univcrsity, 664-14 Duckjin-Dong, Duckjin-Gu, Chonju, Chonbuk, 561 -756, S. Korca [email protected]
Abstract. Mobile information scrviccs havc to provide some dcgrcc of information adaptability onto the current service context, such as mobile user's location, network status. This paper deals with a predictive prefetching scheme for reducing the latcncy to gct rcfrcshcd information appropriatcd on currcnt location. It makes use of the velocity mobility model to exploit location knowledge about thc tcrminal andlor uscr's mobility bchavior. With considcring thc uscr's moving speed and direction, the prefetching zone has been proposed to effectively limit the prefetched information into the most likely future location context whilst to prcscrvc thc prcfctching bcncfits. Thc proposcd schcmc has bccn evaluated with a simulator in the context adaptability point of view.
1 Introduction With prospcrity of hardware tcchnology, computing devices arc gctting small and higher performance. Moreover, by cooperation of the wireless communication with hand-hold computer, computing users can access their own information anytime, anywhcrc, cvcn thcy movc around. Evcn if many cxisting computing scrviccs may bc logically cxtcndcd into this ncw coming computing paradigm, its inhcrcnt charactcristics, such as user, terminal andlor service mobility, require new service methodologies. That is, mobile information services have to provide some degree of information adaptability onto thc currcnt scrvicc contcxt, whcrc contcxt is dcfincd as thc charactcristics of thc uscr's situation likc as currcnt location, and thc uscr supporting cnvironment like as network status (bandwidth, broadcast etc.) and mobile terminal capabilities (size, color etc.). This feature is generally formalized as context-aware mobile information scrviccs [I]. In thc uscr and/or tcrminal mobility cnvironrncnt, location contcxt may bc changcd whilst the user is still interested in the corresponding answer. Therefore, it is required to bc adaptivcly providcd a rcsponsc appropriate to thc currcnt location: a prcviously providcd rcsponsc must bc rcfrcshcd, if thc corresponding request is still activc aftcr a This work was supported in part by University Research Program of Ministry of Information nd Communication in South Korca, and Ccntcr for Advanccd Imagc and Information of Chonbuk National Univcrsity in South Korca. P.M.A. Sloot et al. (Eds.): ICCS 2002, LNCS 2331, pp. 1128−1136, 2002. Springer-Verlag Berlin Heidelberg 2002
Using Predictive Prefetching to Improve Location Awareness 1129
change of location that invalidates the old response [2]. Thus, thc latcncy to get rcfreshed information appropriated on the current context is the most critical service parameter, especially in the frequently context changed environment. With the long latcncy, the uscr may suffcr thc long timc duration provided with information not appropriate for the currcnt location. To rcducc the latcncy, a wcll-cstablishcd tcchnique, prefetching, can be adapted to fetch information in advance, usually based on what contexts are highly utilized [3]. Thus, it is very important to make use of some dcgrcc of knowlcdgc to limit thc amount of prcfctchcd information; othcrwisc, unncccssarily prcfctchcd information can ovcrwhclm the benefit of latcncy rcduction with communication and memory cost. Therefore, prefetching scheme is usually based on some sort of prediction strategy about the information that will be needed in thc ncar futurc. Thc prcdictivc prcfctching concept has bccn applied and evaluated to rcducc thc latency in retrieving a Web document on the wired environment, based on the user's navigation pattcrn [4]. It shows that thc prcdictivc prcfctching rcduccs significantly in thc avcragc acccss timc. The work in [3] providcs the cffcctivcncss of prcfctching policies in the location-aware mobile information services. The work considered only the uniform mobility pattern based on the Markov model, and did not consider how to cffcctivcly confinc thc prcfctching information to rcducc communication cost. In the rcfcrcncc [S], a semantic based caching schcmc has bccn prcscntcd to acccss location dependent data. The work is just centric to the cache information replacement strategy to utilize the semantic locality in terms of location. In this papcr, a prcdictivc prcfctching schcmc is proposcd by cxploiting uscr movcmcnt knowlcdgc to limit thc prcfctching to the most likcly futurc contexts. Bccausc the location is the most distinguished context for the context aware mobile information system, our discussion is limited in location context after this. The rest of the papcr is organized as follows. In Scction 2, a gcncral architccturc of mobilc information scrvicc is prcscntcd, along with the problem definition. In Scction 3, a velocity based mobility model is presented and a predictive prefetching scheme is formalized with the mobility model. Using a simulation, a numerical result is shown in Section 4. Finally a conclusion is given in Section 5.
2 Mobile Information Service For mobile information service, a general architecture has been utilized to describe the location-aware feature [2], [3], [S], [6], [7]. The architecture consists of 3 logical componcnts; mobilc dcvicc that possibly movcs around with a uscr, information scrvcr that would bc located somcwhcrc in Intcrnct, and location-aware scrvicc manager that acts as intermediary between the mobile device and the information server. The manager performs a sequence of operations to retrieve information associated with the currcnt location from the scrvcr, and to load it on thc uscr dcvicc. To do this, it periodically monitors thc currcnt uscr location, and checks whcthcr it belongs to the current information scope. If an out-of-scope condition is detected, the service manager restarts to retrieve the corresponding information on current location. A mobile uscr is assumcd to cquippcd with a ccrtain mechanism to obtain its currcnt location,
1130 G. Cho
such as thc GPS or indoor infrarcd scnsors. For a given mobile information service, the set of possible location contexts is assumed to be divided into separated subsets, such as rooms, areas. Then a separate piccc of information is associatcd to cach subsct, such as cquipmcnt availablc in a room, map of a givcn arca. A unit arca, which a uscr may movc around with thc samc information piece, is divided into adjacent geographical portions. Each geographical portion has corresponding portion information that must be provided by the information scrvicc as rcsponsc to a uscr's rcqucst within that gcographical portion. Gcographical portions arc assumcd to bc cqual shapc rcprcscntcd as quadrangles. A discrete time model is used for a user's movement. At the end of each time slot, the user can remain in the same quadrangle, or move to an adjacent quadrangle through one of thc sharcd cdgcs. At a ccrtain point of timc, thc choicc of rcmaining insidc or moving outsidc thc quadranglc of a uscr is just dcpcndcd on thc currcnt location and thc moving direction and speed.
Information sewice time) Information acquisition delay) Move detection delay)
Fig. 1. Mobile information service delay on a given geographical portion
Fig. 1. shows thc dclay to gct corrcsponding information for a givcn gcographical portion (the quadrangle stands for a geographical portion). On crossing a geographical portion, thc latcncy can bc classified with two timc factors; thc dclay nccdcd to dctcct that thc providcd portion information is out of its currcnt contcxt, so to : movc detection delay, and the delay needed to retrieve and start loading the new information for current geographical portion, so t,: information acquisition delay. By reducing two dclays, a rnobilc uscr can bc blcsscd cnough with thc currcnt contcxt information, so t2: information scrvicc timc. Thc cffcct of thc first factor can bc improvcd by reducing the time interval between two consecutive location checks. The most common technique to improve the second factor is to prefetch information that will be rcquircd in thc ncar futurc. With thc practical rcquircmcnts such as communication and storagc limitation of rnobilc dcvicc, thc amount of prcfctchcd information must be minimized as much as possible, whilst the prefetching amount reduction must not hurt the prefetching benefits.
Using Predictive Prefetching to Improve Location Awareness 1131
3 Mobility Model and Predictive Prefetching At a givcn location, thc uscr's futurc location can bc rclativcly rcprcscntcd by thc moving direction and speed. Therefore, the moving behavior of mobile users can be now modeled from 3 parameters, current location, moving speed and direction. 3.1 Velocity based Mobility Model To model both the direction and speed of a moving user, a velocity Vis defined as a >, where I/, and 1: are integers. A pair of values, 1/, and 4, is the vector < I/,, projcctions of its spccd vcctor on x axis and y axis rcspcctivcly (in practically, cach axis can stand for the east or the north respectively). The signs of l/, and I/,, and the ratio of the absolute value of I/, and that of determine the directions of the movemcnts. Thc modcl assumcs that a uscr movcs around in a 2-dmcnstional spacc at a spccd that kccps constant during any obscrvcd time period and may changc aftcr a given unit time. So, at a certain point of time, when a user's current location has been given, the location in the future can be specified by considering the movement velocity. This conccpt has bccn mostly borrowcd from [ 5 ] .
4
4
3.2 Predictive Prefetching Strategy Generally, a natural prefetching strategy is to prefetch the portion information corresponding to a sct of gcographical portions that arc within a ccrtain dcgrcc of outbound distancc d from thc currcnt location. Thcsc gcographical portions consist of a prefetching zone (shortly PZ). To form a PZ, one possible strategy is to include all the quadrangles within outbound distance d fraction (so, radius d) of rings 0, 1, 2, .. .n [3]. No rcmotc loading is ncccssary unlcss thc uscr movcs around within its currcnt PZ. Whcn thc uscr cntcrs a portion outside its currcnt PZ, thc gcographical portion becomes the new starting position to build a new PZ. The new PZ may include the portion information for quadrangles within distance d ring. Hcrc, our conccrn is to cxploit somc dcgrcc of prcdiction about thc uscr bchavior, in ordcr to bc prcfctchcd only thc information for most likcly futurc contcxts. Whcn a user crosses a PZ, the moving speed is applied to estimate the distance d, and the moving direction is utilized to decide the width w of PZ. Now, a PZ shapes as a rectangle [7], rathcr than a ring. This is bascd on that, if a uscr movcs with high spccd, it would bc vcry rcasonablc to sct thc distancc higher than that of low spccd movcmcnt. Similarly, if a user's behavior tends to move with high degree of obliquity, it would bc vcry rcasonablc to sct thc width widcr than that of low dcgrcc of obliquity. Our stratcgy to prcdictivc prcfctching is as follows. Thc distancc d of a rcctanglc PZ is to get the upper bound integer of the user's speed, that is, the squared root of the
1.-
as Thus, the signs of sum of squares of two speed values Y, and 6, speed values should be ignored because the distance considers the speed characteristic only. In addition, if onc of spccd valucs is zcro, it must bc rcplaccd with thc othcr
1132 G. Cho
non-zero value. Similarly, the width w of a rectangle PZ can be obtained with the upper bound integer of the mean value of the sum i f two speed values, V. and , vxI+
Iv,I
that is . Note that the signs of speed values should be ignored because the direction reflects the degree of obliquity-rather than that of dimension. Differently with the distance, zero valued speed will be ignored in the width. Fig. 2. shows an example for the predictive prefetching reconstruction idea for three different moves. The dark rectangles depict the new PZ that is constructed on each crossing of the PZ. For each moving with the velocity 12, 1>, <3, 5> and <5,0>, the size of PZs is calculated as the distance d = 3,6 and 8, the width w = 2,4 and 3 respectively.
Fig. 2. A predictive prefetching reconstruction example
As shown in the Fig. 2., the gap between the PZs is inevitable. The location aware manager may start to build a new PZ and to prefetch the new PZ, just after it finds out the fact of crossing the current PZ, so the move detection delay in Fig. 1. Then, until the device is refreshed with information for the new PZ, so the information acquisition delay in Fig. 1., the mobile user therefore does not have any information appropriate to its current context. Due to the mobility feature, the user is in the out-of-scope condition of an effective information service. This situation can be improved with defining the edge geographical portions of a PZ into the crossing zone; it should be also defined with considering its moving speed and direction. When a user meets a geographical portion in the crossing zone, it starts to prefetch the new PZ in advance.
Using Predictive Prefetching to Improve Location Awareness 1133
4 Numerical Result In order to provide the cffcctivcncss of proposed scheme, a comparison has bccn made between the ring PZ approach and the rectangle PZ one, through a simulation study. For simulation runs, a set of system parameters has been borrowed from other works, as shown in Table 1. Some numerical result has bccn measured for the amount of prcfctchcd portion information, the ratio of utilization of the prcfctchcd portion information, and the service time for end user. The first is significant to reduce the communication cost, the second stands for how much the prefetched portion information is made use of rcal information scrvicc, and the last is meaningful to configure thc uscr's satisfaction for thc scrvicc. The simulator has bccn implcmcntcd in C language using an event-based simulator CSIM [8]. Table 1. System paramctcrs for simulation Parameter Wired bandwidth L4'ircles bandnidth Movc dctcction time
Description bandwidth of the wired line
Value 800 kbps - 1.2 Mbps
I bandwidth ol'thc \~~ircIcss medium 1 80 kbps - 120 kbps I
1 time
to detect the fact a user
I
1
100 ms
I
4.1 User Move Scenario
In the velocity based mobility model, a user moves around in a 2-dmenstional portion's geographical space at a spccd that kccps constant during any givcn unit timc pcriod. Thcn it changes its spccd and dircction, and moves constantly within the unit time. Therefore, the model is much significant when a user's moving pattern is uniform. Practically, this is analogous to many rcal life cases, such as driving a car on highways or walking a street. A typical uscr move sccnario has bccn utilized for the simulation. A user moves around 20 by 20 geographical portions, as shaped in Fig. 2. It starts from the position (10, lo), and moves 20 times, sequenced as (2, -4), (3,2), (3,4), (-5, O), (2, -51, (-3, -3), (5, - 4 , (0, 31, (4, - 4 , (-1, 3), (3, 3), (0, 51, (-2, 5), (-3, 01, (-5, -21, (5, -2), (-5, O), (-2,4), (-3, -61, (4, -5). 4.2 Simulation Experiments
With the givcn uscr mobility sccnario, the numbcr of prcfctchcd portion information has been evaluated to show the effectiveness of prefetching strategies by comparing three different cases; prefetching with ring PZ, prefetching with rectangle PZ, pre-
1134 G. Cho
fetching is not applied. The same simulation parameters are used for each case, whilst different strategies are used to build a PZ. For the prefetching with ring PZ, radius of degree of fairness the ring is a critical simulation parameter. In order to provide -some
dv:
Y: is adapted for with that of prefetching with rectangle PZ, the speed value calculating the radius of ring PZ. Fig. 3. shows the number of prefetched portion information for each cases. The prefetching with ring PZ has relatively large variation depending on the speed value. Thus, some of portion information would be overlapped between that of the previous PZ and that of the current PZ. After excluding these overlapped portions, 1,105 portions are totally prefetched for the ring PZ strategy, and mean number of prefetched portion information is 55.25 for each move. In the rectangle PZ scheme, the number of prefetched portion information is stable; total number of prefetched portion information is 335, mean number of prefetched portion information is 16.75. In the case of prefetching is not applied, each portion information has to be acquired for every geographical portions in turn, on the path of the user moves through. 1
1
1
1
1
1
1
1
+
1
prefetching with ring PZ prefetching with rectangle PZ ~refetchmgis not applied 1 1 1 I
1 2
3
4
1 1
1 1 1 l
l
5
6
1 1
1 1 1
1 1
1 1 1
~
~
~
1
7
8
9
10 11 12 13 14 15 16 17 18 19 20
move sequence
Fig. 3. The number of prefetched portion information
Now, it is meaningful to figure out how much the prefetched portion information would be actually participated for the user's location-aware service. Fig. 4. shows the utilization ratio of prefetched portion information for two prefetching strategies, so the ring PZ and the rectangle PZ. Note that the utilization ratio is just depended on the user's mobility pattern. With the given user move scenario, the utilization of rectangle PZ is much better than that of ring PZ; the former shows that over 30% of prefetched portion information is utilized for real service, whilst the later figures around 10%. We believe that the predictive prefetching with considering the user's move behavior plays great role in building a prefetching method for mobile information service.
Using Predictive Prefetching to Improve Location Awareness 1135
move sequence Fig. 4. Utilization ratio of the prefetched portion information
Fig. 5. shows the service time which has been spent to provide the corresponding information on each move. In the case of the prefetching is not applied, the time can be obtained by adding the service delay for getting each of portion information that the user has visited. So, mean service time is 570 sec. for each PZ. With the rectangle PZ strategy, the service delay is applied for only when a user crosses the current PZ. Once the service delay has been given on the portion, the other portion information within current PZ would be delivered in concurrently during the first portion information is serviced to end user. So the delay is the same for getting one of portion information on crossing PZ, and can be reduced as about 95 sec.
move sequence
Fig. 5. Information service Time
1136 G. Cho
5 Conclusion In thc location awarc mobile information scrvicc, prcfctching has a tradcoff bctwccn the amount of information transferred and the time delay to get the current context. In this paper we proposed a predictive prefetching scheme to limit the prefetched portion information to thc most likcly futurc contcxts, whilst to prcscrvc thc prcfctching bcncfits. This has bccn achicvcd by defining the prcfctching zonc with distancc and width based on the user's moving behavior, that is, speed and direction. Using a simulator, the effectiveness of the proposed prefetching strategy has been measured with rcspcct to thc scrvicc cffcctivcncss. Thc proposcd prcfctching with rcctanglc PZ was highly bcncficial for both the prcfctchcd data amount as well as the latency reduction. The idea in this paper is expected to be able to be extended to the other mobile server contexts, such as network bandwidth, mobile device type.
References 1. Brown, P. J., Bovey, J. D., Chen, X.: Context-aware Applications: From the Laboratory to the Marketplace. IEEE Personal Communications, Vol. 4, No. 5 (1997) 55-65 2. Couderc, P., Kermarrec, A. -M.: Improving Level of Service for Mobile Users Using Context-Awareness. Proc. 18th IEEE Symp. on Reliable Distributed Systems (1 999) 24-33 3. Persone, V. N., Grassi, V., Morlupi, A.: Modeling and Evaluation of Prefetching Polices for Context-Aware Information Services. Proc. ACM Conf on Mobile Computing and Networking (1998) 55-65 4. Padmanabhan, V. N., Mogul, J. C.: Using Predictive Prefetching to improve World Wide Web Latency.-ACM Computer Communication Review, Vol. 26, No. 3 (1 998) 22-36 5. Ren, Q., Dunham, M. H.: Using Semantic Caching to Manage Location Dependent Data in Mobile Computing. Proc. ACM Conf. on Mobile Computing and Networking (2000) 21022 1 6. Kovacs, E., Rohrle, K., Schiemann, B.: Adaptive Mobile Access to Context-aware Services. Proc. 1st Symp. on Agent Systems and Applications (1999) 190-201 7. Kim, M. J., Cho, G. H., Cho, I. J.: A Prefetching Scheme for Context-Aware Mobile Information Services, Proc. 3rd Conf. on Advanced Communication Technology (2001) 123-125 8. Mesquite Software Inc.: CSTMI 8 Simulation Engine (1997)
Dynamic and Stochastic Properties of Molecular Systems: From Simple Liquids to Enzymes Igor V. Morozov1,2 , Guenri E. Norman2,3 , and Vladimir V. Stegailov3,2 1 Department of Physics, Chair of General Physics and Wave Processes, Lomonosov Moscow State University, Vorob’evy Gory, 119899, Moscow, Russia, [email protected] 2 Institute for High Energy Densities, Associated Institute for High Temperatures, Russian Academy of Sciences (IHED-IVTAN), Izhorskaya 13/19, Moscow 127412, Russia, henry [email protected] 3 Moscow Institute of Physics and Technology, Institutskii per. 9, 141700, Dolgoprudnyi, Russia [email protected]
Abstract. Molecular dynamics method (MDM) supplies to the solution of fundamental contradiction between macroscopic irreversibility and microscopic reversibility with data which help to reveal the origin of stochastization in many-particle systems. The relation between dynamic memory time tm , fluctuation of energy dE and K-entropy (Lyapunov exponent) is treated. MDM is a method which retains Newtonian dynamics only at the times less than tm and carries out a statistical averaging over initial conditions along the trajectory run. Meaning of tm for real systems is related to the quantum uncertainty, which is always finite for any classical system and influence upon particle trajectories in a coarsegraining manner. Relaxation of kinetic energy to equilibrium state was studied by MDM for non-equilibrium strongly coupled plasmas. Two stages of relaxation were observed: initial fast non-Boltzmann oscillatory stage and further relatively slow Boltzmann relaxation. Violation of the microscopic reversibility principle in some enzymatic reactions is discussed.
Monte Carlo method is referred to stochastic methods of molecular simulation whereas molecular dynamics method (MDM) is usually called a dynamic method. The objective of the present paper is to show that MDM possesses both dynamic and stochastic features. Moreover if MDM had no hidden stochastic features MDM would not probably be able to achieve well-known successive results. Another objective is to present examples of MDM simulation of non-equilibrium relaxation when stochastic features influence the dynamics. The topic is related to the occurrence of the irreversibility in the case of the classical molecular systems which has been discussing since Boltzmann-Zermelo debate. The present state was given at the round table ”Microscopic origin of macroscopic irreversibility” at the XX International conference on statistical physics (Paris, 1998) [1, 2], see also [3]. P.M.A. Sloot et al. (Eds.): ICCS 2002, LNCS 2331, pp. 1137−1146, 2002. Springer-Verlag Berlin Heidelberg 2002
1138 I.V. Morozov, G.E. Norman, and V.V. Stegailov
1
Molecular Dynamics Method
The idea of MDM is very simple: all possible classical systems and media are simulated by a set of N moving atoms and/or molecules, which interact with each other (e.g., see [4–11]). The numerical integration of the corresponding system of Newton equations mi
dvi (t) = Fi [r(t)], dt
dri (t) = vi (t) dt
(1)
results in determination of the trajectories of all particles {r, v}. Here, mi , vi , ri , and Fi are the mass, velocity, and coordinate of the ith particle and the force acting on this particle, respectively (i = 1, . . . , N ); the vi and ri values explicitly depend only on time t; Fi depends only on the coordinates of particles; r(t) is the set of the coordinates of all particles, r(t) = {r1 (t), r2 (t), . . . , rN (t)}; v(t) is defined similarly; and Fi = −
∂ U (r1 , r2 , . . . , rN ), ∂ri (t)
(2)
where U is the potential energy. Function U (forces F) is assumed to be given in MDM. The total energy E of the system is the sum of the kinetic T and potential U energies, E = T + U,
T =
N % mv2 i=1
2
i
(3)
Set (1) is exponentially unstable for a system of more than two particles (e.g., see [3–15]). The parameter that determines the degree of instability, that is, the rate of divergence of initially close phase trajectories, is Lyapunov exponent or K-entropy K. It can be determined by several ways. As example for a given U function and particles of the same mass m and for identical initial conditions corresponding to the kth point on an equilibrium molecular-dynamical trajectory, solutions {r(t), v(t)} to system (1) are found in steps ∆t and trajectories {r" (t), v" (t)} are calculated in steps ∆t" . Averaged differences of the coordinates (velocities) of the first and second trajectories are determined at coinciding time moments, N,I ! 2 " $2 1 %# " ∆v (t) = vjk (t) − vjk (t) , NI
(4)
N,I ! 2 " $2 1 %# " ∆r (t) = rjk (t) − rjk (t) , NI
(5)
j,k
j,k
To improve accuracy, averaging over k = 1, . . . , I is also performed. In some transient time tl the differences become exponentially increasing with the same
Dynamic and Stochastic Properties of Molecular Systems 1139
value of K,
! 2 " ∆v (t) = A exp(Kt), ! 2 " ∆r (t) = B exp(Kt) at tl < t < tm .
(6)
The values of A and B are determined by the difference of ∆t and ∆t" . An example for the three-dimensional Lennard-Jones system is given in Fig. 1. We use reduced units in which m = # = σ = 1, and time is measured in (mσ2 /#)1/2 units. 10 1
<∆v2> 2vT2
10-1 -2
10
10-3
t'm <∆r2> L2
tl
10-4 10-5 10-6 10-7
t
10-8 0
1
2
3
4
5
Fig. 1. Normalized averaged differences of velocities ∆v2 and coordinates ∆r2 at coinciding time moments along two trajectories calculated for identical initial conditions in steps ∆t = 0.001 and ∆t" = 0.0001; L is the calculation cell edge length, N = 64, r = 0.5, and T = 0.44.
! " The exponential increase of ∆v2 (t) is limited by the finite value of the thermal velocity of particles. Thus after the time t"m t"m ≈ K −1 ln (6kT /Am) ,
(7)
where T is the temperature, saturation is reached. t"m is a dynamic memory time [14–17]. For t > t"m ! 2 " ! " ∆v (t) = 2 v 2 = 6kT /m, (8) ! 2 " ! 2 " " " ∆r (t) = 6D(t − tm ) + ∆r (tm ) , (9) where 3kT /m is the square!of thermal √ the diffusion coeffi" velocity vT and D is cient. Estimates show that ∆r2 (tm ) = rav , where rav = ( 2nσ)−1 is the mean path of particles between collisions.
1140 I.V. Morozov, G.E. Norman, and V.V. Stegailov
The dynamic memory times are determined by calculating at the same ∆t value and different ∆t" values of ∆t/2, ∆t/5, ∆t/10, etc. The limiting value when ∆t" /∆t → 0 is the dynamic memory time tm for a given system and the selected numerical integration step ∆t [14, 15]. During numerical integration, the system completely “forgets” its initial conditions in time tm , and the calculated molecular-dynamical trajectory completely ceases to correlate with the initial hypothetical Newtonian trajectory. In other words it determines the time interval during which the behavior of a dynamic system can be predicted from initial conditions and deterministic equations of motion [16, 18]. The calculated dependencies of Ktm on the numerical integration step ∆t can be presented in the form Ktm = −n ln(∆t) + const,
(10)
where n is determined by the order of accuracy of the numerical integration scheme, or, in another form, K(tm1 − tm2 ) = n ln(∆t2 /∆t1 ),
(11)
where tm1 and tm2 are the dynamic memory times for the steps ∆t1 and ∆t2 , respectively. This result does not depend on either temperature, density, or the special features of the system under study [14, 15]. Because of the approximate character of numerical integration, energy E [Eq. (3)] is constant only in average. The E value fluctuates about the average value from step to step, and the trajectory obtained in molecular dynamics calculations does not lie on the E = const surface, in contrast to exact solutions to Newton Eqs. (1) . This trajectory is situated in! some"layer of thickness ∆E > 0 near the E = const surface [7, 8]. The value ∆E 2 ∼ ∆tn depends on the accuracy and the scheme of numerical integration [7, 8, 22–25]. Therefore #! "$ (12) Ktm = − ln ∆E 2 + const Equation (12) relates the K-entropy and the dynamic memory time to the noise level in the dynamic system. This equation corresponds to the concepts developed in [16, 18]. It follows from (10)-(12) that tm grows no faster than logarithmically as the accuracy of numerical integration increases. The available computation facilities allow ∆E to be decreased by 5 orders of magnitude even with the use of refined numerical schemes [23–25]. This would only increase tm two times. Estimates of dynamic memory times showed that, in real systems, tm lies in the picosecond range. It means that tm is much less than MDM run. So MDM is a method which retains Newtonian dynamics only at the times less than tm and carries out a statistical averaging over initial conditions along the trajectory run. The K-values were calculated by MDM for systems of neutral particles [3, 7–11, ?,20], two-component [15] and one-component [17] plasmas and primitive polymer model [21]. The values of K turn out to be the same for both velocities and coordinates deviations. It is also seen that the K values for electrons and
Dynamic and Stochastic Properties of Molecular Systems 1141
ions are close to "each other at the initial stage of divergence. At t = tme the ! quantity ∆v2 (t) for electrons reaches its saturation value and, therefore, at t > tme only ion trajectories continue to diverge exponentially with another value of K-entropy depending on M/m as Ki ∼ (M/m)−1/2 . The ratio tme /tmi is a fixed value. The dependence of tmi on the electron-ion mass ratio also fits the square root law. System of 10 polymer molecules with atom-atom interaction potential and periodic boundary conditions was studied in [21]. Each model molecule consisted of 6 atoms with constant interatomic distances and variable angles φ between links. Divergence of velocities ∆v2 (t) and coordinates ∆r2 (t) for both atoms and molecule center-of-masses as well as angles ∆φ2 (t) were calculated. All the five dependencies follow the exponential law before saturation. All the five exponents turned out to be equal to each other, as for electrons and ions in plasmas. One can expect that it is a general conclusion for systems with different degrees of freedom.
2
Stochastisity of dynamic systems
Kravtsov et al. [16, 18] considered the measuring noise, fluctuation forces and uncertainty in knowledge of differential deterministic equations of the system as the reasons why tm has a finite value. It is a characteristic of a simulation model in [7, 14, 15, 17]. The time tm was related to the concept of quasi-classical trajectory, which takes into account small but finite quantum effects in classical systems: broadening of particle wave packets and diffraction effects at scattering [14, 15, 26, 27], to weak inelastic processes [18]. Quasi-classical trajectories themselves are irreversible. The intrinsic irreversibility in quantum mechanics originates from the measurement procedure which is probabilistic by definition. Our premise coincide with the Karl Popper’s conviction foundation stone that “nontrivial probabilistic conclusions can only be derived (and thus explained) with the help of probabilistic premises” [28]. The probabilistic premise we use is the quantum nature of any motion which is used to be considered as a deterministic classical one. The idea was inspired by an old remark of John von Neumann [29] and Landau [30] that any irreversibility might be related to the probabilistic character of measurement procedure in quantum mechanics. Estimates of dynamic memory times were obtained for molecular dynamics numerical schemes. Since tm values very weakly (logarithmically) depend on the noise level, it allowed us to extend qualitative conclusions to real systems of atoms, in which the finiteness of the dynamic memory time is caused by quantum uncertainty. Though the primary source of the stochastic noise is probabilistic character of measurement procedure there are other factors which remarkably increase the noise value and permit to forget about quantum uncertainty at simulation. For example it is water molecule background that creates the stochastic noise in electrolytes. One is able to add Langevin forces into (1) and apply MDM to study their influence on dynamic properties of Coulomb system [31]. The dependence
1142 I.V. Morozov, G.E. Norman, and V.V. Stegailov
of the dynamic memory time on the value of Langevin force is presented in Fig. 2. It is seen that collisions of ions with water molecules does not change essentially the value of tm . 1.6
t/
e
1.2
0.8
0.4
0 0.0001
0.001
0.01
0.1
1
Fig. 2. Dynamic memory time for different values of Langevin force. The dashed line corresponds the level of the Langevin force which acts on the ions in the water solution.
3
Boltzmann and Non-Boltzmann Relaxation
Boltzmann equation is a fundamental equation of kinetic theory of gases; there are numerous attempts to modify the equation and extend it to dense media as well [32, 33]. Kinetic theory of gases deals only with probabilities of collision results. It is an initial assumption at the derivation of Boltzmann equation. Another fundamental assumption is Stosszahlansatz, which means that molecules are statistically independent. Molecular chaos hypothesis a base of the kinetic theory, i.e. it is implied that molecule motion is stochastized. However it is apparent that dynamic motion precedes to stochastic processes [13, 32, 33]. It is supposed that dynamic motion governed by intermolecular interactions defines the values of collision cross-sections but does not influence time dependence of kinetic processes. One can expect that Boltzmann description of kinetic processes is valid only for the times greater than tm . MDM can be a powerful tool for studying non-Boltzmann relaxation phenomena in more or less dense media. Some non-equilibrium processes have been already studied with MDM for example in [14, 19, 34–38]. MDM was applied in [39] to the study of electron and ion kinetic energy relaxation in strongly coupled plasmas. Two-component fully ionized system of 2N single-charged particles with masses m (electrons) and M (ions) was
Dynamic and Stochastic Properties of Molecular Systems 1143
considered . It is assumed that the particles of the same charge interact via the Coulomb potential, whereas the interaction between particles with different charges was described by the effective pair potential (“pseudo-potential”). The nonideality was characterized by a parameter γ = e2 n−1/3 /kT , where n = ne +ni is the total density of charged particles. The values of γ were taken in the interval from 0.2 to 3. The details of the plasma model and numerical integration scheme are presented in [15]. The following procedure was used to prepare the ensemble of nonequilibrium plasma states. Equilibrium trajectory was generated by MD for a given value of γ. Then a set of I = 50 − 200 statistically independent configurations were taken from this run with the velocities of electrons and ions dropped to zero. Thus the ensemble of initial states of nonequilibrium plasma was obtained. MD simulations were carried out for every of these initial states and the results were averaged over the ensemble. N,I & 2 An example of relaxation of average kinetic energy T (t) = N1I vjk (t) is j,k
presented in Fig. 3a for the total relaxation period . The values of T for both electrons and ions are normalized by the final equilibrium value. The nonideality parameter γf in final equilibrium state differs from initial γ value. The time is measured in periods of plasma oscillations τe = 2π/Ωe . Fig. 3a reveals Boltzmann character for time t > 5τe . It is evident from the insertion where the difference between the electron and ion kinetic energies is presented in the semi-logarithmic scale. The character of this long time relaxation agrees with earlier results [34, 36].
1.4
a)
T/Teq
b)
1.4
1
1.2
1.2
tm
1
1
1
3
0.8
(Te-Ti)/Teq
2
0.1
0.8
2
0.6
0.6
1
10-2 10-4
0.4
0.4
t/
0.01
10
0.2 0
40
20
80
30
t/
e 40
e 120
10-6
0.2
0.01
0 0
0.4
2
0.1
0.8
t/
e 1.2
Fig. 3. The kinetic energy for electrons (1), ions (2) and average value (3): a) at Boltzmann relaxation stage; b) for times less than the dynamic memory time. The equilibrium dynamic memory time is given by a vertical arrow. γ = 1, γf = 3.3, M/m = 100, 2N = 200.
1144 I.V. Morozov, G.E. Norman, and V.V. Stegailov
At the times less than 0.1τe both electron and ion kinetic energies increase according to the quadratic fit (Fig. 3b). Then the electron kinetic energy passes through the maximum and undergoes several oscillations damping at t ≈ tm while the ion kinetic energy increases monotonously. The possibility of two stages of relaxation was noted in [36]. The relative importance of non-Boltzmann relaxation stage for different nonideality parameters can be derived from Fig. 3. It is seen (as an example from the velocity autocorrelation function decay time) that it decreases with decrease of plasma nonideality. The calculation shows as well that the oscillatory character of non-Boltzmann relaxation vanishes when the nonideality parameter becomes less than γ = 0.5.
3
t/
e 2
1
0 0
1
2
3
Fig. 4. γ-dependencies of various electron relaxation times. Crosses are dynamic memory time tm , asterisks — inverse Lyapunov exponent, open (filled) circles correspond the values of 0.1 (e−1 ) of normalized electron velocity autocorrelation function, triangles (reverse triangles) are the inverse frequencies corresponding to the minimum (maximum) of dynamic structure facture. Curve and strait line are drawn in order to guide the eye. M/m = 100, 2N = 200.
Another example of a particular relaxation process is related to arising of irreversibility in the case of enzyme catalysis. Microscopic reversibility principle is a fundamental principle of physical and chemical kinetics. In particular it means that all intermediate states coincide for forward and backward chemical reactions. However, Vinogradov [40] obtained experimental evidence for different pathways for direct and reverse enzymatic reactions in the case of hydrolysis and synthesis of ATP and some other mitochondrial molecular machines and supposed that ”the one-way traffic” principle is realized at the level of a single enzyme.
Dynamic and Stochastic Properties of Molecular Systems 1145
Since microscopic reversibility principle follows from the time reversibility of fundamental dynamic equations, the occurrence of the irreversibility in the case of enzyme catalysis might be similar to the case of the classical molecular systems. The latter problem has been existing as intermediate between physics and philosophy. Hypothesis [40] switches it to experiment and applied science. If there are two pathways along the hypersuface of potential energy between initial and final states there should be at least two bifurcation points. It is not Maxwell demon but Lyapunov instability, stochastic terms and asymmetry of complicated potential relief with developed system of relatively hard valence bonds that define the local choice of reaction pathway in the bifurcation point [41]. The physical sense of stochastic terms is related to thermal fluctuations of the relief and noise produced by collisions with water molecules while the main features of the relief do not depend on time essentially. Molecular simulation example [42, 43] for a primitive model confirms this conclusion. The local choice is determined by the local parameters. The situation is equivalent to the statement that there is no thermodynamic equilibrium in the area around the bifurcation point and the theory of transient state is not valid here. The work is supported by RFBR (grants 00-02-16310a, 01-02-06382mas, 0102-06384mas).
References 1. Prigogine, I.: Physica A 263 (1999) 528–539 2. Lebowitz, J.L.: Physica A 263 (1999) 516–527 3. Hoover, W.G.: Time Reversibility, Computer Simulation and Chaos. World Scientific, Singapore (1999) 4. Ciccotti, G., Hoover W.G. (eds.): Molecular-Dynamics Simulation of Statistical Mechanical Systems. Proc. Int. School of Physics ”Enrico Fermi”, Course 97. North-Holland, Amsterdam, (1986) 5. Allen, M.P., Tildesley D.J.: Computer Simulation of Liquids. Clarendon, Oxford (1987) 6. van Gunsteren W.F.: In: Truhler, D. (ed.): Mathematical Frontiers in Computational Chemical Physics. Springer, New York (1988), 136–151. 7. Valuev A.A., Norman G.E., Podlipchuk V.Yu.: In: Samarskii, A.A., Kalitkin N.N. (eds.): Mathematical Modelling. Nauka, Moscow, (1989) 5–40 (in Russian) 8. Norman, G.E., Podlipchuk, V.Yu., Valuev, A.A.: J. Moscow Phys. Soc. Institute of Physics Publishing, UK 2 (1992) 7–21 9. Hoover, W.G.: Computational Statistical Mechanics. Elsevier, Amsterdam (1991) 10. Rapaport, D.C.: The Art of Molecular Dynamics Simulations, Parag. 3.8, 5.5.1. Cambridge University Press, Cambridge (1995) 11. Frenkel, D., Smith, B.: Understanding Molecular Simulations, Parag. 4.3.4. Akademic Press, London (1996) 12. Stoddard, S.D., Ford, J.: Phys. Rev. A 8 (1973) 1504–1513 13. Zaslavsky, G.M.: Stochastisity of dynamic systems. Nauka, Moscow (1984); Harwood, Chur (1985) 14. Norman, G.E., Stegailov, V.V.: Zh. Eksp. Theor. Phys. 119, (2001) 1011–1020; J. of Experim. and Theor. Physics 92 (2001) 879–886
1146 I.V. Morozov, G.E. Norman, and V.V. Stegailov 15. Morozov, I.V., Norman, G.E., Valuev, A.A.: Phys. Rev. E 63 036405 (2001) 1–9 16. Kravtsov, Yu.A.: In: Kravtsov, Yu.A. (ed.): Limits of Predictability. Springer, Berlin (1993) 173–204 17. Ueshima, Y., Nishihara, K., Barnett, D.M., Tajima, T., Furukawa, H.: Phys. Rev. E 55 (1997) 3439-3449 18. Gertsenshtein, M.E., Kravtsov, Yu.A.: Zh. Eksp. Theor. Phys. 118 (2000) 761–763; J. of Experim. and Theor. Physics 91 (2000) 658–660 19. Hoover, W.G., Posch, H.A.: Phys. Rev. A 38 (1998) 473–480 20. Kwon, K.-H., Park, B.-Y.: J. Chem. Phys. 107 (1997) 5171-5179 21. Norman, G.E., Yaroshchuk, A.I.: (to appear) 22. Norman, G.E., Podlipchuk, V.Yu., Valuev, A.A.: Mol. Simul. 9 (1993) 417–424 23. Rowlands, G.J.: Computational Physics 97 (1991) 235–239 24. Lopez-Marcos, M.A., Sanz-Serna, J.M., Diaz, J.C.: J. Comput. Appl. Math. 67 (1996) 173–179 25. Lopez-Marcos, M.A., Sanz-Serna, J.M., Skeel, R.D.: SIAM J. Sci. Comput. 18 (1997) 223–230 26. Kaklyugin, A.S., Norman, G.E.: Zh. Ross. Khem. Ob-va im. D.I. Mendeleeva 44(3) (2000) 7–20 (in Russian) 27. Kaklyugin, A.S., Norman, G.E.: J. Moscow Phys. Soc. Allerton Press, USA 5 (1995) 167-180 28. Popper, K.: Unended Quest. An Intellectual Autobiography. Fontana/Collins, Glasgow (1978) 29. von Neumann, J.: Z. Phys. 57 (1929) 30–37 30. Landau, L.D., Lifshitz, E.M.: Course of Theoretical Physics, Vol. 5, Statistical Physics, Part 1. Nauka, Moscow (1995); Pergamon, Oxford (1980); Quantum Mechanics: Non-Relativistic Theory, Parag. 8, Vol. 3. 4th ed. Nauka, Moscow, (1989); 3rd ed. Pergamon, New York, (1977) 31. Ebeling, W., Morozov, I.V., Norman, G.E.: (to appear) 32. Balescu, R.: Equilibrium and Nonequilibrium Statistical Mechanics. London Wiley, New York (1975). 33. Zubarev, D.N., Morozov, V.G., Roepke, G.: Statistical Mechanics of Nonequilibrium Processes. Akademie-Verlag, Berlin (1996) 34. Hansen, J.P., McDonald, I.R.: Phys. Lett. 97A (1983) 42–45 35. Norman, G.E., Valuev, A.A.: In: Kalman, G., Rommel, M., Blagoev, K. (eds.): Strongly Coupled Coulomb Systems. Plenum Press, New York (1998) 103–116 36. Norman, G.E., Valuev, A.A., Valuev, I.A.: J. de Physique (France) 10(Pr5) (2000) 255–258 37. Hoover, W.G., Kum, O., Posch, H.A.: Phys. Rev. E 53 (1996) 2123–2132 38. Dellago, C., Hoover, W.G.: Phys. Rev. E 62 (2000) 6275–6281 39. Morozov, I.V., Norman, G.E.: (to appear) 40. Vinogradov, A.D.: J. Exper. Biology 203 (2000) 41–49; Biochim. Biophys. Acta 1364 (1998) 169–185 41. Kaklyugin, A.S., Norman, G.E.: Zhurnal Ross. Khem. Ob-va im D.I. Mendeleeva (Mendeleev Chemistry Journal) 45(1) (2001) 3–8 (in Russian) 42. Norman, G.E., Stegailov, V.V.: ibid. 45(1) (2001) 9–11 43. Norman, G.E., Stegailov, V.V.: Comp. Phys. Comm. (to appear)
Determinism and Chaos in Decay of Metastable States Vladimir V. Stegailov Moscow Institute of Physics and Technology, Institutskii per. 9, 141700, Dolgoprudnyi, Russia [email protected]
Abstract. The problem of numerical investigation of metastable states decay is described in this work on example of melting of the superheated solid crystal simulated within the framework of molecular dynamics method. Its application in the case of non-equilibrium processes has certain difficulties, including the averaging procedure. In this work an original technique of averaging over the ensemble of configuration is presented. The question of the instability of the phase space trajectories of many-particle system (i.e. chaotic character of motion) and its consequences for simulation are also discussed.
1
Introduction
Melting is still not completely understood phenomenon. There exists the question of estimation of the maximum possible degree of crystal superheating. This problem attracts certain interest in connection with the experiments dealing with intensive ultrafast energy contributions where specific conditions for superheating are realized [I]. Superheated crystal is metastable, therefore it melts in some finite time. It is obvious that the decay of ordered phase happens sooner at higher degrees of superheating. The utmost superheating therefore should be characterized by the values of life time comparable to the period of crystal node oscillations. At this time scale it is possible to apply molecular dynamics (MD) method of numerical simulation to investigate melting at high degrees of superheating on the microscopic level [2,3]. In this work I dwell on the peculiarities of application of the MD method in study of this non-equilibrium process on example of direct calculation of nucleation rate and melting front propagation velocity.
2
Model and Calculation Technique
The model under consideration is the fcc-lattice of particles interacting via homogeneous potential U = E ( f ) m . Calculations were made for rn= 12. To model an initially surface-free and defect-free ideal crystal periodic boundary conditions were used. For numerical integration Euler-Stormer 2-nd order scheme was P.M.A. Sloot et al. (Eds.): ICCS 2002, LNCS 2331, pp. 1147−1153, 2002. Springer-Verlag Berlin Heidelberg 2002
1148 V.V. Stegailov
applied. Number of particles in the main cell was N = 108; 256; 500. Homogeneous potential allows to describe thermodynamic state of the system in terms of only one parameter X = ( N O ~ / ~ ~ V ) ( ~ B where T / E )V~is/ the ~ , main cell volume and T is the temperature. MD-system was transferred from a stable ordered equilibrium state t o a metastable superheated one by means of either isochoric heating or isothermic stretching or their combination. A degree of superheating can be characterized p ~ 3 / n: the larger is X - l the higher is the superby the quantity of X - l heating. When the cryst a1 was in a sufficiently metastable state the calculation of dynamics of isolated system was performed. Sharp fall of kinetic energy of the system Ekin manifests transition t o disordered state (Fig.1). From Ekin( t ) dependence the value of metastable configuration lifetime tlipe can be derived. However the values of tlipe are very different for the MD-runs calculated from N
Fig. 1. Time dependence of mean kinetic energy of particles. A phase transition from ordered to disordered state is shown. Subfigure: t l i f e values from MD-runs with identical initial conditions but different integration time steps At
one and the same initial configuration with various time-steps At. Besides there is no convergence when At + 0 (Fig.1). This fact is rather confusing because it is not clear what value of tlipe one should attribute t o the dynamical system under consider at ion.
3
Role of Instability
The MD method based on the numerical integration of the corresponding system of Newton equations
that results in determination of the trajectories of all particles { r( t ),v ( t ) ) ,where r = ( r l ,r2,. . . ,r N )and v = ( v l ,~ 2 , . .. ,V N ) .
Determinism and Chaos in Decay of Metastable States
1149
Set (1) is exponentially unstable for a system of more than two particles (e.g., see [4-171). The parameter that determines the degree of instability, that is, the rate of divergence of initially close phase trajectories, is averaged Lyapunov exponent or K-entropy K . Let us consider solutions t o system (1) for identical initial conditions corresponding to some point on an MD-trajectory: {r(t),v(t)) (found in step At) and {r'(t), v' (t)) (in step At'). Averaged differences of the coordinates and velocities of the first and second trajectories are determined at coinciding time moments (Fig.2, a). In some transient time the differences become exponentially increasing (Fig.2, b). The values of A and B are determined by the difference of integration steps At and At'. The exponential increase of (Au2(t)) is limited by the finite value of the thermal velocity of particles uT; (Ar2 (t)) is limited by the mean amplitude of atom oscillations in crystal. Since exponential growth of
Fig. 2. Divergences of velocities (squares) and coordinates (triangles) at coinciding moments of time along two trajectories calculated from identical initial conditions with time-steps At = 0.001 and tk = 0.00101. Time, length and velocity are given in reduced units where E = a = 1
(Ar2(t)) and (Au2(t)) two initially coinciding MD-trajectories lose any relation very quickly. Let t h denote time when divergences come t o saturation (Fig.2) and therefore these trajectories become completely uncorrelated. Value of t h is a function of At and At'. One should mention that for the analyzed on Fig.1 configuration t h << tlipe, hence MD-trajectories calculated with various At becomes uncorrelated much sooner than the transition t o disordered phase happens. This fact explains chaotic dependence of tlipe on At. However it is still vague how to determine an exact value of lifetime that is intrinsic property of the dynamical many-particle system. In [14-161 the notion of dynamical memory time t, is used. By its definition t h ( A t t , At) + t,(At) when At' + 0 and At is fixed. It is known from [14-161 that the dependence of t h on At' at fixed At is rather weak so it can be easily estimated as t, 21 t h .
1150 V.V. Stegailov
The physical sense of time t, consists in the following. During numerical integration after the time t, the MD-trajectory calculated with time step At completely "forgets" its initial conditions. It means that the MD-trajectory ceases to correlate with the hypothetical Newtonian trajectory (an exact solution of the set (??)). In other words the value of t, determines the time interval during which the behavior of molecular-dynamical system can be predicted from initial conditions and deterministic equations of motion at a certain level of accuracy defined by At value and a particular scheme of numerical integration. Such a definition of t, correlates with that given in [18,19]. It follows from [14-161 that t, grows no faster than logarithmically as the accuracy of numerical integration increases; in the same works it is shown that the available computation facilities allow only t o increase t, only two times even with the use of refined numerical schemes [20-221. It means that t, is much less than usual MD-run. So MD method retains Newtonian dynamics only at the short time intervals less than t,. Therefore an exact value of lifetime tlipe of the given metastable configuration can be obtained if tlipe < t,. This condition allows to explore only such superheating st at es that actually are not st ill metastable but practically unstable. By means of present computational capabilities one can calculate much longer time periods that is of more interest for physics. However we should accept that in this case it is impossible to calculate an exact dynamical trajectory of nonequilibrium process and to derive an exact value of tlipe for a given metastable configuration. To obtain results that can be interpret as a real property of dynamical system and not as a numerical artifact the results should be somehow averaged.
4
Averaging Procedure
The simplest way t o do it is to check if the tlipe values have some distribution or they do not. For example, we can calculate n = 100 MD-runs with identical initial configuration and At = 0.01,0.0099,0.0098,. . .,0.001, and then to count in how many realizations tlife E [t, t St]. The performed calculations showed than such distributions actually exist (Fig.3) and their shape weakly changes starting with n 100. Since tlipe >> t h these distributions of lifetime over a set of At is equivalent t o the distributions over different initial configurations corresponding to the same degree of superheating, i.e. thermodynamically indistinguishable. That is an exact situation in real experiments with metastable state decay: either crystallization of supercooled liquid or melting of superheated crystal [23]. Melting is considered as a process of nucleation. Formation of the new phase nuclei can be described in terms of a stochastic Poisson process [23]. Let X denote a probability of liquid phase nucleus formation in short time interval [t, t +St]. Then it can be shown that P ( t ) = exp(-At) is the probability that no nucleus will appear to the t moment. Let us consider no realizations (MD-runs) of metastable state decay. Let n denote the number of those non-decayed to the
+
>
Determinism and Chaos in Decay of Metastable States
Fig. 3. Examples of t l i f edistributions for N = 500: a - 1 / X = 1.3088, b 1.2879; n is the number of MD-runs where t l i f eE [t,t St].
+
-
1151
1/X =
t moment, then n(t) = no exp(-At). This dependence is perfectly confirmed by numerical results (see Fig.4). From the such distributions one can obtain the most probable lifetime for a superheated crystal at the certain degree of superheating tkfe = A-l. Unlike trife the value of tkfe is a physical property of the concerned dynamical model. Superheated state of defect-free crystal is characterized by the homogeneous nucleation rate J , that is average number of nuclei formed in the unit volume in the unit time interval. In our MD-model we can estimate J as (tkfev)-' where V is the main cell volume. From the physical point of view it is interesting to
- - - - - Exponential decay fit
Fig. 4. Distributions of number of MD-runs non-decayed to the t moment
now how depend nucleation rate J on the degree of superheating (the latter is characterized by parameter 1 / X p ~ 3 / n ) Following . the calculation procedure described in this section for a set of initial configurations corresponding t o various values of 1 / X we derived J ( l / X ) dependence presented at Fig.5. Although data points at Fig.5 relate to different N they all form one well determined dependence. It is one more confirmation that this data is not an numerical artifact but a property of the investigated model. In conclusion I'd like t o remark that by means of the described procedure distributions (similar t o Fig.3) of tmelting,i.e. time of the transition on Fig.1, were calculated for different 1 / X values. Melting front propagation velocity u
1152 V.V. Stegailov
can be estimated by L/tLelting, where tLelting corresponds to the maximum of these distributions and L is the main cell edge length. Numerical results are shown on Fig.5. In the investigating limit of extreme superheatings u approaches the value of the sound velocity.
5
Summary
Molecular dynamics met hod contrary t o Mont e-Carlo met hod is a technique of numerical simulation that deals exactly with dynamics of many-particle system. At the same time detailed analysis shows that actually calculation of real dynamics restricted to the quite short time intervals. In contrast to equilibrium phenomena, non-equilibrium processes are much more affected by the instability that is intrinsic to many-particle systems. At longer periods one should take into account specific statistical averaging over an ensemble of thermodynamically indistinguishable configurations. In this work it was shown how in spite of such problems one can obtain consistent physical results by means of appropriate interpretation of numerical data.
Fig. 5. Dependence of nucleation rate J and melting front propagation velocity v on the parameter 1/X, i.e. on the degree of superheating
6
Acknowledgements
I am very grateful t o my scientific supervisor prof. G.E.Norman for his careful guidance and essential remarks. I also appreciate interesting discussions with M.N.Krivoguz and his useful comments. Special thanks to A. A.Shevtsov who provided me with necessary computational resources. The work is supported by Russian Foundation for Basic Research (grants 00-02-16310 and 01-02-06384).
Determinism and Chaos in Decay of Metastable States
1153
References 1. Bonnes, D. A., Brown, J. M.: Bulk Superheating of Solid KBr and CsBr with Shock Waves. Phys. Rev. Lett. 71 (1993) 2931-2934 2. Jin, Z. H., Gumbsch, P., Lu, K., Ma, E.: Melting Mechanisms at the Limit of Superheating. Phys. Rev. Lett. 87 (2001) 055703-1-4 3. Krivoguz, M. N., Norman, G. E.: Spinodal of Superheated Solid Metal. Doklady Physics 46 (2001) 463-466 4. Hoover, W. G.: Time Reversibility, Computer Simulation and Chaos. World Scientific, Singapore (1999) 5. Ciccotti, G., Hoover, W. G. (eds.) : Molecular-Dynamics Simulation of Statistical Mechanical Systems. Proc. Int. School of Physics "Enrico Fermi", course 97, NorthHolland, Amsterdam (1986) 6. van Gunsteren, W. F.: in Truhler, D. (ed.): Mathematical Frontiers in Computational Chemical Physics. Springer-Verlag, New York (1988) 136 7. Valuev, A. A., Norman, G. E., Podlipchuk, V. Yu.: Molecular Dynamics Method: Theory and Applications. In: Samarskii, A. A., Kalitkin, N. N. (eds.): Mathematical Modelling. Nauka, Moscow (1989) 5 (in russian) 8. Allen, M. P., Tildesley, D. J.: Computer Simulation of Liquids. Clarendon, Oxford (1987) 9. Norman, G. E., Podlipchuk, V. Yu., Valuev, A. A.: J. Moscow Phys. Soc. (Institute of Physics Publishing, UK) 2 (1992) 7 10. Hoover, W. G.: Computational Statistical Mechanics. Elsevier, Amsterdam (1991) 11. Rapaport, D. C.: The Art of Molecular Dynamics Simulations. Cambridge University Press, Cambridge (1995) 12. Frenkel, D., Smith, B.: Understanding Molecular Simulations. Akademic Press, London (1996) 13. Zaslavsky, G.M.: Stochastisity of dynamic systems. Nauka, Moscow (1984); Harwood, Chur (1985) 14. Norman, G.E., Stegailov, V.V.: Zh. Eksp. Theor. Phys. 119 (2001) 1011 [J. of Experim. and Theor. Physics 92 (2001) 8791 15. Norman, G.E., Stegailov, V.V.: Stochastic and Dynamic Properties of Molecular Dynamics Systems: Simple Liquids, Plasma and Electrolytes, Polymers. Computer Physics Communications (proc. of the Europhysics Conf. on Computational Physics 2001) to be published 16. Morozov, I.V., Norman, G.E., Valuev, A.A.: Stochastic Properties of strongly coupled plasmas. Phys. Rev. E bf 63 (2001) 036405 17. Ueshima, Y., Nishihara, K., Barnett, D.M., Tajima, T., Furukawa, H.: Partice Simulation of Lyapunov Exponents in One-Component strongly coupled plasmas. Phys. Rev. E bf 55 (1997) 3439 18. Kravtsov, Yu.A. in: Kravtsov, Yu.A. (ed.) : Limits of Predictability. Springer, Berlin (1993) 173 19. Gertsenshtein, M.E., Kravtsov, Yu.A.: Zh. Eksp. Theor. Phys. 118 (2000) 761 [J. of Experim. and Theor. Physics 91 (2000) 6581 20. Rowlands, G.J.: Computational Physics 97 (1991) 235 21. Lopez-Marcos, M.A., Sanz-Serna, J.M., Diaz, J.C.: Are Gauss-Legendre Method Useful in Molecular Dynamics? J. Comput. Appl. Math. 67 (1996) 173 22. Lopez-Marcos, M.A., Sanz-Serna, J.M., Skeel, R.D .: Explicit Symplectic Integrators Using Hessian-Vector Products. SIAM J. Sci. Comput. 18 (1997) 223 23. Skripov, V. P., Koverda, V. P.: Spontaneous Crystallization of Supercooled Liquid. Nauka, Moscow (1984) (in russian)
Regular and Chaotic Motions of the P arametrically F orced Pendulum: Theory and Simulations Eugene I. Butikov St. Petersburg State Univ ersity, Russia E-mail: butik [email protected]
Abstract New types of regular and chaotic behaviour
of the parametrically
driven pendulum are discovered with the help of computer simulations. A simple qualitative physical explanation is suggested to the phenomenon of subharmonic resonances.
An approximate quantitative theory based on
the suggested approach is developed. The spectral composition of the subharmonic resonances is investigated quantitatively, and their boundaries in the parameter space are determined. pendulum
stability
The conditions of the inverted
are determined with a greater precision than they
have been known earlier. A close relationship between the upper limit of stability of the dynamically stabilized inverted pendulum and parametric resonance of the hanging down pendulum is established.
Most of the
newly discovered modes are waiting a plausible physical explanation.
1 Introduction An ordinary rigid planar pendulum whose suspension point is driven periodically is a paradigm of contemporary nonlinear dynamics. Being famous primarily due to its outstanding role in the history of science, this rather simple mec hanical system is also interesting because the dierential equation that describes its motion is frequently encountered in various problems of modern ph ysics. Mechanical analogues of dierent physical systems allow a direct visualization of their motion (at least with the help of sim ulations) and therefore can be very useful in gaining an intuitive understanding of complex phenomena. Depending on the frequency and amplitude of forced oscillations of the suspension point, this apparently simple mec hanical system exhibits an incredibly rich variety of nonlinear phenomena characterized by amazingly dierent types of motion. Some modes of suc h parametrically excited pendulum are quite simple indeed and agree well with our intuition, while others are very complicated and counterintuitive. Besides the commonly kno wn phenomenon of parametric resonance, the pendulum can execute man y other kinds of regular behavior. P.M.A. Sloot et al. (Eds.): ICCS 2002, LNCS 2331, pp. 1154−1169, 2002. Springer-Verlag Berlin Heidelberg 2002
Regular and Chaotic Motions of the Parametrically Forced Pendulum
1155
Among them we encounter a synchronized non-uniform unidirectional rotation in a full circle with a period that equals either the driving period or an integer multiple of this period. More complicated regular modes are formed by combined rotational and oscillatory motions synchronized (locked in phase) with oscillations of the pivot. Dierent competing modes can coexist at the same values of the driving amplitude and frequency. Which of these modes is eventually established when the transient is over depends on the starting conditions. Behavior of the pendulum whose axis is forced to oscillate with a frequency from certain intervals (and at large enough driving amplitudes) can be irregular, chaotic. The pendulum makes several revolutions in one direction, then swings for a while with permanently (and randomly) changing amplitude, then rotates again in the former or in the opposite direction, and so forth. For other values of the driving frequency and/or amplitude, the chaotic motion can be purely rotational, or, vice versa, purely oscillatory, without revolutions. The pendulum can make, say, one oscillation during each two driving periods (like at ordinary parametric resonance), but in each next cycle the motion (and the phase orbit) is slightly (and randomly) dierent from the previous cycle. Other chaotic modes are characterized by protracted oscillations with randomly varying amplitude alternated from time to time with full revolutions to one or the other side (intermittency). The parametrically forced pendulum can serve as an excellent physical model for studying general laws of the dynamical chaos as well as various complicated modes of regular behavior in simple nonlinear systems. A widely known interesting feature in the behavior of a rigid pendulum whose suspension point is forced to vibrate with a high frequency along the vertical line is the dynamic stabilization of its inverted position. Among recent new discoveries regarding the inverted pendulum, the most important are the destabilization of the (dynamically stabilized) inverted position at large driving amplitudes through excitation of period-2 (\ utter") oscillations [1]-[2], and the existence of n-periodic \multiple-nodding" regular oscillations [3]. In this paper we present a quite simple qualitative physical explanation of these phenomena. We show that the excitation of period-2 \ utter" mode is closely (intimately) related with the commonly known conditions of parametric instability of the non-inverted pendulum, and that the so-called \multiplenodding" oscillations (which exist both for the inverted and hanging down pendulum) can be treated as high order subharmonic resonances of the parametrically driven pendulum. The spectral composition of the subharmonic resonances in the low-amplitude limit is investigated quantitatively, and the boundaries of the region in the parameter space are determined in which these resonances can exist. The conditions of the inverted pendulum stability are determined with a greater precision than they have been known earlier. We report also for the rst time about several new types of regular and chaotic behaviour of the parametrically driven pendulum discovered with the help of computer simulations. Most of these exotic new modes are rather counterintuitive. They are still waiting a plausible physical explanation. Understanding such complicated behavior of this simple system is certainly a challenge to our physical intuition.
1156 E.I. Butikov
2 The physical system We consider the rigid planar pendulum whose axis is forced to execute a given harmonic oscillation along the vertical line with a frequency ! and an amplitude a, i.e., the motion of the axis is described by the following equation: z (t) = a sin !t
or z (t) = a cos !t:
(1)
The force of inertia Fin (t) exerted on the bob in the non-inertial frame of reference associated with the pivot also has the same sinusoidal dependence on time. This force is equivalent to a periodic modulation of the force of gravity. The simulation is based on a numerical integration of the exact dierential equation for the momentary angular de ection '(t). This equation includes the torque of the force of gravity and the instantaneous value of the torque exerted on the pendulum by the force of inertia that depends explicitly on time t: ' + 2 '_ + (!02
a 2 ! sin !t) sin ' = 0: l
(2)
The second term of Eq. (2) takes into account the braking frictional torque, assumed to be proportional to the momentary angular velocity '_ in the mathematical model of the simulated system. The damping constant is inversely proportional to the quality factor Q commonly used to characterize the viscous friction: Q = !0=2 . We note that oscillations about the inverted position can be formally described by the same dierential equation, Eq. (2), with negative values of !02 = g=l. When this control parameter !02 is diminished through zero to negative values, the constant (gravitational) torque in Eq. (2) also reduces down to zero and then changes its sign to the opposite. Such a \gravity" tends to bring the pendulum into the inverted position ' = , destabilizing the equilibrium position ' = 0 of the unforced pendulum.
3 Subharmonic resonances An understanding about pendulum's behavior in the case of rapid oscillations of its pivot is an important prerequisite for the physical explanation of subharmonic resonances (\multiple-nodding" oscillations). Details of the physical mechanism responsible for the dynamical stabilization of the inverted pendulum can be found in [4]. The principal idea is utterly simple: Although the mean value of the force of inertia Fin (t), averaged over the short period of these oscillations, is zero, the averaged over the period value of its torque about the axis is not zero. The reason is that both the force Fin (t) and the arm of this force vary in time in the same way synchronously with the axis' vibrations. This non-zero torque tends to align the pendulum along the direction of forced oscillations of the axis. For given values of the driving frequency and amplitude, the mean torque of the force of inertia depends only on the angle of the pendulum's de ection from the direction of the pivot's vibration.
Regular and Chaotic Motions of the Parametrically Forced Pendulum
1157
In the absence of gravity the inertial torque gives a clear physical explanation of existence of the two stable equilibrium positions that correspond to the two preferable orientations of the pendulum's rod along the direction of the pivot's vibration. With gravity, the inverted pendulum is stable with respect to small deviations from this position provided the mean torque of the force of inertia is greater than the torque of the force of gravity that tends to tip the pendulum down. This occurs when the following condition is ful lled: a2!2 > 2gl, or (a=l)2 > 2(!0=!)2 (see, e.g., [4]) However, this is only an approximate criterion of dynamic stability of the inverted pendulum, which is valid only for small amplitudes of forced vibrations of the pivot (a l). Below we obtain a more precise criterion [see Eq. (5)]. The complicated motion of the pendulum whose axis is vibrating at a high frequency can be considered approximately as a superposition of two rather simple components: a \slow" or \smooth" component (t), whose variation during a period of forced vibrations is small, and a \fast" (or \vibrational") component. This approach was rst used by Kapitza [5] in 1951. Being de ected from the vertical position by an angle that does not exceed max [where cos max = 2gl=(a2 !2 )], the pendulum will execute relatively slow oscillations about this inverted position. This slow motion is executed both under the mean torque of the force of inertia and the force of gravity. Rapid oscillations caused by forced vibrations of the axis superimpose on this slow motion of the pendulum. With friction, the slow motion gradually damps, and the pendulum wobbles up settling eventually in the inverted position. Similar behavior of the pendulum can be observed when it is de ected from the lower vertical position. The frequencies !up and !down of small slow oscillations about the inverted and hanging down vertical positions are given by the 2 = !2 (a=l)2 =2 !2 and !2 2 2 2 approximate expressions !up 0 p down = ! (a=l) =2 + !0 , respectively. These formulas yeld !slow = !(a=l)= 2 for the frequency of small slow oscillations of the pendulum with vibrating axis in the absence of the gravitational force. When the driving amplitude and frequency lie within certain ranges, the pendulum, instead of gradually approaching the equilibrium position (either dynamically stabilized inverted position or ordinary downward position) by the process of damped slow oscillations, can be trapped in a n-periodic limit cycle locked in phase to the rapid forced vibration of the axis. In such oscillations the phase trajectory repeats itself after n driving periods T . Since the motion has period nT , and the frequency of its fundamental harmonic equals !=n (where ! is the driving frequency, this phenomenon can be called a subharmonic resonance of n-th order. For the inverted pendulum with a vibrating pivot, periodic oscillations of this type were rst described by Acheson [3], who called them \multiple-nodding" oscillations. An example of such stationary oscillations whose period equals six periods of the axis is shown in Fig. 1. The left-hand upper part of the gure shows the complicated spatial trajectory of the pendulum's bob at these multiple-nodding oscillations. The left-hand lower part shows the closed looping trajectory in the phase plane (', '_ ). Right-hand
1158 E.I. Butikov
Figure 1: The spatial path, phase orbit, and graphs of stationary oscillations with the period that equals six periods of the oscillating axis. The graphs are obtained by a numerical integration of the exact dierential equation, Eq. (2), for the momentary angular de ection side of Fig. 1, alongside the graphs of '(t) and '_ (t), shows also their harmonic components and the graphs of the pivot oscillations. The fundamental harmonic whose period equals six driving periods dominates in the spectrum. We may treat it as a subharmonic (as an \undertone") of the driving oscillation. This principal harmonic describes the smooth component of the compound period-6 oscillation. We emphasize that the modes of regular n-periodic oscillations (subharmonic resonances), which have been discovered in investigations of the dynamically stabilized inverted pendulum, are not speci c for the inverted pendulum. Similar oscillations can be executed also (at appropriate values of the driving parameters) about the ordinary (downward hanging) equilibrium position. Actually, the origin of subharmonic resonances is independent of gravity, because such synchronized with the pivot \multiple-nodding" oscillations can occur in the absence of gravity about any of the two equivalent dynamically stabilized equilibrium positions of the pendulum with a vibrating axis. The natural slow oscillatory motion is almost periodic (exactly periodic in the absence of friction). A subharmonic resonance of order n can occur if one cycle of this slow motion covers approximately n driving periods, that is, when the driving frequency ! is close to an integer multiple n of the natural frequency of slow oscillations near either the inverted or the ordinary equilibrium position: ! = n!up or ! = n!down . In this case the phase locking can occur, in which one cycle of the slow motion is completed exactly during n driving periods. Syn-
Regular and Chaotic Motions of the Parametrically Forced Pendulum
1159
chronization of these modes with the oscillations of the pivot creates conditions for systematic supplying the pendulum with the energy needed to compensate for dissipation, and the whole process becomes exactly periodic. The slow motion (with a small angular excursion) can be described by a sinusoidal time dependence. Assuming !down;up = !=n (n driving cycles during one cycle of the slow oscillation), we nd for the minimal driving amplitudes (for the boundaries of the subharmonic resonances) the values mmin =
p2(1=n2 k);
(3)
where k = (!0=!)2 . The limit of this expression at n ! 1 gives the mentioned earlier condition of stability of the inverted pendulum: mmin = p p approximate 2k = 2(!0 =!) (here k < 0, jkj = j!02=!2 j). The spectrum of stationary n-periodic oscillations consists primarily of the fundamental harmonic A sin(!t=n) with the frequency !=n, and two high harmonics of the orders n 1 and n + 1. To improve the theoretical values for the boundaries of subharmonic resonances, Eq. (3), we use a trial solution '(t) with unequal amplitudes of the two high harmonics. Since oscillations at the boundaries have in nitely small amplitudes, we can exploit instead of Eq. (2) the linearized (Mathieu) equation (with = 0). Thus we obtain the critical (minimal) driving amplitude mmin at which n-period mode '(t) can exist: 2 [n6k(k
m2min = 4 n
1)2
n4(3k2 + 1) + n2 (3k + 2) [n2 (1 k) + 1]
1]
:
(4)
The limit of mmin , Eq. (4), at n ! 1 gives an improved formula for the lower boundary of the dynamic stabilization of the inverted p pendulum instead 2k: of the commonly known approximate criterion mmin = mmin =
p
2k(1
k)
(k < 0):
(5)
The minimal amplitude mmin that provides the dynamic stabilization is shown as a function of k = (!0=!)2 (inverse normalized driving frequency squared) by the left curve (n ! 1) in Fig. 2. The other curves to the right from this boundary show the dependence on k of minimal driving amplitudes for the subharmonic resonances of several orders (the rst curve for n = 6 and the others for n values diminishing down to n = 2 from left to right). At positive values of k these curves correspond to the subharmonic resonances of the hanging down parametrically excited pendulum. Subharmonic oscillations of a given order n (for n > 2) are possible to the left of k = 1=n2, that is, for the driving frequency ! > n!0 . The curves in Fig. 2 show that as the driving frequency ! is increased beyond the value n!0 (i.e., as k is decreased from the critical value 1=n2 toward zero), the threshold driving amplitude (over which n-order subharmonic oscillations are possible) rapidly increases. The limit of a very high driving frequency (!=!0 ! 1), in which the gravitational force is insigni cant compared with the force of inertia, or, which is essentially the same, the limit of zero gravity (!0=! ! 0), corresponds to k = 0, that is, to the
1160 E.I. Butikov
Figure 2: The driving amplitude at the boundaries of the dynamic stabilization of the inverted pendulum and subharmonic resonances points of intersection of the curves in Fig. 2 with the m-axis. The continuations of these curves further to negative k values describe the transition through zero gravity to the \gravity" directed upward, which is equivalent to the case of an inverted pendulum in ordinary (directed downward) gravitational eld. Therefore the same curves at negative k values give the threshold driving amplitudes for subharmonic resonances of the inverted pendulum. 1 Smooth non-harmonic oscillations of a nite angular excursion are characterized by a greater period than the small-amplitude harmonic oscillations executed just over the parabolic bottom of this well. Therefore large-amplitude period-6 oscillations shown in Fig. 1 (their swing equals 55Æ) occur at a considerably greater value of the driving amplitude (a = 0:265 l) than the critical (threshold) value amin = 0:226 l. By virtue of the dependence of the period of non-harmonic smooth motion on the swing, several modes of subharmonic resonance with dierent values of n can coexist at the same amplitude and frequency of the pivot. 1 Actually the curves in Fig. 2 are plotted not according to Eq. (4), but rather with the help of a somewhat more comlicated formula (not cited in this paper), which is obtained by holding one more high order harmonic component in the trial function.
Regular and Chaotic Motions of the Parametrically Forced Pendulum
1161
4 The upper boundary of the dynamic stability When the amplitude a of the pivot vibrations is increased beyond a certain critical value amax , the dynamically stabilized inverted position of the pendulum loses its stability. After a disturbance the pendulum does not come to rest in the up position, no matter how small the release angle, but instead eventually settles into a nite amplitude steady-state oscillation (about the inverted vertical position) whose period is twice the driving period. This loss of stability of the inverted pendulum has been rst described by Blackburn et al. [1] (the \ utter" mode) and demonstrated experimentally in [2]. The latest numerical investigation of the bifurcations associated with the stability of the inverted state can be found in [6]. The curve with n = 2 in Fig. 2 shows clearly that both the ordinary parametric resonance and the period-2 \ utter" mode that destroys the dynamic stability of the inverted state belong essentially to the same branch of possible steadystate period-2 oscillations of the parametrically excited pendulum. Indeed, the two branches of this curve near k = 0:25 (that is, at ! 2!0 ) describe the well known boundaries of the principle parametric resonance. Therefore the upper boundary of dynamic stability for the inverted pendulum can be found directly from the linearized dierential equation of the system. In the p case !0 = 0 (which corresponds to the absence of gravity) we nd mmax = 3( 13 3)=4 = 0:454, and the corresponding ratio of amplitudes of the third harmonic to the funp damental one equals A3=A1 = ( 13 3)=6 = 0:101. A somewhat more complicated calculation in which higher harmonics (up to the 7th) in '(t) are taken into account yields for mmax and A3 =A1 the values that coincide (within the assumed accuracy) with those cited above. These values agree well with the simulation experiment in conditions of the absence of gravity (!0 = 0) and very small angular excursion of the pendulum. When the normalized amplitude of the pivot m = a=l exceeds the critical value mmax = 0:454, the swing of the period-2 \ utter" oscillation (amplitude A1 of the fundamental p harmonic) increases in proportion to the square root of this excess: A1 / a amax . This dependence follows from the nonlinear dierential equation of the pendulum, Eq. (2), if sin ' in it is approximated as ' '3 =6, and agrees well with the simulation experiment for amplitudes up to 45Æ. As the normalized amplitude m = a=l of the pivot is increased over the value 0:555, the symmetry-breaking bifurcation occurs: The angular excursions of the pendulum to one side and to the other become dierent, destroying the spatial symmetry of the oscillation and hence the symmetry of the phase orbit. As the pivot amplitude is increased further, after m = 0:565 the system undergoes a sequence of period-doubling bifurcations, and nally, at m = 0:56622 (for Q = 20), the oscillatory motion of the pendulum becomes replaced, at the end of a very long chaotic transient, by a regular unidirectional period-1 rotation. Similar (though more complicated) theoretical investigation of the boundary conditions for period-2 stationary oscillations in the presence of gravity allows us to obtain the dependence of the critical (destabilizing) amplitude mmax of the pivot on the driving frequency !. In terms of k = (!0=!)2 this dependence
1162 E.I. Butikov
has the following form:
p
mmax = ( 117 232k + 80k2
9 + 4k)=4:
(6)
The graph of this boundary is shown in Fig. 2 by the curve marked as n = 2. The critical driving amplitude tends to zero as k ! 1=4 (as ! ! 2!0). This condition corresponds to ordinary parametric resonance of the hanging down pendulum, which is excited if the driving frequency equals twice the natural frequency. For k > 1=4 (! < 2!0) Eq. (6) yields negative m whose absolute value jmj corresponds to stationary oscillations at the other boundary (to the right of k = 0:25, see Fig. 2). If the driving frequency exceeds 2!0 (that is, if k < 0:25), a nite driving amplitude is required for in nitely small steady parametric oscillations even in the absence of friction. The continuation of the curve n = 2 to the region of negative k values corresponds to the transition from ordinary downward gravity through zero to \negative," or upward \gravity," or, which is the same, to the case of inverted pendulum in ordinary (directed down) gravitational eld. Thus, the same formula, Eq. (6), gives the driving amplitude (as a function of the driving frequency) at which both the equilibrium position of the hanging down pendulum is destabilized due to excitation of ordinary parametric oscillations, and the dynamically stabilized inverted equilibrium position is destabilized due to excitation of period-2 \ utter" oscillations. We can interpret this as an indication that both phenomena are closely related and have common physical nature. All the curves that correspond to subharmonic resonances of higher orders (n > 2) lie between this curve and the lower boundary of dynamical stabilization of the inverted pendulum.
5 New types of regular and chaotic motions In this section we report about several modes of regular and chaotic behavior of the parametrically driven pendulum, which we have discovered recently in the simulation experiments. As far as we know, such modes haven't been described in literature. Figure 3 shows a regular period-8 motion of the pendulum, which can be characterized as a subharmonic resonance of a fractional order, speci cally, of the order 8/3 in this example. Here the amplitude of the fundamental harmonic (whose frequency equals !=8) is much smaller than the amplitude of the third harmonic (frequency 3!=8). This third harmonic dominates in the spectrum, and can be regarded as the principal one, while the fundamental harmonic can be regarded as its third subharmonic. Considerable contributions to the spectrum are given also by the 5th and 11th harmonics of the fundamental frequency. Approximate boundary conditions for small-amplitude stationary oscillations of this type (n=3-order subresonance) can be found analytically from the linearized dierential equation by a method similar to that used above for n-order subresonance: we can try as '(t) a solution consisting of spectral components with frequencies 3!t=n, (n 3)!t=n, and (n + 3)!t=n:
Regular and Chaotic Motions of the Parametrically Forced Pendulum
1163
Figure 3: The spatial path, phase orbit, and graphs of stationary oscillations that can be treated as a subharmonic resonance of a fractional order (8/3)
'(t) = A3 sin(3!t=n) + An 3 sin[(n 3)!t=n] + An+3 sin[(n + 3)!t=n]: (7) For the parametrically driven pendulum in the absence of gravity such a calculation gives the following expression for the minimal driving amplitude:
mmin =
p 3 2(n2 32) p : n2 n2 + 32
(8)
The analytical results of calculations for n 8 agree well with the simulations, especially if one more high harmonic is included in the trial function '(t). One more type of regular behavior is shown in Fig. 4. This mode can be characterized as resulting from a multiplication of the period of a subharmonic resonance, speci cally, as tripling of the six-order subresonance in this example. Comparing this gure with Fig. 1, we see that in both cases the motion is quite similar during any cycle of six consecutive driving periods each, but in Fig. 4 the motion during each next cycle of six periods is slightly dierent from the preceding cycle. After three such cycles (of six driving periods each) the phase orbit becomes closed and then repeats itself, so the period of this stationary motion equals 18 driving periods. However, the harmonic component whose period equals six driving periods dominates in the spectrum (just like in the spectrum of period-6 oscillations in Fig. 1), while the fundamental harmonic (frequency !=18) of a small amplitude is responsible only for tiny divergences between the adjoining cycles consisting of six driving periods.
1164 E.I. Butikov
Figure 4: The spatial path, phase orbit, and graphs of period-18 oscillations Such multiplications of the period are characteristic of large amplitude oscillations at subharmonic resonances both for the inverted and hanging down pendulum. Figure. 5 shows a stationary oscillation with a period that equals ten driving periods. This large amplitude motion can be treated as originating from a period-2 oscillation (that is, from ordinary principal parametric resonance) by a ve-fold multiplication of the period. The harmonic component with half the driving frequency (!=2) dominates in the spectrum. But in contrast to the preceding example, the divergences between adjoining cycles consisting of two driving periods each are generated by the contribution of a harmonic with the frequency 3!=10 rather than of the fundamental harmonic (frequency !=10) whose amplitude is much smaller. One more example of complicated steady-state oscillation is shown in Fig. 6. This period-30 motion can be treated as generated from the period-2 principal parametric resonance rst by ve-fold multiplication of the period (resulting in period-10 oscillation), and then by next multiplication (tripling) of the period. Such large-period stationary regimes are characterized by small domains of attraction consisting of several disjoint islands on the phase plane. Other modes of regular behavior are formed by unidirectional period-2 or period-4 (or even period-8) rotation of the pendulum or by oscillations alternating with revolutions to one or to both sides in turn. Such modes have periods constituting several driving periods. At large enough driving amplitudes the pendulum exhibits various chaotic regimes. Chaotic behaviour of nonlinear systems has been a subject of intense interest during recent decades, and the forced pendulum serves as an excellent physical model for studying general laws of the dynamical chaos [6] { [14].
Regular and Chaotic Motions of the Parametrically Forced Pendulum
1165
Figure 5: The spatial path, phase orbit, and graphs of period-10 oscillations Next we describe several dierent kinds of chaotic regimes, which for the time being have not been mentioned in literature. Poincare mapping, that is, a stroboscopic picture of the phase plane for the pendulum taken once during each driving cycle after initial transients have died away, gives an obvious and convenient means to distinguish between regular periodic behavior and persisting chaos. A steady-state subharmonic of order n would bee seen in the Poincare map as a systematic jumping between n xed mapping points. When the pendulum motion is chaotic, the points of Poincare sections wander randomly, never exactly repeating. Their behavior in the phase plane gives an impression of the strange attractor for the motion in question. Figure. 7 shows an example of a purely oscillatory two-band chaotic attractor for which the set of Poincare sections consists of two disjoint islands. This attractor is characterized by a fairly large domain of attraction in the phase plane. The two islands of the Poincare map are visited regularly (strictly in turn) by the representing point, but within each island the point wanders irregularly from cycle to cycle. This means that for this kind of motion the ow in the phase plane is chaotic, but the distance between any two initially close phase points within this attractor remains limited in the progress of time: The greatest possible distance in the phase plane is determined by the size of these islands of the Poincare map. Figure. 8 shows the chaotic attractor that corresponds to a slightly reduced friction, while all other parameters are unchanged. Gradual reduction of friction causes the islands of Poincare sections to grow and coalesce, and to form nally a strip-shaped set occupying considerable region of the phase plane. As in the preceding example, each cycle of these oscillations (consisting of two driving
1166 E.I. Butikov
Figure 6: The spatial path, phase orbit, and graphs of period-30 oscillations. periods) slightly but randomly varies from the preceding one. However, in this case the large and almost constant amplitude of oscillations occasionally (after a large but unpredictable number of cycles) considerably reduces or, vice versa, increases (sometimes so that the pendulum makes a full revolution over the top). These decrements and increments result sometimes in switching the phase of oscillations: the pendulum motion, say, to the right side that occurred during even driving cycles is replaced by the motion in the opposite direction. During long intervals between these seldom events the motion of the pendulum is purely oscillatory with only slightly (and randomly) varying amplitude. This kind of intermittent irregular behavior diers from the well-known so-called tumbling chaotic attractor that exists over a relatively broad range of parameter space. The tumbling attractor is characterized by random oscillations (whose amplitude varies strongly from cycle to cycle), often alternated with full revolutions to one or the other side. Figure 9 illustrates one more kind of strange attractors. In this example the motion is always purely oscillatory, and nearly repeats itself after each six driving periods. The six bands of Poincare sections make two groups of three isolated islands each. The representing point visits these groups in alternation. It also visits the islands of each group in a quite de nite order, but within each island the points continue to bounce from one place to another without any apparent order. The six-band attractor has a rather extended (and very complicated in shape) domain of attraction. Nevertheless, at these values of the control parameters the system exhibits multiple asymptotic states: The chaotic attractor coexists with several periodic regimes. Chaotic regimes exist also for purely rotational motions. Poincare sections
Regular and Chaotic Motions of the Parametrically Forced Pendulum
1167
Figure 7: Chaotic attractor with a two-band set of Poincare sections for such rotational chaotic attractors can make several isolated islands in the phase plane. A possible scenario of transition to such chaotic modes from unidirectional regular rotation lies through an in nite sequence of period-doubling bifurcations occurring when a control parameter (the driving amplitude or frequency or the braking frictional torque) is slowly varied without interrupting the motion of the pendulum. However, there is no unique route to chaos for more complicated chaotic regimes described above.
6 Concluding remarks The behavior of the parametrically excited pendulum discussed in this paper is much richer in various modes than we can expect for such a simple physical system relying on our intuition. Its nonlinear large-amplitude motions can hardly be called \simple." The simulations show that variations of the parameter set (dimensionless driving amplitude a=l, normalized driving frequency !=!0 , and quality factor Q) result in dierent regular and chaotic types of dynamical behavior. In this paper we have touched only a small part of existing stationary states, regular and chaotic motions of the parametrically driven pendulum. The pendulum's dynamics exhibits a great variety of other asymptotic rotational, oscillatory, and combined (both rotational and oscillatory) multiple-periodic stationary states (attractors), whose basins of attraction are characterized by a surprisingly complex (fractal) structure. Computer simulations reveal also intricate sequences of bifurcations, leading to numerous intriguing chaotic regimes. Most of them remained beyond the scope of this paper, and those mentioned
1168 E.I. Butikov
Figure 8: Chaotic attractor with a strip-like set of Poincare sections. here are still waiting a plausible physical explanation. With good reason we can suppose that this seemingly simple physical system is inexhaustible.
References [1]
Blackburn J A, Smith H J T, Groenbech-Jensen N 1992 Stability and Hopf bifurcations in an inverted pendulum
[2]
60 (10) 909
{ 911
A 448 89 { 95
Butikov E I 2001 On the dynamic stabilization of an inverted pendulum Phys.
[5]
{ 908
Acheson D J 1995 Multiple-nodding oscillations of a driven inverted pendulum Proc. Roy. Soc. London
[4]
60 (10) 903
Smith H J T, Blackburn J A 1992 Experimental study of an inverted pendulum Am. J. Phys.
[3]
Am. J. Phys.
69 (7) 755 { 768
Am. J.
Kapitza P L 1951 Dynamic stability of the pendulum with vibrating suspension point
Soviet Physics { JETP
papers of P. L. Kapitza
21
(5) 588 { 597 (in Russian), see also
Collected
edited by D. Ter Haar, Pergamon, London (1965), v. 2,
pp. 714 { 726.
[6]
Sang-Yoon Kim and Bambi Hu 1998 Bifurcations and transitions to chaos in an inverted pendulum
[7]
Phys. Rev.
E
58, (3) 3028
{ 3035
McLaughlin J B 1981 Period-doubling bifurcations and chaotic motion for a parametrically forced pendulum
J. Stat. Physics
24 (2) 375 { 388
Regular and Chaotic Motions of the Parametrically Forced Pendulum
1169
Figure 9: An oscillatory six-band chaotic attractor.
[8]
[9]
[10]
[11]
Koch B P, Leven R W, Pompe B, and Wilke C 1983 Experimental evidence for chaotic behavior of a parametrically forced pendulum Phys. Lett. A 96 (5) 219 { 224 Leven R W, Pompe B, Wilke C, and Koch B P 1985 Experiments on periodic and chaotic motions of a parametrically forced pendulum Physica D 16 (3) 371 { 384 Willem van de Water and Marc Hoppenbrouwers 1991 Unstable periodic orbits in the parametrically excited pendulum Phys. Rev. A 44 (10) 6388 - 6398 Starrett J and Tagg R 1995 Control of a chaotic parametrically driven pendulum 74, (11) 1974 { 1977
Phys. Rev. Lett.
[12]
Cliord M J and Bishop S R 1998 Inverted oscillations of a driven pendulum A 454 2811 { 2817
Proc. Roy. Soc. London
[13]
[14]
Sudor D J, Bishop S R 1999 Inverted dynamics of a tilted parametric pendulum Eur. J. Mech. A/Solids 18 517 { 526 Bishop S R, Sudor D J 1999 The `not quite' inverted pendulum 9 (1) 273 { 285
Bifurcation and Chaos
Int. Journ.
Lyapunov Instability and Collective Tangent Space Dynamics of Fluids Harald A. Posch and Ch. Forster Institut fu ¨r Experimentalphysik, Universit¨ at Wien, Boltzmanngasse 5, A-1090 Vienna, Austria
Abstract. The phase space trajectories of many body systems charateristic of isimple fluids are highly unstable. We quantify this instability by a set of Lyapunov exponents, which are the rates of exponential divergence, or convergence, of infinitesimal perturbations along selected directions in phase space. It is demonstrated that the perturbation associated with the maximum Lyapunov exponent is localized in space. This localization persists in the large-particle limit, regardless of the interaction potential. The perturbations belonging to the smallest positive exponents, however, are sensitive to the potential. For hard particles they form well-defined long-wavelength modes. The modes could not be observed for systems interacting with a soft potential due to surprisingly large fluctuations of the local exponents.
1
Lyapunov spectra
Recently, molecular dynamics simulations have been used to study many body systems representing simple fluids or solids from the point of view of dynamical system theory. Due to the convex dispersive surface of the atoms, the phase trajectory of such systems is highly unstable and leads to an exponential growth, or decay, of small (infintesimal) perturbationis of an initial state along specified directions in phase space. This so-called Lyapunov instability is described by a set of rate constants, the Lyapunov exponents {λl , l = 1, . . . , D}, to which we refer as the Lyapunov spectrum. Conventionally, the exponents are taken to be ordered by size, λl ≥ λl+1 . There are altogether D = 2dN exponents, where d is the dimension of space, N is the number of particles, and D is the dimension of the phase space. For fluids in nonequilibrium steady states close links between the Lyapunov spectrum and macroscopic dynamical properties, such as transport coefficients, irreversible entropy production, and the Second Law of thermodynamcis have been established [1–4]. This important result provided the motivation for us to examine the spatial structure of the various perturbed states associated with the various exponents. Here we present some of our results for two simple many-body systems representing dense two-dimensional fluids in thermodynmic equilibrium. The first model consists of N hard disks (HD) interacting with hard elastic collisions, the second of N soft disks interacting with a purely repulsive Weeks-Chandler-Anderson (WCA) potential. P.M.A. Sloot et al. (Eds.): ICCS 2002, LNCS 2331, pp. 1170−1175, 2002. Springer-Verlag Berlin Heidelberg 2002
Lyapunov Instability and Collective Tangent Space Dynamics of Fluids
1171
The instantaneous state of a planar particle system is given by the 4Ndimensional phase space vector Γ = {ri , pi , ; i = 1, . . . , N }, where ri and pi denote the respective position and linear momentum of molecule i. An infinitesimal perturbation δΓ = {δri , δpi ; i = 1, . . . , N } evolves according to motione equations obtained by linearizing the equations of motion for Γ(t). For ergodic systems there exist D = 4N orthonormal initial vectors {δΓl (0); l = 1, . . . , 4N } in tangent space, such that the Lyapunov exponents λl = lim
t→∞
1 |δΓl (t)| ln , l = 1, . . . , 4N. t |δΓl (0)|
(1)
exist and are independent of the initial state. Geometrically, the Lyapunov spectrum describes the stretching and contraction along linearly-independent phase space directions of an infinitesimal hypersphere co-moving with the flow. For equilibrium systems the symplectic nature of the motion equations assures the conjugate pairing rule to hold [5]: the exponents appear in pairs, λl = λ4N +1−l = 0, so that only half of the spectrum {λ1≤l≤2N } needs to be cqaluclated. The sum of all Lyapunov exponents vanishes, which, according to Liouville’s theorem, expresses the fact that the phase volume is strictly conserved for Hamiltonian systems. Six if the exponents {λ2N −2≤l≤2N +3 } always vanish as a consequence of the conservation of energy, momentum, and center of mass, and of the nonexponential time evolution of a perturbation vector parallel to the phase flow. For the computation of a complete spectrum a variant of a classical algorithm by Benettin et al. and Shimada et al. is used [6, 7]. It follows the time evolution of the reference trajectory and of an orthonormal set of tangent vectors {δΓl (t); l = 1, . . . , 4N }, where the latter is periodically re-orthonormalized with a GramSchmidt (GS) procedure after consecutive time intercals ∆tGS . The Lyapunov exponents are determined from the time-averaged renormalization factors. For the hard disk systems the free evolution of the particles is interrupted by hard elastic collisions and a linearized collision map needs to be calculated as was demonstrated by Dellago et al. [9]. Although we make use of the conjugate pairing symmetry and compute only the positive branch of the spectrum, we are presently restricted to about 1000 particles by our available computer resources. For our numerical work reduced units are used. In the case of the WeeksChendler-Anderson interaction potential, 4[(σ/r)12 − (σ/r)6 ] + , r < 21/6 σ φ(r) = , (2) 0, r ≥ 21/6 σ. the particle mass m, the particle diameter σ, and the time (mσ2 /)1/2 are unity. In this work we restrict our discussion to a thermodynamic state with a total energy per particle, E/N , also equal to unity. For the hard-disk fluid (N mσ2 /K)1/2 is the unit of time, where K is the total kinetic energy, which is equal to the total energy E of the system. There is no potential energy in this case. The reduced temperature T = K/N , where Boltzmann’s constant is also taken unity. In the following, Lyapunov exponents for the two model fluids will be compared for equal temperatures (and not for equal total energy). This requires a
1172 H.A. Posch and C. Forster
rescaling of the hard-disk exponents by a factor of KW CA /KHD to account for the difference in temperature. All our simulations are for a reduced density ρ ≡ N/V = 0.7, where the simulation box is a square with a volume V = L2 and a side length L. Periodic boundaries are used throughout. Figure 1. Lyapunov spectrum of a dense two-dimensional fluid consisting of N = 100 particles at a density ρ = 0.7 and a temperature T = 0.75. The label WCA refers to a smooth Weeks-Chandler-Anderson interaction potential, whereas HD is for the hard disk system. As an example we compare in Fig. 1 the Lyapunov spectrum of a WCA fluid with N = 100 particles to an analogous spectrum for a hard disk system at the same temperature (T = 0.75) and density (ρ = 0.7). A renormalized index l/2N is used on the abscissa. It is surprising that the two spectra differ so much in shape and in magnitude. The difference persists in the thermodynamic limit. The step-like structure of the hard disk spectrum for l/2N close to 1 is an indication of a coherent wave-like shape of the associated perturbation. We defer the discussion of the so-called Lyapunov modes to Section 3.
2
Fluctuating local exponents
We infer from Equ. (1) that the Lyapunov exponents are time averages over an (infinitely) long trajectory and are global properties of the system. This time average can be written as τ λl = lim λ l (Γ(t))dt ≡ λ l , (3) τ →∞
0
iwhere the (implicitely) time-dependent function λ l (Γ(t)) depends on the state Γ(t) in phase space the system occupies at time t. Thus, λ l (Γ) is called a local Lyapunov exponent. It may be estimated from λ l (Γ(t)) =
1 |δΓl (Γ(t + ∆tGS )| , ln ∆tGS |δΓl (Γ(t)|
(4)
where t and t+∆tGS refer to times immediately after consecutive Gram-Schmidt re-orthonormalization steps. Its time average, denoted by · · ·, along a trajectory gives the global exponent λl . The local exponents fluctuate considerably along a trajectory. This is demonstrated in Fig. ??, where we have plotted the 2 second moment λ l as a function of the Lyapunov index 1 ≤ l ≤ 4N for a system of 16 particles, both for the WCA and HD particle interactions. l = 1 refers to the maximum exponent, and l = 64 to the most negative exponent. The points for 30 ≤ l ≤ 35 correspond to the 6 vanishing exponents and are not shown. We infer from this figure that for the particles interacting with the smooth WCA potential the fluctuations of the local exponents, whose average give rise to global exponents approaching zero for l → 2N . For the hard disk system, however, the relative improtance of the fluctuations also becomes minimal in this limit. We shall return to this point in Section 3.
Lyapunov Instability and Collective Tangent Space Dynamics of Fluids
1173
2
We note that the computation of the second moments λ l for the hard disk system requires some care. Due to the hard core collisions they depend strongly 2 on ∆tGS for small ∆tGS . The mean square deviations, λ l − λ l 2 , vary with 1/∆tGS for small ∆tGS , as is demonstrated in Fig. 3 for the maximum local exponent. However, the shape of the fluctuation spectrum is hardly affected by the size of the renormalization interval.
3
The maximum exponent
The maximum Lyapunov exponent is the rate constant for the fastest growth of phase space perturbations in a system. There is strong numerical evidence for the existence of the thermodynamic limit {N → ∞, ρ = N/V constant } for λ1 and, hence, for the whole spectrum. Furthermore, the associated perturbation is strongly localized in space. This may be demonstrated by projecting the tangent vector δΓ1 onto the subspaces spanned by the perturbation components contributed by the individual particles. The squared norm of this projection, δi2 (t) ≡ (δri )2l + (δpi )2l , indicates how active a particle i is engaged in the growth process of the pertubation associated with λ1 . In Fig. 4 δi2 (t) is plotted along the vertical for all particles of a hard disk system at the respective positions (xi , yi ) of the disks in space, and the ensuing surface is interpolated over a periodic grid covering the simulation box. A strong localization of the active particles is observed at any instant of time. Similar, albeit slightly broader peaks are observed for the WCA system. This localization is a consequence of two mechanisms: firstly, after a collision the delta-vector components of two colliding molecules are linear functions of their pre-collision values and have only a chance of further growth if their values before the collision were already far above average. Secondly, each renormalization step tends to reduce the (already small) components of the other non-colliding particles even further. Thus, the competition for maximum growth of tangent vector components favors the collision pair with the largest components. The localization also persists in the thermodynamic limit. To show this we follow Milanovi´c et al. [8], square all 4N components of the perturbation vector δΓ1 and order the squared components [δΓ1 ]2j ; j = 1, . . . , 4N according to size. By adding them up, starting with the largest, we determine the smallest number of terms, A ≡ 4N C1,Θ , required for the sum to exceed a threshold Θ. Then, C1,Θ = A/4N may be taken as a relative measure for the number of components actively contributing to λ1 : 4N C1,Θ 2 Θ≤ [δΓ1 ]2i ≥ [δΓ1 ]2j for i < j. [δΓ1 ]s , (5) s=1
Here, · · · implies a time average. Obviously, C1,1 = 1. In Fig. 5 C1,Θ is shown for Θ = 0.98 as a function of the particle number N , both for the WCA fluid and for the hard disk system. It converges to zero if our data are extrapolated to the
1174 H.A. Posch and C. Forster
thermodynamic limit, N → ∞. This supports our assertion that in an infinite system only a vanishing part of the tangent-vector components (and, hence, of the particles) contributes significantly to the maximum Lyapunov exponent at any instant of time.
4
Lyapunov modes
We have mentioned already the appearance of a step-like structure in the Lyapunov sepctrum of the hard disk system for the positive exponents closest to zero. They are a consequence of coherent wave-like spatial patterns generated by the perturbation vector components associated with the individual particles. In Fig. 6 this is visualized by plotting the perturbations in the x (bottom surface) and y direction (top surface), {δxi , i = 1, . . . , N } and {δyi , i = 1, . . . , N }, respectively, along the vertical direction at the instantaneous particle positions (xi , yi ) of all particles i. This figure depicts a transversal Lyapunov mode, for which the perturbation is parpendicular to the wave vector, for a hard disk system consisting of N = 1024 particles and for a perturbation vector associated with the smallest positive exponent λ2045 . An analogous plot for δpx and δpy l = 2045 is identical to that of δx and δy in Fig. 6, with the same phase for the waves. This is a consequence of the fact that the perturbations are solutions first-order differential equation instead of second. Furthermore, the exponents for 2042 ≤ l ≤ 2045 are equal. The four-fold degeneracy of non-propagating transversal modes, and an analogous eight-fold degeneracy of propagating longitudinal modes, are responsible for a complicated step structure for l close to 2N , which has been studied in detail in Refs. XXXXX. The wave length of the modes and the value of the corresponding exponents are determined by the linear extension L of the simulation box. There is a kind of linear dispersion relation [10] according to which the smallest positive exponent is proprtional to 1/L. This assures that for a simulation box with aspect ratio 1 there is no positive lower bound for the positive exponents of hard disk systems in the thermodynamic limit. So far, our discussion of modes is only for the hard disk fluid. In spite of a considerable computational effort we have not yet been able to indentify modes for two-dimensional fluid systems with a soft interaction potential such as WCA or similar potentials. The reason for this surprising fact seems to be the very strong fluctions of the local exponents as discussed in Section 2. The fluctuations obscure any mode in the system in spite of considerable averaging and make a positive identification very difficult. Three-dimensional systems are just beyond computational feasibility at present, although the use of parallel machines may change this scenario soon. We are grateful to Christoph Dellago, Robin Hirschl, Bill Hoover, and Ljubo Milanovi´c for many illuminating discussions. This work was supported by the Austrian Fonds zur F¨ orderung der wissenschaftlichen Forschung, grants P11428PHY and P15348-PHY.
Lyapunov Instability and Collective Tangent Space Dynamics of Fluids
1175
References 1. Posch, H. A., and Hoover, Wm. G.: Equilibrium and nonequilibrium Lyapunov spectra for dense fluids and solids. Phys Rev. A 39 (1989) 2175–2188 2. Gaspard, P.: Chaos, Scattering, and Statistical Mechanics, (Cambridge University Press, 1998). 3. Hoover, Wm. G.: Computational Statistical Mechanics, (Elsevier, New York, 1999) 4. Dorfman, J.R.: An Introduction to Chaos in Nonequilibrium Statistical Mechanics, (Cambridge University Press, 1999) 5. Ruelle, D.: J. Stat. Phys. 95 (1999) 393 6. Benettin, G., Galgani, L., Giorgilli, A., and Strelcyn, J. M.: Meccanica 15 (1980) 9 7. Shimada, I., and Nagashima, T.: Proc. Theor. Phys. 61 (1979) 1605 8. Milanovi´c, Lj., and Posch, H. A.: Localized and delocalized modes in the tangentspace dynamcs of planar hard dumbbell fluids. J. Molec. Liquids (2002), in press 9. Dellago, Ch., Posch, P. H., and Hoover, Wm. G.: Phys. Rev. E 53 (1996) 1485 10.
Deterministic Computation Towards Indeterminism Bogdanov A.V., Gevorkyan A.S., Stankova E.N., Pavlova M.I. Institute for High-Performance Computing and Information Systems Fontanka emb. 6, 194291, St-Petersburg, Russia [email protected], [email protected], [email protected], [email protected]
Abstract. In the present work we propose some interpretation of the results of the direct simulation of quantum chaos.
1
Introduction
At the early stage of quantum mechanics development, Albert Einstein has written a work in which the question, which has become a focus of physicians attention several decades later, was touched upon. The question was: what will the classic chaotic system become in terms of quantum mechanics. He has particularly set apart the three-body system. In an effort to realize the problem and get closer to its chaos solution in essential quantum area, M. Gutzwiller, a well known physician, have conditionally subdivided all the existing knowledge in physics into three areas [1]: 1) elementary classical mechanics, which only allows for simple regular system behaviour (regular classical area R); 2) classical chaotic dynamic systems of Poincare systems (P area); 3) regular quantum mechanics, which interpretations are being considered during last 80 years (Q area). The above areas are connected by a separate bounds. Thus, Bor’s correspondence principal works between R and Q areas, transferring quantum mechanics into classical Newton’s mechanics within the limit h ¯ → 0. Q and P areas are connected by Kolmogorov’s - Arnold’s - Mozer’s theorem (KAM). Let’s note that KAM theorem allows to determine the excitations , which cause the chaotic behaviour of regular systems. Inspite of well known work by Wu and Parisi [2], which allows to describe Q-systems with the help of P -systems in thermodynamic limit under certain circumstances, the general principle connecting P and Q is not yet determined. Assuming the existence of the fourth area - quantum chaos area Qch , M. Gutzwiller adds that it rather serves for the puzzle description than for a good problem formulation. It is evident that the task formulated correctly in Qch area is a most general one and must transfer to the abovementioned areas in its limits. The problem of quantum chaos was studied as the example of quantum multichannel scattering in collinear three-body system [3,4]. It was shown than this task can be transformed into a problem of unharmonic oscillator with non-trivial P.M.A. Sloot et al. (Eds.): ICCS 2002, LNCS 2331, pp. 1176−1183, 2002. Springer-Verlag Berlin Heidelberg 2002
Deterministic Computation towards Indeterminism
1177
time (internal time). Let’s note, that in a model considered internal time is determined by a system of two non-linear differential equations of the second order. In [5] this problem was studied on the example of chemical reaction and in [6] it was applied to surface scattering. The ab initio computation even of the simple three-body systems is a challenge for some generation of computational physicists, so some new approach was proposed in [7] and beautiful example of distributed computation was demonstrated in [8]. In the present work we propose some interpretation of the results, obtained in [7,8], and thus give our view of quantum chaos origination.
2
Formulation of the problem
The quantum multichannel scattering in the framework of collinear model is realized accordingly to follow scheme: A + (BC)m (AB) m+C A+B+C (1) A + (BC)n → A + (BC)m ∗ (ABC) → (AB)m + C A+B+C with m and n being the vibrational quantum numbers correspondingly in (in) and (out) scattering channels. As it was shown elsewhere [3,4] the problem of quantum evolution (1) can be strictly formulated as a motion of image point with a reduced mass µ0 over the manyfold M , that is stratificated Lagrange surface Sp , in the moving over Sp local coordinate system. In our case there is standard definition of the surface Sp Sp = x1 , x2 ; 2µ0 E − V x1 , x2 > 0 , µ0 =
mA mB mC mA + mB + mC
1/2
(2) ,
where mA , mB , mC being the masses of corresponding particles, E and V x1 , x2 being correspondingly the total energy and interaction potential of the system. The metric on the surface Sp in our case is introduced in the following way gik = P02 x1 , x2 δik , (3) P02 x1 , x2 = 2µ0 E − V x1 , x2 . As to the motion of the local coordinate system, it is determined by the projection of the image point motion over the extremal ray ext of the Lagrange manyfold Sp . Note, that for scattering problem (1) there are two extremal rays on a surface Sp : one is connecting the (in) channel with the (out) channel of particle
1178 A.V. Bogdanov et al.
rearrangement and the other is connecting the (in) channel with the (out) channel, where all three particles are free. ¿From now on we shall study only the case of rearrangement final channel. Let us introduce curvilinear coordinates x1 , x2 in Euclidean space R2 along the projection of the rearrangement extremal ray ¯ ext and x2 is changing in orthog¯ ext in a such way, that x1 is changing along onal direction. In such a case the trajectory of image point is determined by the following system of the second order differential equations: xk;ss + {}kij xi;s xj;s = 0
(i, j, k = 1, 2)
(4)
where {}kij = (1/2)g kl (glj;i + gil;j − gij;l ), gij;k = ∂xk gij . As to the law of local coordinate system motion, it is given by the solution x1 (s). Based on this solution the quantum evolution of the system on the manyfold M is determined by the equation (see [4]) 2 h ¯ ∆(x1 (s),x2 ) + P02 x1 (s) , x2 Ψ = 0,
(5)
with the operator ∆(x1 (s),x2 ) of the form
1 1 1 ∆(x1 (s),x2 ) = γ − 2 ∂x1 (s) γ ij γ 2 ∂x1 (s) + ∂x2 γ ij γ 2 ∂x2 .
(6)
As to the metric tensor of the manyfold M , it has the following form [4]: 2 λ x1 (s) 1+ , ρ1 (x1 (s))
γ11 = γ22 =
(7)
2
2
1+
γ12 = γ21 = 0,
x ρ2 (x1 (s))
,
γ = γ11 γ22 > 0,
with λ being de Broglie wave length on ext and ρ1 , ρ2 being the principle curvatures of the surface Sp in the point x1 ∈ in the directions of coordinates x1 , x2 changes P0 x1 (s) , 0 , ρ1 = P0;x1 (x1 (s) , 0) h ¯ , λ= 1 P0 (x (s) , 0)
P0 x1 (s) , 0 ρ2 = , P0;x2 (x1 (s) , 0)
P0;xi
dP x1 (s) , x2 = . dxi
(8)
Note, that the main difference of (5) from Schr¨ odinger equation comes from the fact, that one of the independent coordinates x1 (s) is the solution of nonlinear difference equations system and so is not a natural parameter of our system and can in certain situations be a chaotic function.
Deterministic Computation towards Indeterminism
3
1179
Reduction of the scattering problem to the problem of quantum harmonic oscillator with internal time
Let us make a coordinate transformations in Eq.(5): 1
−1 x(s) 1 √ τ = Eki P x , 0 γdx1 , 0
(9)
i − 12 1 2 P x ,0 x , z= h ¯ Ek with Eki being the kinetic energy of particle A in the (in) channel, the function 1 2 P x , x = 2µ0 Eki − V (x1 , x2 ) and with image point on the curve ext it is just the momentum of point. image 1 2 By expanding of P x , x over the coordinate x2 up to the second order we can reduce the scattering equation (5) to the problem of quantum harmonic oscillator with variable frequency in the external field, depending on internal time τ x1 , x2 . E.g. in the case of zero external field the exact wave function of the system without some constant phase, unimportant for the expression of the scattering matrix, is of the form
1/2
(Ωin /π) Ψ (n; τ ) = 2n n! |ξ| +
exp
12 ×
Ei τ
τ 1 dτ − n+ Ωin 2+ h ¯ 2 |ξ| k
−∞
√ Ωin 1 ˙ −1 2 1 −1 2 ˙ z , z − pp z Hn + ξξ 2 2 |ξ| ξ˙ = dτ ξ,
p˙ = dτ p,
(10)
p x1 (s) = P x1 (s) , 0 ,
with the function ξ (τ ) being the solution of the classical oscillator equation ξ¨ + Ω 2 (τ ) ξ = 0,
Ω 2 (τ ) = −
Eki
2 2 1 p;k p;kk + , + 2 ρ2 p p
2
p
k=1
p;k = with asymptotic conditions
dp dxk
(11)
1180 A.V. Bogdanov et al.
∼
ξ(τ ) −−−−−→ exp (iΩin τ ) , τ →−∞
∼
(12)
ξ(τ ) −−−−−→ C1 exp (iΩin τ ) − C2 exp (iΩout τ ) . τ →+∞
Note, that internal time τ is directly determined by the solution of x1 (s) and therefore includes all peculiarities if x1 behaviour. The transition probability for that case is of the form [3,4]: # $ $2 $$2 $ $2 $$ $ C2 $ $ $ C2 $ $ (n> −n< )/2 (n< )! Wmn = 1 − $$ $$ $P(n> +n< )/2 1 − $$ $$ $ , (13) $ (n> )! C1 $ C1 n being the Legandre polywhere n< = min (m, n), n> = max (m, n) and Pm nomial.
4
The study of the internal time dependence versus natural parameter of the problem - standard time
Now we are able to turn to the prove of the possibility of the quantum chaos initiation in the wave function (10) and as a result in the probability (13). It is enough to show, that the solution x1 (s) with certain initial conditions behaves unstable or chaotically. With that purpose on the example of elementary reaction ∗ Li + (F H) → (LiF H) → (LiF ) + H we studied carefully the behaviour of the image point trajectories on Lagrange surface Sp . It was shown that with collision energy Eki = 1.4eV and for fixed transversal vibrations energy Ev = 1.02eV the image point trajectory is stable. The whole region of kinetic energies is splited to regular subregions, and dependingly from which subregion trajectory starts it goes either to (out) channel (Fig.1(a)) or reflects back in the (in) channel (Fig.1(b)). With a further change of kinetic energy the image point trajectory in the interaction region starts orbiting, that corresponds to the creation of the res∗ onance complex (LiF H) , and after that leave the interaction region either to (out) (Fig.1(c)) or return to (in) channel. In such a case the image point trajectories diverge and this divergence is exponential, as can be seen from the study of the Lyapunov parameters. That is for those initial conditions the evolution in the correspondent classical problem is chaotic and so the motion of the local coordinate system is chaotic too. It is easy to see that in such situation the behaviour of x1 (s) is also chaotic and the same is true for internal time, that is the natural parameter of quantum evolution problem. It can be shown, that chaotic behaviour of the internal time is followed by the stochastic behaviour of the model equation solution ξ (τ (s)) and the same is true for the wave function (10) and transition probability (13). In such a way the possibility of violation of quantum determinism and quantum chaos organization was shown on the example of the wave function of the simple model of multichannel scattering.
Deterministic Computation towards Indeterminism
1181
Fig. 1. Dependence of Lyapunov exponent over time parameter s for the case of rearrangement reaction going through resonance state.
Those results may seem strange if we take into account, that original problem (i.e. Schr¨ odinger equation with asymptotic scattering conditions) was quite deterministic (i.e. was good candidate for possessing unique solution), outside of standard interpretation of quantum mechanical quantities. At the same time if one looks carefully at the final version of the scattering probabilities it is clear, that difference between stochastic and regular regimes are not of principal importance, actually the ansatz of solution in our approach for two cases is the same. The only difference comes from the fact, that when orbiting in interaction region starts, the initial conditions for outcoming bunch of trajectories become undetermined, that can be regarded in terms of fluctuations of the initial stratificated Lagrange surface Sp ,just as in the case of vacuum fluctuations in quantum field theory [5].
5
Conclusion
In this work it was shown that the representation developed by the authors includes not only Plank’s constant h ¯ , but new energetic parameter as well. Thus, when the energy of the particles collision exceeds certain critical value (which is different for the different systems), solution for internal time τ coincides with an ordinary time - natural t parameter. In this case, the movement equation for the system of bodies transforms to common nonstationary Schr¨ odinger’s equation. The scattering process is in fact a direct process for this case. But everything is quite different when the collision occurs below the critical energy specified. As it is shown, in such a case the solution for internal time τ in
1182 A.V. Bogdanov et al.
a definite range of t has an oscillational character. Moreover, for all the extreme points the derivative of τ by t has a jump of the first kind, while the phase portrait of reactive (parametric) oscillator has bifurcations. Let’s note that these are the collision modes with the strong interference effects, i.e. the problem becomes essentially multichannel and includes the phase of resonant state formation. At a small decrease of collisions energy, a number of internal time oscillations grows dramatically. In this case the system loses all the information about its initial state completely. Chaos arises in a wave function, which then self-organizes into a new order within the limit τ → ∞. Mathematically it becomes possible as a result of common wave equation irreversibility by time. Let’s stress that the above result supports the transitional complex theory, developed by Eyring and Polanyi on the basis of evristic considerations, the essence of which is statistical description of chemical reactions. The amplitude of regrouping transition in three-body system is investigated in the work on example of Li + (F H)n → (LiF )m + H reaction and it is shown, that in the area where the number of internal time peculiarities is high, it has an accidental value. It is also shown that the representation developed satisfies the limit transitions in the areas specified, including transition from Qch area into P area. The latter occurs under h ¯ → 0 and at Eki < Ec , where Ec is critical energy and Eki is a collision energy. It is possible to give very simple interpretation of the above results in terms of initial Lagrange surface fluctuations in strong interaction region.
References [1] M. C. Gutzwiller, Chaos in Classical and Quantum Mechanics, Springer, Berlin, 1990. [2] E. Nelson, Phys. Rev., (1966), v. 150, p. 1079. [3] A. V. Bogdanov, A. S. Gevorkyan, Three-body multichannel scattering as a model of irreversible quantum mechanics, Proceedings of the International Symposium on Nonlinear Theory and its Applications, Hilton Hawaiian Village, 1997, V.2, pp.693696. [4] A. V. Bogdanov, A. S. Gevorkyan, Multichannel Scattering Closed Tree-Body System as a Example of Irreversible Quantum Mechanics, Preprint IHPCDB-2, 1997, pp. 1-20. [5] A.V. Bogdanov, A.S. Gevorkyan, A.G. Grigoryan, Random Motion of Quantum Harmonic Oscillator. Thermodynamics of Nonrelativistic Vacuum, in Proceedings of Int. Conference ”Trends in Math Physics”, Tennessee, USA, October 14-17, 1998, pp.79-107. [6] A.V. Bogdanov, A.S. Gevorkyan, Theory of Quantum Reactive Harmonic Oscillator under Brownian Motion, in Proceedings of the International Workshop on Quantum Systems, Minsk, Belarus, June 3-7, 1996, pp.26-33. [7] A.V. Bogdanov, A.S. Gevorkyan, A.G. Grigoryan, S.A. Matveev, Internal Time Peculiarities as a Cause of Bifurcations Arising in Classical Trajectory Problem and Quantum Chaos Creation in Three-Body System, in Proceedings of Int. Symposium ”Synchronization, Pattern Formation, and Spatio-Temporal Chaos in Coupled Chaotic Oscillators” Santyago de Compostela, Spain, June 7-10, 1998;
Deterministic Computation towards Indeterminism
1183
[8] A.V. Bogdanov, A.S. Gevorkyan, A.G. Grigoryan, First principle calculations of quantum chaos in framework of random quantum reactive harmonic oscillator theory, in Proceedings of 6th Int. Conference on High Performance Computing and Networking Europe (HPCN Europe ’98), Amsterdam, The Netherlands, April, 1998
Splitting Phenomena in Wa ve Packet Propagation I. A. Valuev1 and B. Esser2 Moscow Institute of Physics and Technology, Department of Molecular and Biological Physics, 141700 Dolgoprudny, Russia 1
2
[email protected]
Institut fur Physik, Humboldt-Universitat Berlin, Invalidenstrasse 110, 10115 Berlin, Germany [email protected] http://www.physik.hu-berlin.de
Abstract. Wave pac ket dynamics oncoupled potentials is considered on the basis of an associated Spin-Boson Hamiltonian. A large number of eigenstates of this Hamiltonian is obtained by numerical diagonalization. Eigenstates display a mixing of adiabatic branc hesas is evident from their Husimi (quantum density) projections. From the eigenstates time dependent Husimi projections are constructed and packet dynamics is investigated. Complex packet dynamics is observed with packet propagation along classical trajectories and an increasing number of pac kets due to splitting events in the region of avoided crossings of these trajectories. Splitting events and their spin dependencies are systematically studied. In particular splitting ratios relating the intensities of packets after a splitting event are derived from the numerical data of pac ket propagation. A new projection technique applied to the state vector is proposed by which the presence of particular packets in the evolution of the system can be established and made accessible to analysis. 1
Introduction and Model
Wave packet dynamics is one of the central topics in quantum evolution with a wide range of applications ranging from from atomic and molecular physics to ph ysical c hemistry (see e.g. [1] and references therein). W e present a numerical investigation of the dynamics of wave packets in a many-potential system, when phase space orbits associated with dierent adiabatic potentials are coupled. Basic to our investigation is the ev olution of quantum states described by the Spin-Boson Hamiltonian
r
^ = + ^1 1 ^x + 1 (P^ 2 + r2 Q^ 2 )^1 + ( p rQ^ + )^z : H (1) 2 2 2 In (1) a quantum particle residing in tw o states is coupled to a vibrational en vironment speci ed by the coordinate Q. The tw ostate quantum system is P.M.A. Sloot et al. (Eds.): ICCS 2002, LNCS 2331, pp. 1184−1192, 2002. Springer-Verlag Berlin Heidelberg 2002
Splitting Phenomena in Wave Packet Propagation 1185
represented by the standard Pauli Spin Matrices ^ (i = x; z ), r is the dimensionless vibrational frequency of the oscillator potential and p is the coupling constant between the dynamics of the particle and the vibrational environment. We note that (1) is a generalized Spin-Boson Hamiltonian containing the parameter , which destroys the parity symmetry of the eigenstates of the more conventional Spin-Boson Hamiltonian in which such a term is absent. A Hamiltonians like (1) can be obtained from dierent physical situations, the particle being e.g. an electron, exciton or any other quasiparticle. To be de nite we will use a \molecular" language and consider the situation when the Hamiltonian (1) is derived from a molecular physics model. We consider a molecular dimer, in which the transfer of an excitation between two monomers constituting the dimer and coupled to a molecular vibration is investigated. Then is the dierence between the energies of the excitation at the two nonequivalent monomers constituting the dimer ( + in (1) is the mean excitation energy, in what follows we omit this term thereby shifting the origin of the energy scale to + ). For the details of the derivation of (1) from a molecular dimer Hamiltonian and in particular the connection between the dimensionless parameters of (1) with a dimer model we refer to [2] and references therein. i
E +
Uad l0
20
l1
r0
r1
10
−
Uad
Q −75
−25
25
75
Adiabtic potentials for the parameter set used. The line of constant energy E = 15 and its crossing points with the potentials (turning points) are shown. The turning points are labeled according to the location of the point (l for "left" and r for "right") and monomer (spin) indicies 0 and 1 correspond to the upper and lower monomer, respectively Fig. 1.
Applying a Born - Oppenheimer stepwise quantization to (1) one obtains the Hamiltonians of the adiabatic reference systems 1 H (Q) = P 2 + Uad (Q): (2) 2
1186 I.A. Valuev and B. Esser
with the adiabatic potentials (Q) = r2 Q2 Uad
2
s
r
2
1 + + p rQ 4 2
;
(3)
In Fig. 1 the adiabatic potentials for a given parameter set are shown. Fixing an energy E phase space orbits can be derived from (2) for each of the adiabatic potentials. Phase space trajectories of isolated monomers at a given energy E can be derived in an analogous way by neglecting the quantum transfer in (1) (discarding ^x ). In what follows we denote the upper and lower monomer of the dimer con guration by the indices (0) and (1), respectively. In spin representation projections of the state vector on such monomer states correspond to projections on the spin up state (upper monomer) and spin down state (lower monomer), respectively. We note that in the semiclassical case adiabatic trajectories can be represented as built from pieces of monomer trajectories. In the analysis of packet propagation derived from the state vector it will be convenient to use projections on such monomer states below. A quantum density in phase space is constructed by using Husimi projections extended by spin variables. For a given state vector j i such densities are derived by projecting j i on a combined basis of standard coherent oscillator states j(Q; P )i, which scan the phase space plane Q; P , multiplied by spin states jsi, jsi = c" j"i + c# j#i via h ((Q; P ); s) = jh j (Q; P ); sij2 :
(4)
Husimi densities are equivalent to Gaussian smoothed Wigner distributions, positive de nite and useful in phase space analysis of quantum states [3]. Here we will use Husimi densities to analyze wave packet dynamics.
2 Numerical procedure and phase space density A large number of eigenstates and eigenvalues of (1) was obtained by a direct matrix diagonalization method for dierent parameters p, r and using the ARPACK package [4]. For the matrix diagonalization a basis of the harmonic oscillator eigenstates extended by the spin variable was used. Here we report results for the particular parameter set p = 20, r = 0:1 and = 10, for which a diagonalization of a matrix of dimension N = 4000 was applied. From this diagonalization the rst 1100 eigenvalues and eigenvectors were used in constructing the statevectors. The Husimi density of a representative eigenstate computed from an eigenvector using (4) is shown in the Fig. 2, where the classical phase space orbits corresponding to the adiabatic potentials at the energy of the eigenvalue of the selected eigenstate are included. From Fig. 2 it is seen that the eigenstate density is located on both of the adiabatic branches, i.e. adiabatic branches are mixed in the eigenstates of (1).
Splitting Phenomena in Wave Packet Propagation 1187
P
7
−7 −50
90
Q
Husimi distribution of the eigenstate number 184. The quantum phase space density is located on both of the adiabatic branches, the corresponding classical phase space orbits of which are shown as lines
Fig. 2.
A detailed analysis of sequences of such eigenstates [2], shows that the components of this density, located on a given adiabatic branch change from one eigenstate to another in a rather irregular fashion. This mixing of adiabatic branches in the spectrum of (1), which can be shown by dierent methods, such as e.g. Bloch projections [5], can be viewed as appearance of spectral randomness and is well known as incipience of quantum chaos [6], when the random features of the spectrum just appear, but regular parts of the spectrum are still intact. Quantum and classical consequences of this behaviour of the Spin-Boson Model have been intensively investigated over the last years [7], [8], [9]. Here we address the dynamical consequences of the mixing of adiabatic branches of the spectrum of (1) for the particular phenomenon of wave packet splitting. As a result of this mixing splitting events of wave packets initially prepared on one adiabatic branch will occur and packets propagating on dierent branches appear. This can be observed by using Husimi projections constructed from a time dependent state vector j (t)i in (4). 3
Splitting analysis of packet propagation
We investigated packet dynamics by propagating numerically test wave packets, which can be constructed initially at the arbitrary positions (Q0 ; P0 ) in phase space as coherent states multiplied by the initial spin j (0)i = j(P0 ; Q0 )ijs0 i. Then time propagation of the state vector j (t)i corresponding to the initial condition was performed by using the numerically obtained eigenstates and eigen-
1188 I.A. Valuev and B. Esser
values. Packet propagation was analyzed in detail by means of Husimi projections (4). In the Fig. 3 a snapshot of such a packet propagation for an initial packet placed at the left turning point of the upper monomer potential with an energy E = 12 is shown. The snapshot is taken at the moment t = 1:1(2=r), i. e. for t = 70, below the time unit 2=r (free oscillator period) is everywhere indicated.
P
9
−9 −70
110
Q Fig. 3. Snapshot of wave packet propagation at a time t = 1:1(2=r ) for an initial wave packet placed at the left turning point of the left monomer (0), energy E=12. For comparison the adiabatic phase space trajectories at the same energy are shown. In the left lower part a splitting event is observed. In the Husimi density the projection spin is equal on both monomers
We observed splitting phenomena of propagated wave packets at each of the crossing points of the monomer phase space trajectories. The region of such crossing points of monomer phase space trajectories is equivalent to the avoided crossings of the adiabatic trajectories shown in the Fig. 3 (in what follows for shortness we will refer to this phase space region simply as monomer crossings). In the Fig. 3 a splitting event is visible in the left lower part of the phase space near such a crossing. The intensity of the propagated and split wave packets was considered in dependence both on the energy E and the spin projection. Packets with spin projections corresponding to the phase space trajectory of the monomer on which propagation occurs turned out to be much more intensive than packets with opposite spin projections (for the parameter set used the intensity ratio was approximately three orders of magnitude). We call the much more intensive packets, for which spin projection corresponds to the monomer phase space trajectory, main packets and the other packets "shadow" packets.
Splitting Phenomena in Wave Packet Propagation 1189
When a main packet reaches a crossing point it splits into two main packets, one main packet propagating on the same monomer phase space trajectory as before, and the other main packet appearing on the trajectory of the other monomer. Both packets then propagate each on their own monomer trajectory with approximately constant Husimi intensities until they reach the next crossing point. Then the packets split again, etc. The result of several splittings of this kind is seen from Fig. 3. Splitting events can be classi ed as primary, secondary and so on in dependence of their time order. For a selected initial condition splitting events can be correspondingly ordered and classi ed into a splitting graph. We present such a splitting graph in the Fig. 6(a), where as a starting point the left turning point of the lower monomer potential and the energy E = 15 were used. In order to minimize the amount of data to be analyzed for packet propagation and splitting events we developed a turnpoint analysis and a numerical program. This program monitors the Husimi intensities of all of the packets resulting from splitting events, when they cross any of the turning points in dependence on their spin projections. The initial packet was also placed at a turning point of a monomer phase space trajectory. 2
Cs
1.5
1
0.5
0
0
10
20
30
40
50
E Fig. 4. Splitting coecient Cs measured as the ratio of the Husimi projections of the packets passing the corresponding turning points after splitting (see text)
First of all we investigated the Husimi intensity for the primary splittings by considering the four turning points as initial conditions for such splitting processes. According to our turnpoint analysis procedure these primary splittings can be classi ed as follows:
1190 I.A. Valuev and B. Esser
fl ; ug ) fr ; dg, fr ; ug fr ; ug ) fl ; dg, fl ; ug fl ; dg ) fr ; ug, fr ; dg fr ; dg ) fl ; ug, fl ; dg 0
1
0
0 1
1 0
0 1
1
0
1
Here on the left sides of the arrows the positions of the initial packets and on the right sides of the two nal packets at their turning points are indicated as 0 for the upper and 1 for the lower monomer, respectively. The letters l, r denote the left, right turning points (see Fig. 1), and spin indices u, d the up and down projections. For shortness here the main packets are considered, when all the projection spins correspond to the turning points of "their" monomer trajectory. In the turnpoint analysis the energy was changed over a broad interval in which well de ned packets are present and the Husimi intensities measured. The analysis of the obtained data showed that the ratio C of the intensity of the packet that appears on the other monomer trajectory to the intensity of the packet that remains on the initial monomer trajectory after a splitting is constant for all primary splitting con gurations and is a function of the initial packet energy only (Fig. 4). All the packets observed in the propagation are due to complex interference inside the state vector (t) of the system. In order to investigate this complex behaviour we introduced special projection states with which it is possible to analyze the process of appearance of packets. Such projection states can be introduced by an arti cially constructed state vector s
jM (t)i = X a j(Q (t); P (t))ijs i; i
i
i
i
(5)
i
which is a superposition of coherent states modelling all packets at a given time t. The packets are indexed by i and contribute to jM (t)i with their coecients a and spin js i that corresponds to the monomer trajectory the packet is propagating on. The coecients a can be derived from the splitting data of the turnpoint analysis. The phase space positions (Q (t); P (t)) are chosen according to the semiclassical motion of the packet centers. We note that this construction provides only the amplitudes a (all a are assumed to be real), because information about the phases cannot be extracted from the Husimi densities. For an initial packet in jM (t)i chosen to be the same as for the exact quantum propagation, it is possible to investigate the correspondence between the reference states jM (t)i and the exact state vector j (t)i. A comparison of the correlation functions h (0)j (t)i and hM (0)jM (t)i shows very similar reccurence features (Fig. 5). The intensities of the reccurence peaks for the exact and model wavefunctions are in good agreement at the early stage of propagation. The reccurence times are in agreement even for longer propagation times, when a lot of packets already exist in the splitting model based on (5) (1584 packets for t = 5(2=r) in Fig. 5). i
i
i
i
i
i
i
Splitting Phenomena in Wave Packet Propagation 1191
|<ψ(0)|ψ(t)>|
1
(a)
0.5
0
|<M(0)|M(t)>|
1
(b)
0.5
0
0
1
2
3 t, 2π/r
4
5
Fig. 5. The correlation functions for the initial packet located at the turning point l1 with initial spin j #i and E = 15: (a) { from numerical propagation and (b) { from the splitting model. Time is measured in periods of the osillator associated with monomers
(a)
1 P(t)
0.5
(b)
0
1 |Ca(t)|
(c)
0.5 0
0
1
2 t, 2π/r
Fig. 6. The splitting dynamics of the state initially located at the turning point l1 with initial spin j #i and E = 15. (a) Splitting event graph. The branchings correspond to splittings of the packets, the packets which change and do not change the monomer trajectory are displayed by the lines going down and up, respectively. (b) Correlation of the numerically propagated wavefunction and the normalized splitting model wavefunction. (c) Correlation of the numerically propagated wavefunction and the packet, classically moving along the lower adiabatic potential
1192 I.A. Valuev and B. Esser
For direct comparison of jM (t)i and j (t)i we use the sum of projections of all packets existing in jM (t)i on j (t)i: P (t) =
Xa
i jh
(t)j(Qi (t); Pi (t))ijsi ij:
(6)
i
From the Fig. 6(b), where P (t) is presented it is seen that jM (t)i is a good approximation to the state vector j (t)i. This shows that this projection technique oers a possibility to analyze the exact state vector. The projection of an individual reference packet moving classically along some phase space trajectory, for example the trajectory of an adiabatic potential, on j (t)i can be used to nd out to what extent this packet is contained in the time evolution. The projection of this type, Ca (t) = hMa (t)j (t)i, where jMa (t)i is the model wavefunction constructed from a packet of constant intensity moving along the lower adiabatic potential without splittings, is shown in Fig. 6(c). The initial state for both the exact state vector and the reference state jMa (t)i is the same and located in the turning point l0 . The absolute value of Ca (t) decays stepwise as the splittings in j (t)i occur. We conclude that the construction of reference states jM (t)i captures essential features of wave packet propagation and splitting displayed by the exact state vector (t) and therefore can be used for wave packet modelling and projection techniques. Following this idea we can make the birth process of packets in the splitting region accessible to direct investigation by projecting the exact state vector on such reference states. Using particular spin states for j(Qi (t); Pi (t))ijsi i in projections, it should be possible to project out the birth processes of packets in the state vector j (t)i in the splitting region.
Acknowledgements Financial support from the Deutsche Forschungsgemeinschaft (DFG) is gratefully acknowledged.
References 1. 2. 3. 4.
J. E. Bay eld, Quantum Evolution, John Wiley and Sons Inc., New York 1999. C. Koch and B. Esser, Phys. Rev. A 61, 22508 (2000). K. Takahashi, Progr. Theor. Phys. Suppl. 98, 109 (1989). R. B. Lehoucq, D. C. Sorensen and C. Yang, Arpack Users Guide: Solution of Large Scale Eigenvalue problems, http://www.caam.rice.edu/software/ARPACK 5. H. Schanz and B. Esser, Z. Phys. B 101, 299 (1996). 6. M. Cibils, Y. Cuche, and G. Muller, Z. Phys. B 97, 565 (1995). 7. L. Muller, J. Stolze, H. Leschke, and P. Nagel, Phys. Rev. A 55, 1022 (1991). 8. H. Schanz and B. Esser, Phys. Rev. A 44, 3375 (1997). 9. R. Steib, J. L. Schoendor, H. J. Korsch, and P. Reineker, Phys. Rev. E 57, 6534 (1998).
An Automated System for Prediction of Icing on the Road Konstantin Korotenko P.P. Shirshov Institute of Oceanology 36 Nakhimovsky pr. Moscow, 117851, Russia http ://www.aha.rul-koroten [email protected]
Abstract. During the period from late autumn to early spring, vast areas in North America, Western Europe, and many other countries experience frequent snow, sleet, ice, and frost. Such adverse weather conditions lead to dangerous driving conditions with consequential effects on road transportation in these areas. A numerical forecasting system is developed for automatic prediction of slippery road conditions at road station sites in northern Europe and North America. The system is based on a road conditions model forced by input from an operational atmospheric limited area model. Synoptic information on cloud cover and observations of temperature, humidity, water and ice on the road from the road station sites are taken into account in a sophisticated initialization procedure involving flux corrections for individual locations. The system is run initially at the Rhode Island University with promising results. Currently, new forecasts 3 h ahead are produced every 20 minutes for 14 road station sites.
1 Introduction An accurate prediction of meteorological parameters such as precipitation, temperature, and humidity close to the ground is of great importance for various applications. For example, warnings about slippery road conditions may be issued if snow, freezing rain, or rime can be forecast with sufficient accuracy. An addition, traffic delays and the risk of accidents may be significantly reduced by specific actions such as road salting. An impressive amount of money is spent on winter road maintenance in many European countries. For example, it is estimated that the total budget for winter road maintenance in the United Kingdom is about $200 million every year. For Denmark, being a smaller country, the corresponding budget is about half of this. The variability from year to year, however, is considerable. Unnecessary road salting should be avoided for economic reasons and due to the risk of environmental damage. This means that optimal salting procedures should be sought. In this context, accurate road weather information is vital, which justifies the efforts that are spent on the development of advanced road conditions models. The present paper concerns the development of a numerical model system for operational forecasting of the road conditions in Rhode Island, USA. The prediction of the road conditions requires the production of accurate forecasts of temperature, humidity, and precipitation at the P.M.A. Sloot et al. (Eds.): ICCS 2002, LNCS 2331, pp. 1193−1200, 2002. Springer-Verlag Berlin Heidelberg 2002
1194 K. Korotenko
duction of accurate forecasts of temperature, humidity, and precipitation at the road surface. To provide this information, two strategies are possible. The first one relies on the manual work of a forecaster who issues warnings of slippery road conditions based on various meteorological tools, for example, synoptic observations and output from atmospheric models. The second possibility is based on automatic output from specialized models, which may involve statistical or deterministic methods. The two approaches can be used in combination if a forecaster supplies certain input data for the automatic system
2
System Basic Formulae and Boundary Conditions
The system is primarily based on earlier models developed by Sass [GI and Baker and Davies [I]. Additional work by Unsworth and Monteith [8], Strub and Powell [7], Louis [4], Jacobs [2] and Manton [5] provided information needed in the parameterization of the atmospheric heat flux terms. The resulting second order diffusion equation, with empirically parameterized flux terrns, is solved by a standard forward in time, centered in space finite difference scheme. The model predicts the continuous vertical temperature profile from the road surface to depths of about two meters in the roadbed. The model also allows predictions of road icing conditions. Atmospheric data, necessary as input to the model, can be supplied either from observations or from a weather forecast model.
2.1 Ground Heat Flux The model is constructed on an unsteady onedimensional heat conduction equation, that is,
where T&. 0 is the temperature at time t and depth z. It is assumed that the road surface and underlying sublayers are horizontally homogeneous so that heat Transfer in horizontal direction can be neglected- The model considers a vertical pillar with unit cross-sectional area, extending to the depth (usually 2 m) deep enough to eliminate the diurnal oscillation of temperature. The equation is solved with finite-difference method [3], along with an initial temperature profile within the road sublayer and upper and lower boundary conditions. The initial condition prescribes the temperature at every grid point in the road sublayers at the beginning of forecast. The lower boundary condition is straightforward and treated as a constant of mean winter soil temperature at two meters. The upper boundary condition is complicated and is expressed by an energy balance equation.
An Automated System for Prediction of Icing on the Road 1195
The gnd spacing is irregular with thin layers close to the surface. The temperature at the bottom layer is determined by a climatological estimate depending on the time of year. The values of the heat conductivity , density Po, and specific heat capacity C, are constant (see Apendix).
2.2 Solar Heat Flux The solar axid infrared radiation is computed from the radiation scheme used in the at the ground is comatmospheric HIRLAM model [6].The net solar flux density @fi puted according to (2) as a linear combination of a clear air part
gfiaand a cloudy
where a is a surface albedo for solar radiation and CMis a total cloud cover determined from a maximum overlap assumption. The clear air term is parameterized in (3):
where S is the solar constant and poo is a reference pressure. The fust tern in (2) depending on the zenith angle concerns the stratospheric absorption due to ozone. A major contribution to the extinction of solar radiation comes from tropospheric absorption due to water vapor, C02, and 03.This is parameterized according to the second term in (3). Here, U ,is the vertically integrated water vapor path throughout the atmosphere. It is linearly scaled by pressure and divided by cos8. The last term involving two contributions describes the effect of scattering. The first contribution arises from scattering of the incoming solar beam, while the second one is a compensating effect due to reflected radiation, which is backscattered from the atmosphere above. The coefficients a6 and a7 (see appendix) that are larger than 1 represent a crude inclusion of effects due to aerosol absorption and scattering, respectively. of (2) is given by (4): The cloudy contribution
gfiC
elwr
In (41, is the solar flux density at the top of the uppermost cloud layer. It is given by a forrnula correspondmg to (3), and the surface albedo appearing in the back-
1196 K. Korotenko
scattermg term of (3) is replaced by an albedo representing the cloudy atmosphere below. The transmittance of flux density from top to bottom of the cloudy atmosphere is described by
f ( p H , p , ) In . (4), the denominator takes into account multiple re-
flections between ground and cloud. The constant ay accounts for absorption in reflected beams.
2.3 Longwave Radiative Heat Flux The outcome of the radiation computations is the net radiation flux (bR, whlch can be partitioned into shortwave and longwave contributions:
In (5), a is the road-surface shortwave albedo, F is a scaled, dimensionless icesnow height, and H is ice-snow height (m) of equivalent water. There is no distinction between ice and snow. In addition, He is a critical value for ice-snow height, 6 is the solar zenith angle, and Rs is the total downward solar flux consisting of direct and difise radiation. Effects of absorption by water vapor and clouds are included, as is scattering of radiation by clear air and clouds. The value RL is the infrared longwave flux reaching the road surface, whch absorbs only a fraction Es = 0.90. It consists of longwave radiation from clear air and clouds, which have an ernissivity less than 1, depending on cloud type and thickness. The ernissivity of clouds in the lower part of the atmosphere is, however, close to 1, provided that the clouds are sufficiently thick (-200 m). The upward emission of longwave radiation is & b T: ,where 0 is the Stefan-Boltzmannconstant, 0 = 5.7* 10' ~ mK4,- and~ T, is the road surface temperature (0. According to observations [6], the albedo a,(6) for natural surfaces increases for large solar zenith angles, but it is almost constant equal to 03, as expressed in (1). For most surfaces, 0 <
a.
=
0.60 for cos(0) >
a. < 0.30, and for asphalt roads it is
reasonable to assume that a, = 0.10. For simplicity, the albedo for large zenith angles has been expressed as a linear function of COS(~). This involves the introduction
An Automated System for Prediction of Icing on the Road 1197
of an additional constant C 2 . The value of C 2 , however, is not well known and should ideally be based on local measurements. The constant cl = 0.60 represents a common albedo for ice and snow, which may have an albedo in the range between 0.35 for slushy snow and 0.80 for fresh snow 161. Because of this simplification, the zenith-angle dependency of albedo is neglected in case of ice or snow. In order to prevent a discontinuous transition between no ice and ice conditions, a simple interpolation term involving the dimension less ice height F is added to a a, ( 6 ) to obtain the albedo
2.4
a(6,F ) .
Longwave Radiative Heat Flux
Traditional drag formulas as pven by (6) and (7) are used to describe the fluxes of sensible and latent heat:
where
where Z = 10 m, C, is the specific heat of moist air at constant pressure, p, is the air density, and 8, and 8,, are potential temperatures at the computation level Z and at the surface, respectively. Similarly, q, and q, are specific humidities at the same levels. The latter is determined by a surface wetness parameter W N c , where Wc= 0.5 kg m-2 (0 < WJW,< l), q,, ( T,) is the saturation specific humidity at the surface temperature T, , and L, is the specific latent heat of evaporation if T, > O°C, otherwise it is the specific heat of sublimation. Here. k is the von Karman constant.
3
Description of the Developed System
The system was developed with a use of Visual Basic 6.0, Compaq Digital Fortran 6.0, ArcView GIs, and Surfer 7.3. The system is currently operational, in rudimentary form. The development of the web-based interface was being coordinated with a similar effort h d e d by the EPA EMPACT program and led by the Narragansett Bay Commission. The system allows access to a base map, GIs data on primary and sec-
1198 K. Korotenko
ondary highways from RI GIs, and llnkages to a variety of supporting web sites. As an example Figure 1 shows the opening page of the web site. It displays a map of RI and shows the location of the RWIS observation site.
~ t 9el k d 78 ( Westerly )
B
Block Island
I$I?~
Fig. 1. Rhode Island State (USA) map. The Road Weather Information System (RIRWIS) sites are depicted by solid circles
An Automated System for Prediction of Icing on the Road 1199
Fa
D
6
12
18
24
Long W a n Radiation, WfnA2
0
6
12
Wind Hxd.d Clouds Type P A W
a
6
Ts
12 Water
18
24
Ice
24
S a w i h l r H l w , WfnA2
D
6
12
24
Latent Plux, W i d 2
Fig. 2. Heat balance terms and the road temperature predicted for the site 72.32W, 44.50N
As an example Figure 2 illustrate the heat balance terms and the road temperature calculated on January 25, 2001 at the location 72.32W, 44.50N. It is seen that changes of the total heat leads to the oscillation of the pavement temperature that leads, in turn, to the formation rime and slippery roads in the evening and night, and wet or dry roads at daytime.
Acknowledgements. Author wishes to thank M. Spaulding, C. Calagan and T. Opishinski for fruitful discussion and support of this work .
1200 K. Korotenko
References 1. Barker, H.W. Davies, J.A.: Formation of Ice on Roads beneath Bridges. Journal of Applied Meteorology, Vo1.29, (1990) 1180-1184. 2. Jacobson, M.Z., Fundamentals of Atmospheric Modeling. Cambridge University Press (1999) 656 p. 3. Korotenko, K.A.: Modeling Turbulent Transport of Matter in the Ocean Surface Layer. Oceanology, Vo1.32, (1992) 5-13. 4. Louis, J. F.: Parameteric Model of Vertical Eddy Fluxes in the Atmosphere. Boundary Layer Meteorology, Vol. 17, (1979) 187-202. 5. Manton, M. J.: A Bulk Model of the Well-Mixed Boundary Layer, Boundary-Layer Meteorology, Vo1.40, (1987) 165-178. 6. Saas, B. H.: A Numerical Forecasting System for the Prediction of Slippery Roads. Journal of Applied Meteorology, Vol. 36 (1996) 80 1-817. 7. Strub, P.T., Powell, T.M.: The Exchange Coefficients for Latent and Sensible Heat Flux over Lakes: Dependence Upon Atmospheric Stability. Boundary-Layer Meteorology, Vol. 40 (1987) 349-361. 8. Unsworth M.H., Monteith, J.L.: Long-Wave Radiation at the Ground. I. Angular Distribution of Incoming Radiation. Quart. J. R .Met. Soc., Vol. 10 1 (1975) 13-24.
Appendix: Model Coefficients Table 1. Model Coefficients
Coefficient a1 a2 a3 a4 as a6 a7 a8
a9 a10
Value 5.56*10-j 3.47*10-j 0.25 600 2. 78*1U5 1.20 1.25 0.80 20 40
Coefficient B1 B2 B3 B4 B5 B6 B7 B8 CG Do0
Value 35 3000 0.60 0.17 0.0082 0.0045 0.4343 2.5*1d 800 5 "1o4
Coefficient k
Ls
L, w s
a EO
AG
p~ CT
Value 0.40 2.83 *lo6 2.50 "1o6 0.5 0.10 0.90 2.0 2400 5.67*10-8
Neural Network Prediction of Short-Term Dynamics of Futures on Deutsche Mark, Libor and S&P500 Ludmila Dmitrieva1, Yuri Kuperin1,2 and Irina Soroka3 1
Department of Physics, Saint-Petersburg State University, Ulyanovskaya str. 1, 198094 Saint-Petersburg, Russia [email protected] 2 School of Management, Saint-Petersburg State University per.Dekabristov 16, 199155 Saint-Petersburg, Russia 3 Baltic Financial Agency, Nevsky pr. 140, 198000 Saint-Petersburg, Russia [email protected]
Abstract. The talk reports neural network modelling and its application to the prediction of short-term financial dynamics in three sectors of financial market: currency, monetary and capital. The methods of nonlinear dynamics, multifractal analysis and wavelets have been used for preprocessing of data in order to optimise the learning procedure and architecture of the neural network. The results presented here show that in all sectors of market mentioned above the useful prediction can be made for out-of-sample data. This is confirmed by statistical estimations of the prediction quality.
1 Introduction In this talk we consider dynamic processes in three sectors of the international financial markets - currency, monetary and capital. Novelty of an approach consists in the analysis of financial dynamics by neural network methods in a combination with the approaches advanced in econophysics [1]. The neural network approach to the analysis and forecasting of the financial time series used in the present talk is based on a paradigm of the complex systems theory and its applicability to the analysis of financial markets [2,3]. The approach we used is original and differs from approaches of other authors [4-7] in the following aspects. While choosing the architecture of a network and a stratagy of forecasting we carried out deep data preprocessing on the basis of methods of complex systems theory: fractal and multifractal analysis, wavelet-analysis, methods of nonlinear and chaotic dynamics [1,2,3]. In the present talk we do not describe stages and methods of this data preprocessing. However the preliminary analysis has allowed to optimize parameters of the neural network, to determine horizon of predictability and to carry out comparison of forecasting quality of different time series from various sectors of the financial market. P.M.A. Sloot et al. (Eds.): ICCS 2002, LNCS 2331, pp. 1201−1208, 2002. Springer-Verlag Berlin Heidelberg 2002
1202 L. Dmitrieva, Y. Kuperin, and I. Soroka
Specifically we studied dynamic processes in the currency, monetary, capital markets in the short-term periods, predicting daily changes of prices of the following financial assets: futures on a rate US dollar - DM (it is designated as DM), futures on the rate of interest LIBOR on eurodollars (ED), futures on American stock index Standard & Poor’s 500 (SP).
2 Method of Prediction and Network Architecture It should be noted that the success or failure of a neural network predictor depends strongly on the user definition of the architecture of the input and desired output. For prediction of data sets under consideration the neural network we had used two inputs. As the first input we used the daily returns expressed as follows: to 1,5% return there corresponded the value 1.5. As the second input we used the mean values for the last 5 days. The presence of noise in analyzed time series can degenerate the learning and generalisation ability of networks. It means that some smoothing of time series data is required. We used the simplest techniques for smoothing, i.e. 5days moving average shifted backwards for one day. Thus such neural network aims to predict smoothed daily return to next day. Among all possible configurations of neural nets we chose the recurrent network with hidden layer feedback into input layer known as the Elman-Jordan Network (see Fig.1).
Fig. 1. Architecture of the Elman-Jordan neural network used for prediction
Neural Network Prediction of Short-Term Dynamics
1203
To our opinion this is one of the most powerful recurrent network. This type of backpropagation network has been successfully used in predicting financial markets because recurrent networks can learn sequences, so they are powerful tools for time series data processing. They have the slight disadvantage of taking longer to train. A backpropagation network with standard connections responds to a given input pattern with exactly the same output pattern every time the input pattern is presented. A recurrent network may respond to the same input pattern differently at different times, depending upon the patterns that have been presented as inputs just previously. Thus, the sequence of the patterns is as important as the input pattern itself. Recurrent networks are trained in the same manner as standard backpropagation networks except that patterns must always be presented in the same order; random selection is not allowed. The one difference in structure is that there is one extra slab in the input layer that is connected to the hidden layer just like the other input slab. This extra slab holds the contents of one of the layers as it existed when the previous pattern was trained. In this way the network sees previous knowledge it had about previous inputs. This extra slab is sometimes called the network’s ’’long term’’ memory. The Elman-Jordan network has logistic f(x)=1/(1+exp(-x)) activation function of neurons in its hidden (recurrent) layer, and linear activation function of neurons in its output layer. This combination is special in that two-layer networks with these activation functions can approximate any function (with a finite number of discontinuities) with arbitrary accuracy. In the present research we changed the logistic activation function by the symmetric logistic function f(x)=(2/(1+exp(-x)))-1. This do not change the properties of the network in principle but allows to speed up the convergence of the algorithm for the given types of series. The only requirement is that the hidden layer must have enough neurons. More hidden neurons are needed as the function being fitted increases in complexity. In the network we used the hidden layer consisted of 100 neurons. One of the hard problems in building successful neural networks is knowing when to stop training. If one trains too little the network will not learn the patterns. If one trains too much, the network will learn the noise or memorise the training patterns and not generalise well with new patterns. We used calibration to optimise the network by applying the current network to an independent test set during training. Calibration finds the optimum network for the data in the test set which means that the network is able to generalise well and give good results on out-of-sample data.
3 Neural Network Prediction of Returns We divided each analysed time series into 3 subsets: training set, test set and production set. As the training set, i.e. the set on which the network was trained to give the correct predictions, we took the first 900 observations. The test set was used for preventing the overfitting of the network and for calibration and included the observations numbering from 901 up to 1100. The production set included observations, which “were not shown” to the network and started from the 1101-th observation up to the end of the series.
1204 L. Dmitrieva, Y. Kuperin, and I. Soroka
The results of predictions for training sets of all analysed financial assets are given in Fig1, Fig2, Fig.3.
Fig. 2. Results of neural network prediction of 5-days moving average of returns for Deutsche mark futures: actual versus predicted returns in percents (top figure), coincidence of signs of predicted and actual returns (middle figure), absolute value of the actual minus predicted returns in percents (bottom figure)
Neural Network Prediction of Short-Term Dynamics
1205
Fig. 3. Results of neural network prediction of 5-days moving average of returns for Eurodollar futures: actual versus predicted returns in percents (top figure), coincidence of signs of predicted and actual returns (middle figure), absolute value of the actual minus predicted returns in percents (bottom figure)
1206 L. Dmitrieva, Y. Kuperin, and I. Soroka
Fig. 4. Results of neural network prediction of 5-days moving average of returns for S&P500 futures: actual versus predicted returns in percents (top figure), coincidence of signs of predicted and actual returns (middle figure), absolute value of the actual minus predicted returns in percents (bottom figure)
Neural Network Prediction of Short-Term Dynamics
1207
The quality of prediction was estimated by the following parameters: • Training time and number of learning epochs - the quantities showing how long the network can improve its predictions to achieve the best results on the test set. By the learning epoch we mean the single presentation to the network of all samples from training set. These parameters can vary depending on the given learning rate and momentum. Leaning rate and momentum are established on the basis of desirable accuracy of the prediction. For the neural network we used both these parameters were equal to 0.003. • Coefficient Q compares the accuracy of the model to the accuracy of a trivial benchmark model or trivial predictor wherein the prediction is just the mean of all of the samples. A perfect fit would result in an Q value of 1, a very good fit near 1, and a very poor fit less than 0. If neural model predictions are worse than one could predict by just using the mean of sample case outputs, the coefficient Q value will be less than 0. • R-squared - the coefficient of determination which is a statistical indicator usually applied to regression analysis being the ratio of predicted values variation to the actual values variation. • Mean absolute error - this is the mean over all patterns of the absolute value of the actual minus predicted values. • Max absolute error - this is the maximum of actual values - predicted values of all patterns. • % of proper predictions of returns signs – this is the ratio of number of samples for which signs of predicted values coincide with signs of actual ones to the number of considered samples. For details the reader is referred to the linear statistical analysis literature [8]. The above characteristics of neural network prediction quality for the analysed series are given in table 1. The table consists of three blocks. The upper one gives characteristics of the network training. The middle one refers to the whole time series, which includes training, test and production sets. The bottom block describes only the results for production sets. The table shows that the best predictions are obtained for S&P500 futures, the worse predictions are obtained for the Eurodollar (ED) futures. Deutsche mark (DM) futures has intermediate predictions. This follows from values of coefficient Q for production set (see the bottom block of table 1) although it hardly can be seen by sight from Fig.3 and Fig.4. It should be noted that despite the approximately equal quality of learning (see values of coefficient Q in the middle block of table 1) the training time for S&P500 is five times bigger than that for DM and training time for DM futures is four times bigger than training time for ED. This obviously means that to find hidden regularities in S&P500 futures is noticeably complicated than in DM futures and all the more in ED futures. At the same time the best quality of prediction is obtained just for S&P500 and the worse for ED. All this points out that hidden regularities found by neural network in S&P500 preserve their character much more longer than that found in ED. In other words ED futures have more unsteady hidden regularities what results in the worse quality of predictions.
1208 L. Dmitrieva, Y. Kuperin, and I. Soroka
Table 1. Numerical characteristics of neural network predictions quality (N denotes the number of learning epochs, τ stands for training time, Nwhole denotes the number of samples in the whole time series while Nprod stands for the number of samples in production set, % of signs means the percent of proper predictions of returns signs)
Characteristics N τ (hours) Nwhole Q r–squared mean abs.er., % max abs.er., % % of signs Nprod Q r–squared mean error, % max.error, % % of signs
S&P500 futures
DM futures
30512 19 1173 0,7408 0,7431 0,182 2,172 86 77 0,8032 0,8062 0.217 0.799 88
ED futures 6779 4 1170 0,7594 0,7612 0,196 1,291 83 74 0,5897 0,6319 0,279 1,046 86
1873 1 1170 0,7436 0,7452 0,179 2,281 83 74 0,4517 0,5697 0,201 1,234 88
In summary one should mention that ultimate goal of any financial forecasting is profitability. The latter is always connected with some trading rule and/or the money management strategy. This problem is out the scope of the present talk (see, however our recent paper [9]).
References 1. Mantegna, R.N., Stenley, H.E:. An Introduction to Econophysics. Correlations and Complexity in Finance. Cambridge University Press (2000) 2. LeBaron, B.: Chaos and Nonlinear Forecastability in Economics and Finance. Philosophical Transactions of the Royal Society of London 348 (1994) 397-404 3. Peters E.E.: Chaos and Order in Capital Market. John Wiley&Sons (1996) 4. Baestaens, D.E., Den Bergh, W.-M.Van, Wood, D: Neural network solutions for trading in financial markets. Pitman Publishing (1994) 5. Refenes, A.-P. (ed.): Neural Networks in the Capital Markets. John Wiley&Sons (1995) 6. Poddig, Th.: Developing Forecasting Models for Integrated Financial Markets using Artificial Neural Networks. Neural Network World 1 (1998) 65 – 80 7. Poddig, Th., Rehkugler, H. A.: World Model of Integrated Financial Markets using Artificial Neural Networks. Journal of Neurocomputing 10 (1996) 251-273 8. Dougherty, Ch.: Introduction to Econometrics. Oxford University Press (1992) 9. Kuperin, Yu.A., Dmitrieva L.A. and Soroka, I.V.: Neural Networks in Financial Market Dynamics Studies. Working paper series 2001-12, Center for Management and Institutional Studies, St.Petersburg State University, St.Petersburg (2001) 1-22
№
Entropies and Predictability of Nonlinear Processes and Time Series Werner Ebeling Saratov State University, Faculty of Physics, Saratov, Russia, werner-ebeling0web. de, home page: www .ebelinge .de
Abstract. We analyze complex model processes and time series with respect to their predictability. The basic idea is that the detection of local order and of intermediate or long-range correlations is the main chance to make predictions about complex processes. The main methods used here are discretization, Zipf analysis and Shannon's conditional entropies. The higher order conditional Shannon entropies and local conditional entropies are calculated for model processes (Fibonacci, Feigenbaum) and for time series (Dow Jones). The results are used for the identification of local maxima of predictability.
1
Introduction
Our everyday experience with the prediction of complex processes is showing us that predictions may be done only with certain probability. Based on our knowledge on the present state and on certain history of the process we make predictions, sometimes we succeed and in other cases the predictions are wrong [I]. Considering a mechanical process, we need only some knowledge about the initial state. The character of the dynamics, regular or chaotic, and the precision of the measurement of the initial states decide about the horizon of predictability. For most complex systems, say e.g. meteorological or financial processes, we have at best a few general ideas about their predictability. The problem we would like to discuss here is, in which cases our chances to predict future states are good and in which cases they are rather bad. Our basic tool to analyze these questions are the conditional entropies introduced by Shannon and used by many workers [2-61. By using the methods of symbolic dynamics any trajectory of a dynamic system is first mapped to a string of letters on certain alphabet [2,4,5]. This string of letters is analyzed then by Shannon's information-theoretical methods.
2
Conditional Entropies
This section is devoted to the introduction of several basic terms stemming from information theory which were mostly used already by Shannon. Let us assume that the processes to be studied are mapped to trajectories on discrete P.M.A. Sloot et al. (Eds.): ICCS 2002, LNCS 2331, pp. 1209−1217, 2002. Springer-Verlag Berlin Heidelberg 2002
1210 W. Ebeling
state spaces (sequences of letters) with the total length L. Let X be the length of the alphabet. Further let A1A2 . :. An be the letters of a given subtrajectory of . A,) be the probability to find in the total length n L. Let further p ( " ) ( .~. ~ trajectory a block (subtrajectory) with the letters A1 . . .A,. Then according to Shannon the entropy per block of length n is:
<
<
From this we derive conditional entropies as h, = (H,+l - H,) log(X) The limit of the dynamic n-gram entropies for large n is the entropy of the source h (called also dynamic entropy or Kolmogorov - Sinai entropy). Further we define
as the average predictability of the state following after a measured n-trajectory. We remember that log (A) is the maximum of the uncertainty, so the predictability is defined as the difference between the maximal and the actual uncertainty. In other words, predictability is the information we get by exploration of the next state in the future in comparision to the available knowledge. In the following we shall use X as the unit of the logarithms. The predictability of processes is closely connected with the dynamic entropies [7]. Let us consider now certain section of length n of the trajectory, a time series, or another sequence of symbols A1 . . .A,, which often is denoted as a subcylinder. We are interested in the uncertainty of the predictions of the state following after this particular subtrajectory of length n. Following again the concepts of Shannon we define the expression
as the conditional uncertainty of the next state (1 step into the future) following behind the measured trajectory A1 . . .A, .Further we define T;)(A~. . .A,) = I - h p ) ( .~. .A,) ~
(4)
as the predictability of the next state following after a measured subtrajectory, which is a quantity between zero and one. We note that the average of the local uncertainty leads us back to Shannon's conditional entropy h, . The predictability may be improved by taking into account longer blocks. In other words, one can gain advantage for predictions by basing the predictions not only on actual states but on whole trajectory blocks which represent the actual state and its history.
3
The conditonal entropy for model processes and time series
The first mathematical model of a nonlinear process was formulated in 1202 by the Italian mathematician Leonardo d a Pisa, better known as Fibonacci, in his
Entropies and Predictability of Nonlinear Processes and Time Series
1211
book Liber Abaci. Fibonacci considered the problem how many rabbit pairs are generated after n breeding sessions assuming the following simple rules: - the game starts with an immature pair, - rabbits mature in one season after birth, - mature rabbit pairs produce one new pair every breeding session, - rabbits never die. This game generates the famous sequence of Fibonacci numbers 1, 1, 2, 3, 5, 8, 13, 21, 34, 55, ...... The Fibonacci model may be encoded as a sequence of zeroths and ones by using the rules 0 + 1 denoting "young rabbits grow old" and 1 -+ 10 standing for "old rabbits stay old and beget young ones". Beginning with a single 0, continued iteration gives 1, 10, 101, 10110, etc., resulting finally in the infinite selfsimilar Fibonacci sequence 1011010110110 ... Alternatively we may formulate the rules by a grammar:
The conditional entropy of the Fibonacci sequence is exactly known [8]. These entropies behave in the limit of large n as
Another well-studied simple model of a nonlinear process is the logistic map:
In order to generate a discrete string from this map we use the bipartition (A = 2)
This way the states are mapped on the symbols 0 and 1 and the process is mapped on binary strings. We denote these sequences as Feigenbaum sequences. The rank-ordered word distributions for Feigenbaum strings were discussed by several authors [2,9,11]. For r = 4 all the words of fixed length are equally distributed and the entropy is maximal h, = 1. For the Feigenbaum accumulation point r, = 3.5699456... we get also a simple word distributions consisting only of one or two steps in dependence on the word length [2,9,11]. The construction rules for the generation of these sequences generate selfsimilar structures. Accordingly the n-gram block entropies satisfy the relations
By specialization of eqs.(l2) we get for n = 2* a result first obtained in 1986 by Grassberger [7]: Hn = log2(3n/2) P3)
1212 W. Ebeling
In a similar way we obtain the H , for all the other sequences [2,9]. For the conditional entropies we get for the Grassberger numbers n = 2,4,8,16,... the conditional entropies In between two Grassberger numbers namely at n = 3,6,12,24,... the dynamic entropies jump to the value according to the next Grassberger number. In this way a simple step function is obtained. We see that the dynamic entropy itself (which is the limit of infinite n ) is zero. For infinite histories the predictability is 1 i.e. 100%. This correponds to a zero Lyapunov exponent X = 0 [7,11]. In the region r > 3.5699 ... the Lyapunov exponent is in general larger zero, corresponding to chaotic states. Then the Pesin theorem h = X may be used in order to obtain a lower border for the conditonal entropies [5]. The convergence to the limit is rather fast. This may be exploited also for the investigation of optimal partitions [5]. Further we may use the knowledge of the lower order entropies and of the limit for the construction of Pade approximations. We mention that similar long range correlations are also generated by intermittent processes [Ill. A special group of discrete intermittent maps of such type was investigated by Szepfaluzy and coworkers [lo]. The following scaling for the approach to the limit was found
The processes considered so far, correspond to the limiting case of processes which are predictable on the basis of a long observation. This property is lost, if <1 noise is added which leads always to an upper limit of the predictability r,,, [11,91. There exist several types of noise, the simplest is the perturbation by a noisy channel. Assuming that the percentage E of the symbols is subject to random flips, the source entropy as a function of may be estimated by Real data behave similar to maps with noise. By combination of a solvable map with local structures on any scale, say the Feigenbaum map, with channel noise (measurement noise)and by adaptation of the parameter a wide variety of shapes of the entropy function h, may be represented. Let us present now an application of these concepts to the analysis of real strong noisy data [4]. Prediction of strong noisy data using classical linear methods usually fails to give accurate predictions and a reliable confidence level. The concept of entropy and local predictability in combination with classical methods is a good candidate to give reliable results. Applications of these concepts to meteorological strings were given in [12,13] and to nerve signals in [5]. In the following our concept will be demonstrated on daily stock index data St: Dow Jones 1900-1999 (27044 trading days). Since the stock index itself has an exponentially growing trend we will use daily logarithmic index changes
Entropies and Predictability of Nonlinear Processes and Time Series
1213
An application of the entropy concept requires a partitioning of the real value data xt into symbols At of an alphabet. To find an optimal partition and alphabet is a process of maximizing the Kolmogorov-Sinai entropy [5]. However for strong noisy signals with short memory an equal frequency of the letters is near to optimal. To be concrete we used X = 3 and At = 0; xt < -0 .OO% (strong decrease in the stock value), At = 2; xt > 0.0034 (strong increase), At = 1 (intermediate) were chosen [4]. In Fig. 1 the result of calculations of the conditional entropy is presented [4].
Fig. 1. C o n d i t i o n a l e n t r o p y h , = Hn+l - H , as a f u n c t i o n of word length n; the strong decrease for n 2 5 is an artefact due to length effects.
We see that the average predictability is rather small. For n 2 4 - 5 the error is growing due to length effects [14]. The further decay seems to be an artefact, the true entropy probably remains constant for n 2 5. Therefore the average uncertainty of the daily stock index is very high und the average predictability is less than 5%.
4
Predictions Based on a Local Analysis
Sometimes the analysis of the average entropies fails to detect existing correlations. On the other hand the average uncertainty of predictions is in many cases (e.g. for the stock market as shown above) higher than 0.9 (i.e. higher than 1.8 bits). Therefore the average predictability is rather low. For practical applications, one is not so much interested in an average value but even more in
1214 W. Ebeling
a concrete prediction based on the the observation of a concrete string of finite length n. In other words one is more interested in concrete predictions than in "average predict abilities" . Therefore we have studied also the predictabilities of the states following right after the particular strings A1 . . .A, which we denoted by r p )(A1 . . . A,) This is a quantity depends on the local "history" A1 . . .A, and fluctuates therfore while going along the string. Another closely related fluctuation quantity is the transinformation, which is connected with the local predictability. For the Fibonacci sequences as well as for the Feigenbaum sequences the local regularities follow from the grammar rules. Sometimes the next letter is nearly predetermined. Let us give just one example. In the Fibonacci sequence as well as in the Feigenbaum sequence the subsequence 00 is forbidden by the grammatical rules. Therefore in the state 0 the predictability of the next state is 1, after the symbol 0 comes the symbol 1 with 100% certainty. The rule that 00 is forbidden, creates a special local order. In the following the existence of local regularities will be demonstrated on the daily stock index data St discussed above. The result of the calculation of the local uncertainty h, (A1, . . . , A,) for the next trading day following behind an observation of n trading days A1, . . . , A, for n = 5 is plotted in Fig. 2. The local uncertainty is almost near one, i.e. the average predictability is very small. However behind certain patterns of stock movements A l l . . . , A, the local predictability reaches 8% - a notable value for the stock market, which in average is near to random. The mean predictability over the full data set is less then 2% (see Fig. 1). The question of the significance of the prediction is treated by calculating a distribution of local uncertainty hz (A1, . . . , A,) by help of surrogates. We constructed surrogate sequences having the same two point probabilities as the original sequence [4]. The level of significance Ii' was calculated as
where ( h i (A1, . . . , A,)) is the mean and c is the standard deviation of the local uncertainty distribution for the word A l l . . . ,A,. Assuming Gaussian statistics lIi'1 5 2 represents confidence greater then 95%. However since the local uncertainty distribution is more exponential like larger IGvalues are required to guarantee significance. For the analyzed data set a word length up to 6 seems to give still reliable results. In Fig. 2 we represented the uncertainty of the state subsequent to six observed states as a function of time, the interval corresponds to the last months of 1987 [4]. The greyvalue codes the level of significance calculated from a first order Markov surrogate. Dark represents a large deviation from the noise level (good significance). It is remarkable that higher local predictabilities coincide with larger levels of significance. This can be seen also from Table 1. Since we used a timeseries over a very long period we have to address the problem of non-stationary by dividing the original timeseries into smaller pieces.
Entropies and Predictability of Nonlinear Processes and Time Series
1215
Fig. 2. Local uncertainty of the the daily Dow Jones index ( i n symbolic representation) which foddows after a n observation of 5 subsequent days. W e represented a n interval corresponding t o the second half of 1987.
Furthermore instead of producing surrogates on the level of Symbols one can discuss surrogates obtained by modells of a stockmarkets like ARCHIGARCHmodells. This has been done in [4]. Analyzing the data in Table 1 we see, that in spite of the fact that the average predictability is very low (about 2%) there are special days, where the predictability is up to 8%, i.e. up to 4 times higher than in avaerage. We remember that in our way of coding 0 stands for a day with a strong downswing of the index, 2 stands for a strong upswing and 1 stands for a day where the index remains nearly constant. Remarkable is, that the highest predictability correponds to the days following the October-Crash in 1987. As a result of these investigations we may state that in spite of the fact that the stock market index index is in average very uncertain, some local order might be detected which is helpful for predictions. Similar resultes were obtained for meteorological data and for nerve signals [5,12,13].
5
Conclusions
Our results show that the dynamic entropies are an appropriate measure for studying the predictability of complex processes. Of particular interest are local studies of the predictabilities after certain local histories. Local minima of the uncertainty may be found in many processes including even the index of the stock market. The basic problem for improving predictions is the detection of middle range and of long range correlations. These correlations are of specific interest since they improve the predictability. If long range correlations exist, one can improve the resultes by basing the predictions at longer observations.
1216 W. Ebeling
Table 1. Sequences of 3-5 daily indices of the Dow Jones with the highestpredictability of the following (nextday) index.
Further we have found that there are specific local substrings, where the uncertainty is much smaller than the average, i.e. the predictability is better than in average. In other words, even for the case of noisy data, there are specific situations where local predictions are possible, since the local predictability is much better than the average predictability. It may be of practical importance to find out all substrings which belong to this particular class. Our results clearly demonstrate that the best chance for predictions is based on the observation of ordered local structures. The entropy-like measures studied here operate on the sentence and the word level. In some sense entropies are the most complete quantitative measures of correlation relations. This is due to the fact that the entropies include also many point-correlations. On the other hand the calculation of the higher order entropies is extremely difficult and a t the present moment there is no hope to extend the entropy analysis to the level of hundreds of letters. In conclusion we may say that a more careful study of the correlations in time series sequences of mediate and long range may contribute to better predictions of complex processes. The author thanks J. Freund, L. Molgedey, T . Poschel, K. Rateitschak, and R. Steuer for many fruitful discussions and a collaboration on special topics of the problems discussed here.
References 1. Feistel, R., Ebeling, W.: Evolution of Complex Systems, Kluwer Academic Publ., Dordrecht 1989. 2. Ebeling, W., Nicolis, W.: Word frequency and entropy of symbolic sequences: a dynamical perspective, Solitons & Fractals 2 (1992) 635-640. 3. Ebeling, W.: Prediction and entropy of sequences with LRO. Physica D 109 (1997) 42-50.
Entropies and Predictability of Nonlinear Processes and Time Series
1217
4. L. Molgedey, W. Ebeling: Local order, entropy and predictability of financial time series, Eur. Phys. J B 15 (2000) 733-737; Physica A 287 (2000) 420-427. 5. R. Steuer. L. Molgedey, W. Ebeling, M.A. Jimenez-Montano: Entropy and optimal partition for data analysis, Eur. Phys. J. B 19 (2001) 265-269. 6. Ebeling, W., Steuer, R., Titchener, M.R.: Partition-based entropies of deterministic and stochastic maps. Stochastics and Dynamics 1 (2001) 45-61. 7. Grassberger, P.: Entropy and complexity. Int. J. Theor. Phys. 25 (1986) 907-915. 8. T. Gramss, T.: Entropy of Fibonacci sequences. Phys. Rev. E 50 (1994) 2616-2620. 9. Ebeling, W., Rateitschak, K.: Symbolic dynamics, entropy and complexity of the Feigenbaum map at the accumulation point. Discrete Dyn. in Nat. & Soc.2 (1998) 187-194. lo. Szepfaluzy, P., Gyorgyi, G.: Entropy of nonlinear maps. Phys. Rev. A 33 (1986) 2852-2860. 11. Freund, J., Ebeling, W., Rateitschak, K.: Self similar sequences and universal scaling of dynamical entropies, Phys. Rev. E 54 (1996) ; Int. J. Bifurc. & Chaos 6 (1996) 611-620. 12. Nicolis, C., Ebeling, W., Baraldi, C.: Markov processes, dynamical entropies and the statistical prediction of mesoscale wheather regimes. Tellus 49 A (1997) 108118. 13. Werner, P.C., Gerstengarbe, F.-W., Ebeling, W.: Changes in the probability of sequences, exit time distribution and dynamical entropy in the Pot sdam temperature record, Theor. Appl. Climatol. 62 (1999) 125-132. 14. Poschel, T., Ebeling, W., Rosk, H.: Guessing probability distributions from small samples, J. Stat. Phys 80 (1995) 1443-1452.
Author Index
Abdalhaq, B. I-513 Ablitt, N. III-285 Abou-Chakra, H. I-743 Abramson, D. II-834 Abubakar, A. III-207 Adam, J.C. III-342 Addison, C. III-755 Adelmann, A. III-362 ˚ Ahlander, K. III-711 Akinjide, E. I-1030 Albada, G.D. van I,176, I-653, I-693 Alberts, C.-P. II-295 Aldous, J.R. II-695 Alexandrov, N. II-688 Alexandrov, V. II-609, II-619, II-688 Alique, A. III-574 Alique, J.R. III-574 Amodio, P. III-429 Amorim, C.L. de I-296 An, W.H. I-862 Andersson, M. III-26 Anido, L. I-78, III-872 Antonelli, L. II-171 Arbenz, P. III-295 Arickx, F. II-679 Arnal, A. II-229 Artoli, A.M. I-361 Astalos, J. I-543 Atanassov, E.I. II-628 Axelsson, O. III-784 Ayala, G. I-458 Baboolal, S. II-452 Bader, M. III-662 Bal, H.E. II-558 Ba=la, P. I-552 Bali´s, B. II-884 Balsoy, O. I-22 Banda, M.K. I-930 Basaruddin, T. I-653 Bassi, A. II-568 Baumann, M. III-419 Baxter, J. I-743 Beazley, D.M. II-824 Bechmann, D. II-62, II-219
Beck, M. II-568 Bekker, H. III-55 Beletskyy, V. I-409 Belleman, R.G. I-693, III-265 Bellucci, D. III-918 ´ Belmonte, O. II-111 Benavent, X. I-458 Benderskaya, G.Y. II-412 Benhamou, F. II-1097 Berg, P.M. van den III-207 Berridge, S. II-510 Berti, G. III-745 Berzins, M. I-523 Bhatia, A. III-227 Bhowmick, S. II-325 Bhuruth, M. II-393 Bianco, M. I-733 Bilardi, G. I-733 Biryukov, V. III-372 Bischof, C.H. II-1069 Blaheta, R. III-830 Blais, J.A.R. III-164 Blanco, V. II-923 Blobel, B. II-982 Bloor, M.S. III-184 Boada, I. II-121 Bogdanov, A.V. III-1176 Bollman, D. II-548 Bolvin, H. I-920 Boojhawon, R. II-393 Borisov, M.O. I-316 Botana, F. II-211 Boukhanovsky, A.V. I-216, I-683 Bourchtein, A. III-813 Bourchtein, L. III-813 Box, F.M.A. III-255 Bregman, B. II-767 Brezany, P. I-484 Bridle, I. I-743 Bright, N. II-894 Brodlie, K. I-853 Broeckhove, J. II-679 Bubak, M. I-484, II-874, II-884 Buchem, M.A. van III-255 B¨ ucker, H.M. II-1069
1220
Author Index
Bungartz, H.-J. III-662 Buono, N. Del III-467, III-526 Burger, M. den II-558 Butikov, E.I. III-1154 Cabaleiro, J.C. II-923 Caeiro, M. I-78, III-872 Cai, X. II-345 Cai, X.-C. I-533 Cai, Z. I-613 Calle, M. III-544 Calvez, C. Le II-364 Campari, E.G. I-763 Carle, A. II-1029 Carracciuolo, L. II-171 Carvalho, L.A.V. de I-236 Cary, J.R. III-334 Castellano, G. I-970 Castiello, C. I-970 Castillo, E. II-374 Ceccarelli, M. II-171 Cetin, N. I-371 Chambarel, A. I-920 Chang, L.-P. I-449 Chapelle, P. I-743 Chen, J. III-994, III-1004 Chen, W.-C. I-449 Cheng, J.-R.C. I-1020 Chiarantoni, E. III-439 Cho, G. III-1128 Cho, Y. II-82, II-275 Choi, B. I-276 Choi, H. III-1090 Choi, Y. III-1100 Choo, H. III-1061, III-1108 Chover, M. II-111 Chow, P. III-755 Christakis, N. I-743 Chung, J.Y. III-1042 Chung, T.M. III-1051 ˇ Ciegis, R. II-422 Clercx, H.J.H. I-1010 Coleman, S.A. I-1077 Cort´es, A. I-513 Cos´ın, C. II-72 Cotofrei, P. I-572 Coufal, D. III-584 Cox, C.L. III-735 Cox, D. II-461 Crocchianti, S. III-908
Croft, N. II-806 Croft, T.N. II-480 Cross, M. I-743, II-806, II-943, II-953 Crutchfield, J.P. I-793 Curley, M.G. II-646 Cyr, E.C. III-735 D’Amore, L. II-171 Daescu, O. III-65, III-227 Dahlblom, O. III-701 Dakowicz, M. III-144 Dam, M. ten I-663 Danelutto, M. II-844 Danˇek, J. III-820 Danilov, V. III-325 Daoud, D.S. I-324 Dasgupta, B. II-442 Datta, A. I-306, III-75 Debelov, V.A. II-13 Debenham, J. I-246 Decyk, V.K. III-342 Degtyarev, A. I-683, III-564 Delgado, A. I-980 Deng, S. III-342 Derksen, J. I-713 Dewar, W. I-419 D´ıaz-B´ an ˜ez, J.M. III-46 Diele, F. III-449, III-476 Din˘cov, D.D. I-813 Dmitrieva, L. III-1201 Dobrucky, M. I-543 Domingo, J. I-458 Dom´ınguez, J.J. III-544 Donangelo, Raul I-236 Doornbos, J. III-255 Dou, J. III-966 Douglas, C.C. III-774 Drees, A. III-372 Drenth, W. I-429 Drkoˇsov´ a, J. III-536 Dumitrescu, A. III-14 Dung, L.P. II-834 Duperrier, R. III-411 Dvorak, M. II-739, II-758, II-777 Ebeling, W. III-1209 Echevarr´ıa, G. II-305 Elia, C. III-467, III-526 Elkouhen, M. II-62 Emmen, A. I-70, II-995
Author Index Engelen, R. van I-419 Engelmann, C II-720 Escobar, J.M. I-335, I-950 Espinosa, E.L. I-704 Esser, B. III-1184 Essert-Villard, C. II-151 Eswaran, V. II-442 Evans, G. I-910 Fagan, M. II-1029 Falgout, R.D. III-632 Fanelli, A.M. I-970 Fathi, B. II-609, II-619 Fatta, G. Di I-286 Feichtinger, D. III-362 Feng, Y. I-399 Ferdinandova, I. I-439 Fernandez, M. II-22 Fern´ andez, M. II-111 Fern´ andez-Iglesias, M.J. I-78 Ferr´ andez, A. I-61 Ferri-Ram´ırez, C. I-166 Figge, M.T. I-803 Filbet, F. III-305 Finkel, R. I-51 Fischer, B. III-202 Fisher, J. III-184 Flanery, R.E. II-864 Fliller, R.P. III-372 Fonseca, R.A. III-342 Fornarelli, G. III-439 Forster, C. III-1170 Forth, S.A. II-1077 Foster, J. II-695 Fox, G. I-22, I-503 Franc, M. II-42 Frank, A. III-662 Froehlich, D. I-543 Frolova, Julia I-226 Fuentes, E. II-568 Funika, W. II-874, II-884 G¨ artner, K. II-355 Gallivan, K. I-419 G´ alvez, A. II-161, II-305 Gang, X.J. I-862 Gannon, D. I-22 Gao, J. III-285 Garc´ıa, G.C. I-704 Gatehouse, P. III-285
Gavrilov, D. III-115 Gavrilova, M.L. III-105, III-115 Geest, R.J. van der III-242, III-255 Geist, G.A. II-720 Gelas, J.-P. II-578 Gervasi, O. III-956 Gervois, A. III-95 Gevorkyan, A.S. III-1176 Giering, R. II-1019 Giesen, J. III-154 Gloukhov, V. I-753 G=lut, B. I-353 Gofen, A. I-562, I-1000 Gold, C. III-135, III-144 Gold, C.M. III-1004 ´ G´ omez-Nieto, M.A. I-704 Gonz´ alez, L. I-137 Gonz´ alez, P. II-923 Gonz´ alez-Santos, G. I-391 Gonz´ alez-Yuste, J.M. I-335, I-950 Goodyer, C. I-523 Gorokhov, O.Yu. I-186 Greiner, G. III-652 Gudmundsson, J. III-26 Guha, S. III-14 Guillaume, P. II-364 Haas, P.C.A. de I-1010 Haber, R.E. III-574 Haber, R.H. III-574 Hakl, F. III-554 Halada, L. I-206 Han, S. I-276, I-643 Han, S.-K. III-1118 Hanjali´c, K. I-266 Hartmann, C. I-980 Hassan, O. II-816 Hayryan, E. III-804 H´egron, G. II-285 Heimbach, P. II-1019 Heitzer, M. I-833 Helmke, U. III-419 Helstrup, H. I-494 Heras, D.B. II-923 Hern´ andez, J.C. III-1024 Hern´ andez-Orallo, J. I-166 Herrero, H. II-374 Herron, M.G. I-1077 Hey, T. I-3 Hill, C. II-1019
1221
1222
Author Index
Hinsen, K. III-691 Hlav´ aˇcek, I. III-840 Hlav´ aˇcek, M. III-554 Hluchy, L. I-543 Hod´ asz, G. I-51 H¨ ofinger, S. II-963 Hoekstra, A.G. I-88, I-361 Holmes, J. III-325 Holmgren, S. III-681 Hong, J. I-643 Hong, S.S. III-1051 Hovland, P. II-1087 Hu, J. III-774 Huard, A. II-364 H¨ ulsemann, F. III-652 Huh, E.-N. III-1071 Huh, U.-Y. III-613 Hungersh¨ ofer, J. III-36 Hurtado, F. III-46 Huˇsek, P. II-520, III-604 Huybers, P. III-85 Hwang, J. III-1080 Iavernaro, F. III-429 Ierotheou, C.S. II-953 Iglesias, A. II-161, II-181, II-191, II-305 Imamiya, A. II-131 Irony, D. II-335 Isasi, P. III-1024 Islas, A.L. III-486 Jacob, R.L. II-748 Jakl, O. III-830 Jancic, J. II-894 Jang, J.-H. I-1068 Janssen, J.P. III-242 Jeltsch, R. III-863 Jimack, P.K. II-797 Jim´enez-Morales, F. I-793 Johansson, P. I-872 John, M. III-154 Johnson, S. II-953 Jolivet, V. II-3 Jurczyk, T. I-353 Kaandorp, J. I-88 Kacsuk, P. II-671 Kalous, R. III-554 Kamerman, D.J. I-117 Kang, D.W. II-32
Kantur, O. II-432 Karagiorgos, G. I-623 Karaivanova, A. II-598 Karl, W. II-933 Karnik, M. II-442 Kashio, K. II-131 Kato, H. II-131 Kato, K. I-990 Katsouleas, T. III-342 Keane, A.J. I-881 Kenjereˇs, S. I-266 Keppens, R. I-940 Kielmann, T. II-558 Kim, C. I-276 Kim, D. II-275 Kim, D.-S. II-82, II-275 Kim, D.S. III-1051 Kim, K. I-276, I-643 Kim, K.H. III-1051 Kim, S. III-1042 Kim, S.J. III-1100 Kim, Y.S. III-1080 Kipfer, P. III-652 Kiryukhin, I. III-564 Klouˇcek, P. II-461 Kohl, J.A. II-864 Kole, J.S. I-803 Kolingerov´ a, I. III-125 Koning, P.J.H. de III-242, III-255 Korobitsin, Victor I-226 Korotenko, K. III-1193 Kotocova, M. I-890 Koulisianis, M.D. I-673 Kowarschik, M. III-642 Kranzlm¨ uller, D. II-913 Kressner, D. I-872 Krishnaswamy, S. I-603 Kroc, J. I-773, III-217 Krol, M. II-767 Krstovi´c, G. I-266 Kr¨ uger, T. III-950 Kudov´ a, P. III-594 Kulikov, G.Y. II-412 Kume, E. I-990 Kuperin, Y. III-1201 Kurzyniec, D. II-709 Kwiesielewicz, M. I-468 La, S.C. I-960 Lad´ anyi, L. I-592
Author Index Lagan` a, A. III-908, III-918, III-926, III-956 Lai, C.-H. II-480 Lang, B. II-1069 Langtangen, H.P. III-764 Larson, J.W. II-748 Lee, B.G. III-1080 Lee, D. III-1090 Lee, D.C. III-1118 Lee, H. I-643 Lee, S. III-342, III-1100 Lee, W.-Y. III-613 Lee, W.J. III-1080 Lef`evre, L. II-578 Leggett, P. II-953 Lehmann, U. II-1049 Leinen, P. II-470 Lemaire, J.-L. III-305 Leopold, C. I-843 Levcopoulos, C. III-26 Levi, G. I-763 Li, S. I-960 Li, X.K. I-910 Li, Z. III-1004 Lien, J. I-494 Lim, I.-S. III-613 Lindemann, J. III-701 Lindenstruth, V. I-494 Lintner, M.D. III-882 Liu, B. II-609, II-619 Liu, Y. I-127 Llamas, M. I-78 Llopis, F. I-61 Lluch, A. II-229 Loader, R.J. II-655, II-665 Lobbiani, M. III-956 Lobosco, M. I-296 Loke, S.W. I-603 Lopez, L. III-526 Lorenzetto, G.P. III-75 Lozano, M. II-22 Lozano, S. III-544 Lu, M. I-613 Lu, W. III-342 Lum, E.B. II-102 Luque, E. I-107, I-513 Lyakh, A. III-194 Ma, A.-N. Ma, K.-L.
III-975, III-984 II-102, III-352
1223
Ma, Y.-a. III-966 Maas, R. I-663 Mac´e, P. II-285 Maimour, M. II-588 Malawski, M. I-484 Malitsky, N. III-372 Malloy, B.A. III-735 Mao, S.-J. III-975, III-984 Mao, X. II-131 Marcelli, G. III-932 Mardal, K.-A. III-764 Margalef, T. I-513 Marir, F. I-41, II-258 Mark, P. van der I-419 Marsden, R.H. II-480 Marsh, A. II-972, II-1012 Martins, Isa H. I-236 Martyniak, J. III-234 Maryˇska, J. III-794 Mascagni, M. II-598, II-635 Masoumi, M.E. I-723 Mattheij, R.M.M. I-1010 Maubach, J.M. I-429 McManus, K. II-806, II-943 Meier Yang, U. III-632 Meijer, H. III-46 Melnik, R.V.N. II-490 Memelli, M. III-926 Meng, I.-H. I-449 Meng, S. I-910 M´eot, F. III-381 Meriste, M. I-156 Merks, R. I-88 M¸etel, P. II-874 Mickelson, S.A. II-739, II-758, II-777 Missirlis, N.M. I-623 Mitchell, M. I-793 Mitchell, W.F. III-672 Modersitzki, J. III-202 Mohr, M. II-528 Monostori, K. I-51 Montenegro, R. I-335, I-950 Monterde, J. II-72, II-229 Montero, G. I-335, I-950 Moore, S.V. II-904 Moore, T. II-568 Moreno, J.A. I-147 Morgan, K. II-816 Mori, P. III-898 Mori, W.B. III-342
1224
Author Index
Morozov, I.V. III-1137 Motus, L. I-156 Mower, J.E. III-174 Mun, Y. III-1061, III-1071 Mun, Y.S. III-1118 Mundani, R. III-662 Munt, P. I-107 Murillo, M. I-533 Murli, A. II-171 Muylle, J. II-787 Nadeem, S.A. II-797 Nagel, K. I-371 Naidoo, R. II-452 Nakai, J. I-256 Narasimhan, G. III-26 Naumann, U. II-1039 Navazo, I. II-121 Nechaev, Yu.I. I-683, III-564 Nedoma, J. III-840 N´emeth, Z. II-729 Neruda, R. III-536, III-594 Neytcheva, M. III-784 Nguyen, G.T. I-543, I-890 Nguyen, T. I-474 Niaei, A. I-723 Nieter, C. III-334 Nipp, K. III-863 Nool, M. I-900 Nord´en, M. III-681 Norman, G.E. III-1137 Norris, B. II-1087 Nowi´ nski, A. I-552 Nowi´ nski, K. I-552 Ochmanska, E. I-1049 Oehsen, J. Barr von III-735 Oger, L. III-95 Okunbor, D. I-1030 Oliveira, S. I-1058 Ong, E.T. II-748 Or=lowski, R. II-874 Orozco, E. II-548 Otero, C. II-315 Otto, K. III-711 Ouazzane, K. I-41, II-258 Overmars, M.H. III-3 Pacifici, L. III-908 Pakalnyt˙e, V. II-422
Pallickara, S. I-22 Pan, Y. III-888 Papatheodorou, T.S. I-673 Park, Y.S. II-32 Parker, S.G. III-719 Parrott, K.A. I-813 Pascoe, J.S. II-655, II-665, II-688 Pataki, M. I-51 Patel, M.K. I-743 Paternoster, B. III-459 Pavani, R. III-516 Pavlova, M.I. III-1176 Pavluˇs, M. III-804 Pedersen, P.W. II-538 Pena, T.F. II-923 P´erez, G. II-268 Perez, M. II-22 Pericleous, K.A. I-813 Perminov, V. I-823 Pesavento, F. I-733 Petrov, E. II-1097 Peyr´e, H. II-219 Pflaum, C. III-622 Pflug, G.Ch. I-206 Pham, C. II-588 Pichoff, N. III-411 Pierce, M. I-503 Piermarini, V. III-908 Plank, J.S. II-568 Plassmann, P.E. I-1020 Plemenos, D. II-3 Plumejeaud, C. I-474 Poellabauer, C. II-894 Poernomo, I. II-854 Politi, T. III-439, III-449 Pombo, J.J. II-923 Posch, H.A. III-1170 Poulingeas, P. II-3 Prodan, A. I-1040 Prodan, R. I-1040 Proot, M.M.J. I-900 Pruneda, R.E. II-374 Pryce, J.D. II-1077 Pucci, G. I-733 Puig-Pey, J. II-161 Pytelkov´ a, R. II-520, III-604 Qiang, J.
III-352
Raedt, H. De
I-803
Author Index Raedt, K. De III-55 Raghavan, P. II-325 Ragni, S. III-476 Rajabi, M.A. III-164 Ralphs, T.K. I-592 Ram´ırez-Quintana, M.J. I-166 Raney, B. I-371 Rappaport, D. III-46 Rasch, A. II-1069 Ratto, M. I-196 Ray, J. III-774 Re, G. Lo I-286 Reiber, J.H.C. III-242, III-255 Reid, J.K. II-1077 Remolar, I. II-111 Ren, C. III-342 Reussner, R. II-854 Reynolds, D.R. II-461 Ribagorda, A. III-1024 Ribelles, J. II-111 Ricci, L. III-898 Richard, P. III-95 Riganelli, A. III-926 Rivera, F.F. II-923 Roanes-Lozano, E. II-52 Roanes-Mac´ıas, E. II-52 Roberts, A.J. II-490 Rocco Sanseverino, C.M. I-147 Rodeiro, J. II-268 Rodr´ıguez, E. I-335, I-950 Rodr´ıguez, J. III-872 Rodr´ıguez, J.S. I-78 R¨ obenack, K. II-1059 R¨ ohrich, D. I-494 Rokne, J. III-105, III-115 Ros, S. III-574 Ruede, U. III-852 R¨ ude, U. III-642, III-652 Ruiz, I.L. I-704 Ruskin, H.J. I-127, I-381, I-399 Rutten, M.C.M. III-255 Ryne, R. III-352 Ryoo, S.T. II-141 Ryu, J. II-82 Saad, Y. II-345 Saarelainen, M. II-1003 Sadrameli, M. I-723 Sadus, R.J. III-932 Saltelli, A. I-196
1225
Saltzman, M.J. I-592 Samulyak, R. III-391 Sandberg, G. III-701 Santos, J. I-78, III-872 Santoso, J. I-653 Santos Costa, V. I-296 Sax, A.F. III-950 Sbert, M. II-249 Schaap, J.A. III-242, III-255 Schenk, O. II-355 Schmidt, H.W. II-854 Schober, C.M. III-486 Schoemaker, R.M. I-1010 ´ II-201 Schramm, E. Schreck, P. II-201 Schrefler, B.A. I-733 Schulz, M. II-933 Schumacher, H. II-510 Schwan, K. II-894 Scotney, B.W. I-1077 Scott, S.L. II-720 Scurr, A.D. I-881 Sea¨ıd, M. I-930 Segers, A. II-767 Seguel, J. II-548 Seider, D. III-622 Sellares, T. III-46 Semoushin, I.V. I-186 Seo, S.H. II-32 Sevastyanov, I.M. II-13 Sever´ yn, O. III-794 Sgura, I. III-449 Shakhov, V.V. III-1108 Shevchuk, V.A. II-500 Shishlo, A. III-325 Shklarski, G. II-335 Shulakov, R. III-265 Sibley, G. II-665 Sierra, J.M. III-1024 Sigalas, M.P. III-942 Silva, L.O. III-342 Skaali, B. I-494 Skaburskas, K. I-633 Skala, V. II-42 Slominski, A. I-22 Sloot, P.M.A. I-88, I-176, I-361, I-653, I-693 Smith, B. II-1087 Smith, K. I-582 Sofroniou, M. III-496, III-506
1226
Author Index
Soma, T. I-1058 Sonnendr¨ ucker, E. III-305 Sørensen, K. II-816 Soroka, I. III-1201 Sosnov, A. II-285 Sosonkina, M. II-345 Soundaralakshmi, S. I-306 Spaletta, G. III-496, III-506 Spinnato, P.F. I-176 Srinivasan, A. II-635 Stanisz-Wallis, K. III-234 Stankova, E.N. III-1176 Star´ y, J. III-830 ˇ edr´ Stˇ y, A. III-536 Stegailov, V.V. III-1137, III-1147 Steinbeck, T. I-494 Stewart, D. I-1058 Stoffel, K. I-572 Stuer, G. II-679 Subasi, D.S. I-324 Sun, M. III-975, III-984, III-1014 Sunderam, V. II-709, II-729 Sunderam, V.S. II-655, II-665 Sung, H. I-643 Suppi, R. I-107 Tadjouddine, M. II-1077 Tai, X.-C. I-345 Tam, L. I-582 Tan, C.J.K. II-383 Taniar, D. I-582 Tao, J. II-933 Tasso, S. III-918 Taylor, J. II-739, II-758, II-777 Teberekidis, V.I. III-942 Tehver, M. I-633 Teranishi, K. II-325 Teti, P. II-844 Thoai, N. II-913 Thorne, D. III-774 Thun´e, M. III-681 Todd, B.D. III-932 Togores, R. II-315 Toledo, S. II-335 Topa, P. I-97, I-783 Topping, B.H.V. II-787 T´ oth, G. I-940 Towfighi, J. I-723 Tran, V.D. I-543, I-890 Trbojevic, D. III-372
Trefethen, A.E. I-3 Trigiante, D. III-429 Troadec, J.P. III-95 Tse, R.O.C. III-135 Tsolis, G.K. I-673 Tsung, F.S. III-342 Tsybulin, V. II-432 Tuminaro, R. III-774 Tyszka, J. I-97 Uden, E. van I-468 Ullaland, K. I-494 Uriot, D. III-411 Urso, A. I-286 Uyar, A. I-22 Vainikko, E. I-633 Valuev, I.A. III-1184 Vankan, W.J. I-663 Vanrumste, B. II-528 Vargas-Jarillo, C. I-391 Vassell, C. III-1032 V´ azquez, P.-P. II-249 Vedru, J. I-633 Vegara, F. I-458 Velthoven, P. van II-767 Verdonck, P.R. III-275 Vergura, S. III-439 Vestbø, A. I-494 Vicedo, J.L. I-61 Vierendeels, J.A. III-275 Vinogradov, O. III-115 Voellmy, A. I-371 Vohral´ık, M. III-794 Volkert, J. II-913 Vorobiev, L.G. III-315 Voss, H. II-403 Vosse, F.N. van de III-255 Vrtic, M. I-371 Walczycka, L. III-234 Walkley, M. I-853 Walshaw, C. II-806, II-943 Walther, A. II-1049 Wang, J.-b. III-966 Wang, L. I-960 Wang, R. I-381 Watanabe, T. I-990 Watson, G. II-834 Weatherill, N. II-816
Author Index Wedemann, Roseli S. I-236 Weiß, C. III-642 Welch, L.R. III-1071 Welch, P.H. II-687, II-695 Westenberg, J.J.M. III-242 Wiebalck, A. I-494 Wiechert, W. III-858 Wierum, J.-M. III-36 Wilde, T. II-864 Wilkinson, M.H.F. I-117 Williams, A. II-806 Wilson, B. II-102, III-352 Wirza, R. III-184 Wism¨ uller, R. II-874, II-884 Wolf, M. II-894 Woo, B.K. III-1051 Wood, J. I-853 Xue, Y.
I-41, III-975, III-984, III-1014
Yan, L. III-966 Yang, G.-Z. III-285 Yang, W.-P. I-449
1227
Yoo, H.S. I-1068 Yoon, K.H. II-32, II-141 York, R.C. III-315 You, L. II-92 Youn, C. I-503 Youn, H.Y. III-1108 Yu, W.H. I-960 Zacharias, J. I-980 Zaj¸aac, K. I-484 Zaplatin, E. III-401 Zaslavsky, A. I-51, I-603 Zerzour, K. I-41, II-258 Zhang, J.J. II-92, II-239 Zhang, Y. III-994, III-1004, III-1014 Zhao, R. III-994, III-1004, III-1014 Zhao, Z. I-693 Zheng, J.J. II-239 Zhong, D.G. I-862 Zhou, L. III-994 Zhu, B. I-399 Zudilova, E.V. I-316