Specification Transformation Navigation - Informatik - FB3

Transcription

Specification Transformation Navigation - Informatik - FB3
!
"#$%&"'()*+
,-./!0"
""!!
Bernd Krieg-Brückner
Preface
As we began preparing this Festschrift in late 2008 and sent out invitations to
submit to Bernd’s friends and present and former members of his group, we
had no way of anticipating the huge resonance to this. We ended up receiving
no less than 33 submissions, including 19 full technical papers. Out of these we
accepted 33 submissions, including 19 full technical papers, for publication in
the Festschrift – a staggering 100% acceptance rate!
Our deepest thanks goes to all the authors for providing the interesting and
humorous contributions. We wish all of you to enjoy the Colloquium. Our congratulations go to Bernd, both on occasion of his birthday, but also regarding his
fruitful and productive scientific life. All the best for the next 60 (or so) years,
Bernd!
February 2009
The Festschrift committee
Bernd Gersdorf
Berthold Hoffmann
Christoph Lüth
Till Mossakowski
Sylvie Rauer
Thomas Röfer
Markus Roggenbach
Lutz Schröder
Shi Hui
Mattias Werner
Table of Contents
Introduction
Specification, Transformation, and Navigation – Research Topics of
Bernd Krieg-Brückner . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
Berthold Hoffmann, Till Mossakowski, Thomas Röfer, Bernd
Gersdorf
Promotionen und Habilitationen . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
Mattias Werner
1
13
Greetings
Glückwünsche zum Sechzigsten . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
Wolfgang Bibel
15
Greetings . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
Erwin R. Catesbeiana, jr.
17
google bilder: bernd krieg brückner . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
Martin Gogolla
18
Grußwort . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
Modar Ibraheem
19
Greetings . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
Christoph Lüth
20
Greetings . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
Klaus Lüttich
22
Greetings . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
Peter Mosses
23
Pencil Sharpener . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
Frieder Nake
24
Birthday Card . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
Zhenyu Qian
26
Grußwort . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
Pedro de la Cruz Ramos
28
Félicitations . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
Sylvie Rauer
29
Greetings . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
Don Sannella
30
To Bernd Krieg-Brückner on the Occasion of his 60th Birthday . . . . . . . . .
Andrzej Tarlecki
32
Grußwort zum 60. Geburtstag . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
Mattias Werner
34
Technical Papers
De Bello Bremense – et effectibus eius in Saravica . . . . . . . . . . . . . . . . . . . . .
Serge Autexier, Dominik Dietrich, Dieter Hutter
35
On Inconsistency and Unsatisfiability . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
Erwin R. Catesbeiana
40
The VSE Refinement Method in Hets . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
Mihai Codescu, Bruno Langenstein, Christian Maeder, Till
Mossakowski
51
Encapsulating Lazy Behavioral Subtyping . . . . . . . . . . . . . . . . . . . . . . . . . . . .
Johan Dovland, Einar Broch Johnsen, Olaf Owe, and Martin
Steffen
72
A Paradox in Estimating Uncertainty . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . .
Udo Frese, Oliver Birbach
88
(A) Vision for 2050 — Context-Based Image Understanding for a
Human-Robot Soccer Match . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 103
Udo Frese, Tim Laue, Oliver Birbach, Jörg Kurlbaum, Thomas
Röfer
On the whereabouts of CSP-CASL – A Survey . . . . . . . . . . . . . . . . . . . . . . . 121
The Processes & Data Gang – Andy Gimblett, Temesghen Kahsai,
Liam O’Reilly, Markus Roggenbach
Enhanced Formal Verification Flow for Circuits Integrating Debugging
and Coverage Analysis . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 140
Daniel Große, Görschwin Fey, Rolf Drechsler
Program Graph Transformation . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 152
Berthold Hoffmann
Type-Safe Red-Black Trees with Java Generics . . . . . . . . . . . . . . . . . . . . . . . 168
Stefan Kahrs
Compensating the Computational Bias of Spreadsheets . . . . . . . . . . . . . . . . 184
Andrea Kohlhase, Michael Kohlhase
Formal Management of CAD/CAM Processes . . . . . . . . . . . . . . . . . . . . . . . . 201
Michael Kohlhase, Johannes Lemburg, Lutz Schröder, Ewaryst
Schulz
Graph Transformation Units to Model Model Transformation – First Ideas 217
Hans-Jörg Kreowski
Test Automation Problems and Their Reduction to Bounded Model
Checking . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 229
Jan Peleska
Regularity based on categorical algebra . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 246
Horst Reichel
15 Years of Rolland . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 255
Thomas Röfer, Christian Mandel, Axel Lankenau, Bernd Gersdorf,
Udo Frese
On Classical and Quantum Process Algebra as Abstract Data Types . . . . 273
Ruqian Lu, Lixing Li, Yun Shang, Xiaoyu Li
AG.com – Bremer Informatik-Arbeitsgemeinschaften in der gymnasialen
Sekundarstufe I . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 291
Holger Schlingloff
Using Orientation Calculi and Route Graphs to Interpret Route
Descriptions . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 303
Hui Shi
Top-down vs. Bottom-up . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 314
Burkhart Wolff
Author Index . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 339
Specification, Transformation, and Navigation
– Research Topics of Bernd Krieg-Brückner
Berthold Hoffmann1 , Till Mossakowski1,2,3 ,
Thomas Röfer3 , and Bernd Gersdorf3
1
FB-3 – Fachbereich Mathematik und Informatik,
Universität Bremen, Postfach 330 440, D-28334 Bremen
{hof,till}@informatik.uni-bremen.de
2
Albert-Ludwigs-Universität Freiburg, Institut für Informatik,
Georges-Köhler-Allee 52, D-79110 Freiburg
[email protected]
3
Deutsches Forschungszentrum für Künstliche Intelligenz GmbH,
Sichere kognitive Systeme, Enrique-Schmidt-Str. 5, D-28359 Bremen
{Till.Mossakowski,Thomas.Roefer,Bernd.Gersdorf}@dfki.de
Scientific Career. Bernd Krieg Brückner studied electrical engineering at the
universities of Hannover and Erlangen. He received his master of computer science at Cornell University in 1971, with a thesis on the implementation of blockoriented languages.
He worked for the research project “Computer-aided Intuition-guided Programming” (CIP ) at TU München, where he received his PhD in 1978, for a
thesis on program specification and transformation. Then he joined Jean Ichbiah’s design team for the Ada programming language.
After a stay at the universities of Stanford and Berkeley in 1979-1980, he
returned to TU München, before he became a professor of computer science at
Universität Bremen in 1982. Here he participated in numerous research projects
in the areas of program specification, transformation, on spatial cognition and
robot navigation (since 1990), and on ambient-assisted living (since 2007). He
was a member of the university’s Zentrum für Kognitionswissenschaften (center for cognitive sciences), and the Technologie-Zentrum Informatik (technology center for computer science), he takes part in the Sonderforschungsbereich
(Collaborative Research Center) “Spatial Cognition: Reasoning, Action, Interaction”, and he co-founded the Bremen branch of Deutsches Forschungszentrum
für Künstliche Intelligenz (German Research Center for Artificial Intelligence).
Research Topics. Three keywords occur frequently in Bernd’s research:
– Specification
– Transformation
– Navigation
The following paragraphs give more details about Bernd’s contributions to these
areas, mention languages and tools developed in these project, and, finally, list
his publications.
1
2
Berthold Hoffmann
Specification and Transformation. Ever since Bernd participated in the CIP
project, he has been interested in the formal specification of program properties,
and in the transformation of specifications and programs so that their correctness
would be preserved, while their operational behaviour is improved. The ultimate
goal of this research is the development of provably correct efficient programs. To
support this goal, one needs languages wherein specification and transformations
can be defined, and tools that analyse specifications and execute transformations.
Language and Tools. Bernd has been involved in the design of quite a number
of specification and programming languages. CIP-L, the language developed in
the cip project at TU München, covers a wide spectrum of levels, ranging from
logical specification to low-level imperative programming. The imperative programming language Ada was developed for the US department of defence. Bernd
designed generics (parameterised packages) and co-authored the formal definition of the language. Bernd designed, with David Luckham, Frieder von Henke,
and Olaf Owe, the specification language Anna for “Annotating Ada Programs”
by assertions and equational properties. For the esprit-project “Program Development by Specification and Transformation” (ProSpecTra), Bernd coordinated the design of another wide-spectrum language, the “ProSpecTra subset
of Anna and Ada” PAnn
d A. Within the same project, a transformation language
TrafoLa was designed based on the object language PAnn
d A. Both languages
have been implemented in the Propsctra system.
The Propsctra experiences have been further elaborated in the project
KORSO. Here, for the first time, the transformation rules themselves and a
calculus for them came into play. The language SpecTraL (co-authored with
Don Sannella) was an ambitious wide-spectrum specification and programming
language, covering all stages of the software development process (from loose
requirements to implementation). SpecTraL was one of the first languages
covering both parameterised specifications and parameterised programs (and
distinguishing them). While SpecTraL has not gone beyond an initial design,
some of its ideas later revived in Casl (see below).
Bernd was also active in forming a European community: he was the coordinator of the COMPASS working group — COMPASS means “a COMPrehensive
Algebraic approach to System Specification and development”. This EU-funded
working group comprised of 19 international partners, and brought together most
European scientists in the area of algebraic specification methods. The volume
[80] was a direct outcome, and an important preparation for, again, Casl.
But before coming to that, the next step was to look for applications and
industrial partners. In the national research project UniForM, a generic framework, instantiated with specific tools for methods to handle communicating distributed systems and real-time requirements, has been developed. Bernd’s vision
of software development by stepwise transformation was realised in the transformation application system TAS. It probably was the first system which had all
transformation rules proved logically correct by a theorem prover.
Specification, Transformation, and Navigation
3
In COMPASS (and also in the IFIP working group 1.3, of which Bernd
is a member), the need for a unification among dozens of existing specification
languages was noticed. Therefore, the Common Framework Initiative (CoFI), an
open initiative with more than 50 researchers from different research groups, was
founded. The outcome was Casl, the Common Algebraic Specification Language.
Bernd played an active role in the process of coming to an agreement on a
common design and was chairman of the CoFI language design task group. In
Bernd’s group in Bremen, several extensions of Casl have been designed, and
the main analysis and proof tool for Casl and extensions, the Heterogeneous
Tool Set Hets, is maintained.
Now the world is not (only) about formal stuff. In the project MMiSS (MultiMedia Instruction in Safe Systems), which provided hypermedia course materials
for teaching formal methods, formal and informal methods have been combined.
For example, a slide or a text can contain a piece of formal specification, and
tools can recognise and exploit this. Currently, Bernd leads the research project
FormalSafe, where the results of the previous efforts are combined into a new tool
DocTip that provides an integration of and a change management for informal
and formal documents. That is, development here not only means transformation
and refinement in the classical formal sense, but also dependency tracing among
evolving formal and informal documents. A proof of concept will be the application of this methodology to the area of safe robotics — this simultaneously
provides a link to the next main research area.
Navigation. Rumour has it that Bernd got interested in the research topic of
Navigation (on the homepage of his working group, this area is called Cognitive
Robotics) because of his wife Dorothea, who is a biologist specialised in the area
of bees. Beside the prominent bee language, the waggle dance, researchers are
also interested in the bee’s astonishing navigational capabilities, and how it is
achieved with such a small brain.
When Bernd entered the area within the FNK interdisciplinary priority program “Cognitive Autonomous Systems” in 1990, artificial neural networks were
one of the major topics in AI research (again). Bernd was one of the founding
members of the Center for Cognitive Sciences (ZKW) in 1992. In the center,
cognitive processes of insects and lower animals were modelled using artificial
neural networks. Bernd became chairman of the DFG Graduate College “Spatial
Orientation and Organisation of Action of Autonomous Systems” that was initiated by the members of the ZKW. Later, Bernd left the ZKW when members
started to conduct research with monkeys. However, within the Graduate College, a device was born that will always be associated with Bernd’s work in the
area of navigation: the Bremen Autonomous Wheelchair “Rolland” (the name
was also invented by Bernd). The research performed with the several instances
of Rolland is described in one of the contributions to this Festschrift.
In 1996, Bernd entered the DFG priority program with the project “Semilocal 3-D Landmarks Based on Image Sequences for Navigation in Dynamic
Environments”, which later was simply called “Robot Navigation”. The project
4
Berthold Hoffmann
ran the full six years of the priority program. Within this project, Bernd developed the concept of the Route Graph as a layered representation for navigational
knowledge. The Route Graph was also the starting point for Bernd to approach
the topics of navigation and robotics from the perspective of Formal Methods.
The DFG project “Formal Fault-Tree Analysis, Specification and Testing of Hybrid Real-Time Systems with Application to Service Robotics” (SafeRobotics)
that Bernd started together with Jan Peleska is another example for this direction.
The success of the priority program resulted in two follow-up projects. One
of them was the DAAD International Quality Network “Spatial Cognition” in
2002-2004, the other, much bigger one is the joint DFG SFB/TR 8 research
center “Spatial Cognition – Reasoning, Action, Interaction” of the universities
of Bremen and Freiburg. It started in 2003 and is one of first trans-regio SFBs
in Germany. Besides being one of the founding members (together with Christian Freksa and Bernhard Nebel), Bernd is principal investigator in four of the
projects in the SFB. In “A1-[RoboMap] – Maps for Robot Navigation”, the concept of the Route Graph was used to navigate the wheelchair Rolland, while it
was and is analysed from a more formal perspective in “I4-[SPIN] – Specification for the Integration of Spatial Concepts”. In “I3-[SharC] – Shared Control
via Dialogues”, dialogues between man and machine in a navigation scenario
are formally modelled. The results of A1-[RoboMap] are currently used to develop an industrial prototype within the transfer project “T1-[Rolland] – Safe
Wheelchair with Driving Assistance”.
Bernd entered a second DFG priority program in 2001, namely “Cooperating
Teams of Mobile Robots in Dynamic Environments”, or short RoboCup, with the
project “Automatic Recognition of Plans and Intentions of Other Mobile Robots
in Competitive, Dynamic Environments” together with the working group of
Christoph Schlieder. This activity resulted in the foundation of the RoboCup
team GermanTeam in the Four-Legged League (now Standard Platform League)
that won the world championship three times.
As the leader of DFKI’s research area Safe and Secure Cognitive Systems,
Bernd started the BMBF Project “SAMS – Safety Component for Autonomous
Mobile Service Robots” in 2006. In the project, a TÜV-certified component is
developed that stops an automated guided vehicle in time to avoid potential
collisions, based on its current speed and the readings of a laser range sensor.
The key idea of the project is to establish the use of formal methods in TÜV
certifications. The project can be seen as a successor of the project SafeRobotics.
Also within the DFKI, the EU project “SHARE-it – Supported Human Autonomy for Recovery and Enhancement of cognitive and motor abilities using
information technologies” started in 2007. In this project, the DFKI develops
a number of assistants based on Rolland and – Bernd’s idea – the Intelligent
Walker (iWalker) to help the elderly.
Ambient-Assisted Living. In 2005, one of the dreams of Bernd became true.
The construction of a new computer science building began, and he became
Specification, Transformation, and Navigation
5
responsible for the design of the new building, now know as “Cartesium”. It
contains a building automation system that covers admission control, automatic
doors, light control, the alarm system, roller shutters, and the heating system,
together with wired and wireless computer networks. The largest lecture room
“Rotunde” (inspired by the “Event-Rotunde Tutzing”) is further equipped with
a comprehensive multi media system. With the student project “Smart Office
Building”, he turned this building into a large laboratory for intelligent office
building services with energy management based on room reservations.
The start of the EU project “SHARE-it – Supported Human Autonomy for
Recovery and Enhancement of cognitive and motor abilities using information
technologies” in 2007 was the definite step into the Ambient Assisted Living
Community. Bernd began the construction of the “Bremen Ambient Assisted
Living Lab” (BaalL), an apartment with an almost complete domestic installation, intelligent household devices and furniture, media control, all within the
Cartesium as a laboratory for Ambient Assisted Living. Inspired by the “Design
for All” principle, the BaalL serves all inhabitants, from children to the elderly
and the physically and mentally impaired. The integration of speech control with
the use of mobility platforms such as Rolland and IntelligentWalker is unique.
He participates in the “Bremen Health Region” (Gesundheitsregion Bremen),
cooperates with the EU sponsored project “i2home”, and is an active member
in the working group “Interface Integration and Interoperability” within “VDEInnovationspartnerschaft AAL” to develop a vision for the world of AAL in the
year 2020.
References
1. B. Krieg. Formal Definition of the Block Concept and some Implementation Models. Master’s thesis, Cornell University, 1971.
2. F. Geiselbrechtinger, W. Hesse, B. Krieg, and H. Scheidig. Lo, the basic layer
of the wide spectrum language L. In W. Brauer, editor, 3. Jahrestagung GI,
number 1 in Lecture Notes in Computer Science, pages 188–197. Springer-Verlag,
Heidelberg, 1973.
3. F. Geiselbrechtinger, W. Hesse, B. Krieg, and H. Scheidig. Language layers,
portability and program structuring. In L.A. Maarsen W.L. v.d. Poel, editor,
Machine Oriented Higher Level Languages, pages 79–104. Elsevier Science, 1974.
4. B. Krieg, R. Gnatz, and R. Höllerer. S0, a low level systems implementation
language. In Günther et al., editor, Proc. Int. Computing Symp. 1973 (Davos),
pages 285–290. North Holland, 1974.
5. B. Krieg. A class of recursive coroutines. In Proc. IFIP Congress 74 (Stockholm),
pages 408–412. North Holland, 1974.
6. B. Krieg-Brückner. Prevention of side-effects from functions. Machine Oriented
Languages Bulletin, IFIP WG 2.4(6), 1977.
7. F. L. Bauer, M. Broy, R. Gnatz, W. Hesse, B. Krieg-Brückner, H. Partsch, P. Pepper, and H. Wössner. Towards a wide spectrum language to support program
specification and program development. SIGPLAN Notices, 13(12):15–24, 1978.
8. F.L. Bauer, M. Broy, R. Gnatz, W. Hesse, and B. Krieg-Brückner. Towards a wide
spectrum language to support program specification and program development.
6
Berthold Hoffmann
9.
10.
11.
12.
13.
14.
15.
16.
17.
18.
19.
20.
21.
22.
In Proc. 5. GI-Fachtagung Programmiersprachen (Braunschweig), pages 73–85.
Springer-Verlag, Heidelberg, 1978.
F.L. Bauer, M. Broy, R. Gnatz, W. Hesse, and B. Krieg-Brückner. A wide spectrum language for program development. In B. Robinet, editor, Program Transformations: Proc. 3rd International Symposium on Programming (Paris), pages
1–15. Dunod; Paris, 1978.
B. Krieg-Brückner. Concrete and Abstract Specification and Program Development by Transformation. Dissertation, TU München; Institut für Informatik,
1978. Technical Report TUM-INFO-7805.
F.L. Bauer, M. Broy, R. Gnatz, W. Hesse, B. Krieg-Brückner, P. Partsch, P. Pepper, and H. Wössner. Towards a wide spectrum language to support program
specification and program development. In F.L. Bauer and M. Broy, editors,
Program Construction, number 69 in Lecture Notes in Computer Science, pages
273–289. Springer-Verlag, Heidelberg, 1979.
Paul F. Albrecht, Phillip E. Garrison, Susan L. Graham, Robert H. Hyerle, Patricia Ip, and Bernd Krieg Brückner. Source-to-source translation: Ada to Pascal
and Pascal to Ada. In Proceedings of the ACM-SIGPLAN symposium on the
Ada programming language, pages 183–193, New York, NY, USA, 1980. ACM.
Manfred Broy and Bernd Krieg-Brückner. Derivation of invariant assertions during program development by transformation. ACM Trans. Program. Lang. Syst.,
2(3):321–337, 1980.
Jean Ichbiah, John G. P. Barnes, J. C. Heliard, Bernd Krieg-Brückner, Olivier
Roubine, and Brian A. Wichmann. Modules and visibility in the Ada programming language. In On the Construction of Programs, pages 153–192. 1980.
Bernd Krieg-Brückner and David C. Luckham. Anna: towards a language for
annotating ada programs. In SIGPLAN ’80: Proceedings of the ACM-SIGPLAN
symposium on Ada programming language, pages 128–138, 1980.
F.L. Bauer, M. Broy, W. Dosch, T. Gnatz, B. Krieg-Brückner, A. Laut, D.C.
Luckmann, M., T.A. Matzner, B. Möller, H. Partsch, P. Pepper, K. Samelson,
R. Steinbrüggen, M. Wirsing, and H. Wössner. Programming in a wide spectrum
language: A collection of examples. Science of Computer Programming, 1(1–
2):73–114, 1981.
J.D. Ichbiah, B. Krieg-Brückner, A. Wichmann, H.F. Ledgard, J-C. Heliard, J.-R.
Abrial, J.P.G. Barnes, M. Woodger, C. Roubine, P.N. Hilfinger, and R. Firth. Reference manual for the Ada programming language: Proposed standard document.
In Duijvestijn and Lockemann [19].
Bernd Krieg-Brückner. Ada and the german pay phone: An illustrative example
of parallel processing. In Duijvestijn and Lockemann [19], pages 122–134.
A. J. W. Duijvestijn and Peter C. Lockemann, editors. Trends in Information
Processing Systems, volume 123 of Lecture Notes in Computer Science. Springer,
1981.
J.D. Ichbiah, B. Krieg-Brückner, A. Wichmann, H.F. Ledgard, J-C. Heliard, J.-R.
Abrial, J.P.G. Barnes, M. Woodger, C. Roubine, P.N. Hilfinger, and R. Firth. Reference Manual for the Ada Programming Language. Number ANSI/MIL-STD1815A-1983 in Ada Companion Series. American National Standards Institute;
Washington, 1983.
B. Krieg-Brückner. Consistency checking in Ada and Anna: a transformational
approach. Ada LETTERS, 3(2):46–54, 1983.
F.L. Bauer, M. Broy, R. Gnatz, W. Hesse, B. Krieg-Brückner, P. Partsch, P. Pepper, and H. Wössner. Towards a wide spectrum language to support program
Specification, Transformation, and Navigation
23.
24.
25.
26.
27.
28.
29.
30.
31.
32.
33.
34.
35.
36.
37.
7
specification and program development (russian translation). In W.H. Arafanova,
editor, Anforderungen und Spezifikationen in der Programmentwicklung; eine Aufsatzsammlung, Matematitscheskoje Objespjetschenije EWM, pages 28–46. MIR;
Moskau, 1984.
B. Krieg-Brückner. Types in the programming language Ada. In M.L. Brodie,
J. Mylopoulos, and J.W. Schmidt, editors, On Conceptual Modelling; Perspectives
from Artificial Intelligence, Data Bases and Programming Languages, pages 385–
408. Springer-Verlag, Heidelberg, 1984.
B. Krieg-Brückner. Language comparison and source-to-source translation. In
P. Pepper, editor, Program Transformation and Programming Environments, volume F8 of NATO ASI Series, pages 299–304. Springer-Verlag, Heidelberg, 1984.
F.L. Bauer, R. Berghammer, M. Broy, W. Dosch, R. Gnatz, F. Geiselbrechtinger,
E. Hangel, W. Hesse, B. Krieg-Brückner, A. Laut, T.A. Matzner, B. Möller,
F. Nickl, H. Partsch, P. Pepper, K. Samelson, and M. Wirsin. The Munich Project
Cip, Vol. 1: The Wide Spectrum Language Cip-L. Number 183 in Lecture Notes
in Computer Science. Springer-Verlag, Heidelberg, 1985.
F.W. von Henke, D.C. Luckham, B. Krieg-Brückner, and O. Owe. Semantic
specification of Ada packages. In J.G.P. Barnes and G.A. Fisher Jr., editors, Ada
in Use, Proc. Ada Int’l Conf. 85 (Paris), Ada Companion Series, pages 185–198.
Cambridge University Press, 1985.
B. Krieg-Brückner. Transformation of interface specifications. In Kreowski [28],
pages 156–170.
Hans-Jörg Kreowski, editor. Recent Trends in Data Type Specification, 3rd Workshop on Theory and Applications of Abstract Data Types, Bremen, November 1316, 1984, Selected Papers, volume 116 of Informatik-Fachberichte. Springer, 1985.
V.D. Donzeau-Gouge, J.-C. Heliard, G. Kahn, B. Krieg-Brückner, and B. Lang.
Formal definition of the Ada programming language. Preliminary version, Honeywell Inc., CII Honeywell Bull and INRIA; Le Chesnay, 1980.
Bernd Krieg-Brückner. Integration of program construction and verification: The
ProSpecTra methodology. In Habermann and Montanari [31], pages 173–194.
A. Nico Habermann and Ugo Montanari, editors. Software Development and
Ada, volume 275 of Lecture Notes in Computer Science. Springer, 1987.
B. Krieg-Brückner. Formalisation of developments: an algebraic approach. In
Rogers [34], pages 491–501.
B. Krieg-Brückner, B. Hoffmann, H. Ganzinger, M. Broy, R. Wilhelm, U. Möncke,
B. Weisgerber, A.D. McGettrick, I.G. Campbell, and G. Winterstein. Program
development by specification and transformation. In Rogers [34], pages 301–312.
M.W. Rogers, editor. Results and Achievements, Proc. Esprit Conf. ’86. North
Holland, 1986.
B. Krieg-Brückner, H. Ganzinger, R. Wilhelm, M. Broy, U. Möncke, B. Weisgerber, A.D. McGettrick, I.G. Campbell, and G. Winterstein. Program development
by specification and transformation in Ada/Anna. In P. Wallis, editor, Ada:
Managing the Transition, Proc. Ada Europe Conf. 86 (Edinburgh), Ada Companion Series, pages 249–260. Cambridge University Press, 1986.
B. Krieg-Brückner. An introduction to Ada language (adaptation to Chinese by
Chen Zushun et. al.). Computer Research and Development, 25(2):37–53, 1987.
B. Krieg-Brückner. Systematic transformation of interface specifications. In L. G.
T. L. Meertens, editor, Program Specification and Transformation, Proc. IFIP
TC2 Working Conf. (Tölz’86), pages 269–291. North Holland, 1987.
8
Berthold Hoffmann
38. D.C. Luckham, F.W. von Henke, B. Krieg-Brückner, and O. Owe. Anna, a Language for Annotating Ada Programs: Reference Manual. Number 260 in Lecture
Notes in Computer Science. Springer-Verlag, Heidelberg, 1987.
39. B. Krieg-Brückner. Algebraic formalisation of program development by transformation. In H. Ganzinger, editor, Proc. European Symp. On Programming ’88,
number 300 in Lecture Notes in Computer Science, pages 34–48. Springer-Verlag,
Heidelberg, 1988.
40. B. Krieg-Brückner. The ProSpecTra methodology of program development. In
J. Zalewski, editor, Proc. IFIP/IFAC Working Conf. on Hardware and Software
for Real Time Process Control (Warsaw), pages 257–271. North Holland, 1988.
41. P. de la Cruz, B. Krieg-Brückner, and A. Perez Riesco. From algebraic specifications to correct Ada programs. In A. Alvarez, editor, Proc. Ada Europe Conf. 89
(Madrid), Ada Companion Series. Cambridge University Press, 1989.
42. B. Krieg-Brückner (ed.). Compass, a comprehensive algebraic approach to system
specification and development: Objectives, state of the art, references. InformatikBericht 6/89, Fachbereich Mathematik u. Informatik; Universität Bremen, 1989.
43. B. Krieg-Brückner. Algebraic specification and functionals for transformational
program and meta-program development. In J. Diaz and F. Orejas, editors, Proc.
TAPSOFT’89 (Barcelona), number 352 in Lecture Notes in Artificial Intelligence,
pages 36–59. Springer-Verlag, Heidelberg, 1989.
44. B. Krieg-Brückner. Algebraic specification with functionals in program development by transformation. In H. Hünke, editor, Proc. Esprit Conf. ’89, pages
302–320. Kluwer Academic Publishers; Dordrecht, The Netherlands, 1989.
45. B. Krieg-Brückner. Program development by specification and transformation
(revised version, invited paper). Technique et Science Informatiques: Special Issue
on Advanced Software Engineering in Esprit, pages 134–149, 1990.
46. B. Krieg-Brückner. Compass, a comprehensive algebraic approach for system
specification and development, Esprit basic research working group 3264. EATCS
Bulletin, 40:144–157, 1990.
47. B. Krieg-Brückner. Program development by specification and transformation. In
G. Kahn, editor, Proc. Summer School on Programming Environments in Esprit
(Sophia Antipolis), 1990.
48. B. Krieg-Brückner. Transformational meta-program development. In M. Broy
and M. Wirsing, editors, Methods of Programming, number 544 in Lecture Notes
in Computer Science, pages 19–34. Springer-Verlag, Heidelberg, 1991.
49. B. Krieg-Brückner, E.W. Karlsen, J. Liu, and O. Traynor. The ProSpecTra
methodology and system: Uniform transformational (meta-) development. In
Søren Prehn and W.J. Toetenel, editors, VDM’91, Formal Software Development
Methods, Vol. 2: Tutorials, number 552 in Lecture Notes in Computer Science,
pages 363–397. Springer-Verlag, Heidelberg, 1991.
50. B. Krieg-Brückner and D. Sannella. Structuring specifications in-the-large and inthe-small: Higher-order functions, dependent types and inheritance in SpecTraL.
In Proc. TAPSOFT’91, number 494 in Lecture Notes in Computer Science, pages
313–336. Springer-Verlag, Heidelberg, 1991.
51. E.W. Karlsen, B. Krieg-Brückner, and O. Traynor. The ProSpecTra system: A
unified development framework. In M. Nivat, C. Rattray, T. Rus, and G. Scollo,
editors, Algebraic Methodology and Software Technology (AMAST’91), Workshop
in Computing, pages 421–433. Springer-Verlag, Heidelberg, 1992.
52. O. Traynor, J. Liu, and B. Krieg-Brückner. Knowledge-based transformational
programming. In In Proc. 4th International Conf. on Software Engineering and
Knowledge Engineering, Capri, Italy, 1992.
Specification, Transformation, and Navigation
9
53. B. Krieg-Brückner. Introduction [to program development by specification and
transformation]. In Hoffmann and Krieg-Brückner [57].
54. Bernd Krieg-Brückner.
A language family for programming and metaprogramming. In Hoffmann and Krieg-Brückner [57], pages 147–148.
55. B. Krieg-Brückner, E.W. Karlsen, J. Liu, and O. Traynor. Uniform transformational development. In Hoffmann and Krieg-Brückner [57].
56. J. Liu and B. Krieg-Brückner. Transformation. In Hoffmann and Krieg-Brückner
[57].
57. Berthold Hoffmann and Bernd Krieg-Brückner, editors. Program Development
by Specification and Transformation, The ProSpecTra Methodology, Language
Family, and System, volume 680 of Lecture Notes in Computer Science. Springer,
1993.
58. B. Krieg-Brückner, J. Liu, B. Wolff, and H. Shi. Towards correctness, efficiency
and reusability of transformational developments. In H. Reichel, editor, Informatik – Wirtschaft – Gesellschaft, 23. GI Jahrestagung 1993, Dresden, Fachgespräch ”Spezifikation und Semantik”, Informatik aktuell, pages 241–252. SpringerVerlag, Heidelberg, 1993.
59. B. Krieg-Brückner, J. Liu, H. Shi, and B. Wolff. Towards correct, efficient and
reusable transformational developments (extended version). In Programmentwicklung durch Spezifikation und Transformation – Bremer Beiträge zum Verbundprojekt Korso (Korrekte Software) [61].
60. B. Krieg-Brückner and W. Menzel. System architecture framework for Korso. In
Programmentwicklung durch Spezifikation und Transformation – Bremer Beiträge
zum Verbundprojekt Korso (Korrekte Software) [61].
61. B. Krieg-Brückner (Hrsg.). Programmentwicklung durch Spezifikation und Transformation – Bremer Beiträge zum Verbundprojekt Korso (Korrekte Software);
Band 2. Informatik Bericht 10/94, Fachbereich Mathematik u. Informatik; Universität Bremen, 1994.
62. B. Krieg-Brückner, J. Liu, H. Shi, and B. Wolff. Towards correct, efficient and
reusable transformational developments. In M. Broy and S. Jähnichen, editors,
Korso: Methods, Languages, and Tools for the Construction of Correct Software.
Final Report, number 1009 in Lecture Notes in Computer Science, pages 270–284.
Springer-Verlag, Heidelberg, 1995.
63. B. Krieg-Brückner. Seven years of Compass. In M. Haveraaen, O. Owe,
and O.-J. Dahl, editors, Recent Trends in Data Type Specification. Proc. 11th
ADT/ Compass Workshop (Oslo 1995), number 1130 in Lecture Notes in Computer Science, pages 1–13. Springer-Verlag, Heidelberg, 1996.
64. Z. Qian and B. Krieg-Brückner. Typed object-oriented functional programming
with late binding. In Proc. 10th European Conf. on Object-Oriented Programming,
number 1098 in Lecture Notes in Computer Science, pages 48–72. Springer-Verlag,
Heidelberg, 1996.
65. Z. Qian and B. Krieg-Brückner. Object-oriented functional programming and type
reconstruction. In M. Haveraaen, O. Owe, and O.-J. Dahl, editors, Recent Trends
in Data Type Specification. Proc. 11th ADT/ Compass Workshop (Oslo 1995),
number 1130 in Lecture Notes in Computer Science, pages 458–477. SpringerVerlag, Heidelberg, 1996.
66. Z. Qian and B. Krieg-Brückner. An approach to object-oriented functional programming. In A. Pnueli and H. Lin, editors, Proc. 1995 Int’l. Workshop on Logic
and Software Engineering, pages 71–89. World Scientific Publishing; Singapore,
1996.
10
Berthold Hoffmann
67. M. Gogolla M. Cerioli, H. Kirchner, B. Krieg-Brückner, Z. Qian, and M. Wolf,
editors. Algebraic System Specification and Development: Survey and Annotated
Bibliography, volume 3 of Monographs of the Bremen Institute of Safe Systems
(BISS). Shaker Verlag, Aachen, 2nd edition edition, 1997.
68. M. Cerioli, A. Haxthausen, B. Krieg-Brückner, and T. Mossakowski. Permissive
subsorted partial logic in Casl. In M. Johnson, editor, Algebraic Methodology and
Software Technology, 6th International Conference, AMAST’97, number 1349 in
Lecture Notes in Computer Science, pages 91–107. Springer-Verlag, Heidelberg,
1997.
69. J. Kollmann, A. Lankenau, A. Bühlmeier, B. Krieg-Brückner, and T. Röfer. Navigation of a kinematically restricted wheelchair by the parti-game algorithm. In
Spatial Reasoning in Mobile Robots and Animals, pages 35–45. AISB-97 Workshop; Manchester University, 1997.
70. S. Werner, B. Krieg-Brückner, H.A. Mallot, K. Schweitzer, and C. Freksa. Spatial
cognition: The role of landmark, route, and survey knowledge in human and robot
navigation. In M. Jarke, K. Pasedach, and K. Pohl, editors, Informatik’97, pages
41–50. Springer-Verlag, Heidelberg, 1997.
71. B. Krieg-Brückner. A taxonomy of spatial knowledge for navigation. In U. Schmid
and F. Wysotzki, editors, Qualitative and Quantitative Approaches to Spatial Inference and the Analysis of Movements. TU Berlin; Computer Science Department;, 1998. Technical Report 98-2.
72. B. Krieg-Brückner, T. Röfer, H.-O. Carmesin, and R. Müller. A taxonomy of
spatial knowledge for navigation and its application to the Bremen autonomous
wheelchair. In C. Freksa, C. Habel, and K. F. Wender, editors, Spatial Cognition,
number 1404 in Lecture Notes in Artificial Intelligence, pages 373–397. SpringerVerlag Berlin, 1998.
73. A. Lankenau, O. Meyer, and B. Krieg-Brückner. Safety in robotics: The Bremen
autonomous wheelchair. In Proceedings of the 5th Int. Workshop on Advanced
Motion Control (AMC ’98), pages 524–529, 1998.
74. T. Mossakowski, Kolyang, and B. Krieg-Brückner. Static semantic analysis and
theorem proving for Casl. In F. Parisi-Pressice, editor, Recent Trends in Algebraic Development Techniques, 12th International Workshop, WADT’97, number
1376 in Lecture Notes in Computer Science, pages 333–348. Springer-Verlag, Heidelberg, 1998.
75. B. Krieg-Brückner. UniForM perspectives for formal methods. In D. Hutter,
W. Stephan, P. Traverso, and M. Ullmann, editors, Applied Formal Methods –
FM-Trends 98. International Workshop on Current Trends in Applied Formal
Methods, number 1641 in Lecture Notes in Computer Science, pages 251–265.
Springer-Verlag, Heidelberg, 1999.
76. B. Krieg-Brückner, J. Peleska, E.-R. Olderog, D. Balzer, and A. Baer. UniForM
Workbench, Universelle Entwicklungsumgebung für Formale Methoden; Schlussbericht, volume 9 of Monographs of the Bremen Institute of Safe Systems (BISS).
Shaker Verlag, Aachen, 1999.
77. B. Krieg-Brückner, J. Peleska, E.-R. Olderog, and A. Baer. The UniForM workbench, a universal development environment for formal methods. In J.M. Wing,
J. Woodcock, and J. Davies, editors, FM’99, Formal Methods. Proceedings, Vol. 2,
number 1709 in Lecture Notes in Computer Science, pages 1186–1205. SpringerVerlag, Heidelberg, 1999.
78. Christoph Lüth, Haykal Tej, Kolyang, and Bernd Krieg-Brückner. TAS and
IsaWin: Tools for transformational program development and theorem proving.
In Finance [79], pages 239–243.
Specification, Transformation, and Navigation
11
79. Jean-Pierre Finance, editor. Fundamental Approaches to Software Engineering (FASE’99) Proceedings, volume 1577 of Lecture Notes in Computer Science.
Springer, 1999.
80. E. Astesiano, H.-J. Kreowski, and B. Krieg-Brückner, editors. Algebraic Foundations of System Specification. IFIP State-of-the-Art Reports. Springer-Verlag,
Heidelberg, 2000.
81. D. Basin and B. Krieg-Brückner. Formalization of the development process. In
E. Astesiano, H.-J. Kreowski, and B. Krieg-Brückner, editors, Algebraic Foundations of System Specification, IFIP State-of-the-Art Reports. Springer-Verlag,
Heidelberg, 2000.
82. T. Mossakowski, A. Haxthausen, and B. Krieg-Brückner. Subsorted partial
higher-order logic as an extension of Casl. In C. Choppy, D. Bert, and P. Mosses,
editors, Recent Trends in Algebraic Development Techniques, 14th International
Workshop, WADT’99, Bonas, France, volume 1827 of Lecture Notes in Computer
Science, pages 126–145. Springer Verlag, London, 2000.
83. S. Werner, B. Krieg-Brückner, and T. Herrmann. Modelling navigational knowledge by route graphs. In C. Freksa, C. Habel, and K.F. Wender, editors, Spatial
Cognition II, volume 1849 of Lecture Notes in Artificial Intelligence, pages 295–
317. Springer-Verlag, Heidelberg, 2000.
84. E. Astesiano, M. Bidoit, B. Krieg-Brückner, H. Kirchner, P. D. Mosses, D. Sannella, and A. Tarlecki. Casl - the common algebraic specification language.
Theoretical Computer Science, 286:153–196, 2002.
85. B. Krieg-Brückner, D. Hutter, A. Lindow, C. Lüth, A. Mahnke, E. Melis, P. Meier,
A. Poetzsch-Heffter, M. Roggenbach, G. Russell, J.-G. Smaus, and M. Wirsing.
Multimedia instruction in safe and secure systems. In M. Wirsing, D. Pattinson,
and R. Hennicker, editors, Recent Trends in Algebraic Development Techniques,
volume 2755 of Lecture Notes in Computer Science, pages 82–117. SpringerVerlag, Heidelberg, 2003.
86. A. Lankenau, T. Röfer, and B. Krieg-Brückner. Self-localization in large-scale
environments for the Bremen autonomous wheelchair. In C. Freksa, W. Brauer,
C. Habel, and K. F. Wender, editors, Spatial Cognition III, number 2685 in Lecture Notes in Artificial Intelligence, pages 34–61. Springer-Verlag Berlin, 2003.
87. CoFI Language Design Group, B. Krieg-Brückner, and P.D. Mosses. Casl syntax
/ Casl summary. In P.D. Mosses, editor, Casl Reference Manual, volume 2960
of Lecture Notes in Computer Science. Springer-Verlag Heidelberg, 2004.
88. B. Krieg-Brückner, U. Frese, K. Lüttich, C. Mandel, T. Mossakowski, and R. Ross.
Specification of an Ontology for Route Graphs. In Freksa et al. [94], pages 390–
412.
89. B. Krieg-Brückner, A. Lindow, C. Lüth, A. Mahnke, and G. Russell. Semantic
interrelation of documents via an ontology. In G. Engels and S. Seehusen, editors, DeLFI 2004, Tagungsband der 2. e-Learning Fachtagung Informatik, 6.-8.
September 2004, Paderborn, Germany, volume 52 of Lecture Notes in Informatics,
pages 271–282. Springer-Verlag, Heidelberg, 2004.
90. K. Lüttich, B. Krieg-Brückner, and T. Mossakowski. Tramway Networks as Route
Graphs. In E. Schnieder and G. Tarnai, editors, FORMS/FORMAT 2004 – Formal Methods for Automation and Safety in Railway and Automotive Systems,
pages 109–119, 2004.
91. K. Lüttich, T. Mossakowski, and B. Krieg-Brückner. Ontologies for the Semantic
Web in Casl. In J. L. Fiadeiro, P. Mosses, and F. Orejas, editors, Recent Trends in
Algebraic Development Techniques, 17th International Workshop (WADT 2004),
12
92.
93.
94.
95.
96.
97.
98.
99.
100.
101.
Berthold Hoffmann
volume 3423 of Lecture Notes in Computer Science, pages 106–125. SpringerVerlag Berlin, 2005.
A. Mahnke and B. Krieg-Brückner. Literate ontology development. In Robert
Meersman, Zahir Tari, and Angelo Corsaro et al., editors, On the Move to Meaningful Internet Systems 2004: OTM 2004 Workshops, volume 3292 of Lecture
Notes in Computer Science, pages 753–757. Springer-Verlag Berlin, 2004.
R. Ross, H. Shi, T. Vierhuff, B. Krieg-Brückner, and J. Bateman. Towards dialogue based shared control of navigating robots. In Freksa et al. [94], pages
478–499.
Christian Freksa, Markus Knauff, Bernd Krieg-Brückner, Bernhard Nebel, and
Thomas Barkowsky, editors. Spatial Cognition IV: Reasoning, Action, Interaction,
volume 3343 of Lecture Notes in Computer Science. Springer, 2004.
Bernd Krieg-Brückner. Towards multimedia instruction in safe and secure systems. In Hutter and Stephan [96], pages 379–395.
Dieter Hutter and Werner Stephan, editors. Mechanizing Mathematical Reasoning, Essays in Honor of Jörg H. Siekmann on the Occasion of His 60th Birthday,
volume 2605 of Lecture Notes in Computer Science. Springer, 2005.
Bernd Krieg-Brückner and Hui Shi. Orientation calculi and route graphs: Towards
semantic representations for route descriptions. In M. Raubal, H. Miller, A. Frank,
and M. Goodchild, editors, Geographic Information Science - Fourth International
Conference, GIScience 2006, volume 4197 of Lecture Notes in Computer Science.
Springer-Verlag Berlin, 2006.
Hui Shi and Bernd Krieg-Brückner. Modelling route instructions for robust
human-robot interaction on navigation tasks. International Journal of Software
and Informatics, 2(1):33–60, 2008.
Hui Shi and Bernd Krieg-Brückner. Qualitative semantic representation of spatial
knowledge in dialogue systems. KI Zeitschrift, 3:59–61, 2008.
B. Krieg-Brückner, H. Shi, C. Fischer, T. Röfer, J. Cui, and K. Schill. Welche
Sicherheitsassistenz brauchen Rollstuhlfahrer? In 2. Deutscher AAL-Kongress.
VDE-Verlag; Berlin-Offenbach, Germany, 2009.
B. Krieg-Brückner, B. Gersdorf, Mathias Döhle, and Kerstin Schill. Technik für
Senioren in spe im Bremen Ambient Assisted Living Lab. In 2. Deutscher AALKongress. VDE-Verlag; Berlin-Offenbach, Germany, 2009.
13
14
Glückwünsche zum Sechzigsten
Wolfgang Bibel
Technische Universität Darmstadt
eAdresse: [email protected]
Lieber Bernd,
Zu Deinem sechzigsten Geburtstag wünsche ich Dir von Herzen alles Gute.
Vor allem Gesundheit, die in diesem erreichten Alter immer wichtiger wird.
Da ich Dir ja über zehn Jahre voraus bin, kann ich Dir aus eigener Erfahrung
verheißen, daß auch die kommenden Jahre so erfüllt sein können wie die vorangegangen. Eigentlich sogar noch erfüllter, weil man ja gelernt hat, Wichtiges
von Unwichtigem zu unterscheiden und sich auf das zu konzentrieren, auf das es
einem persönlich wirklich ankommt.
Du bist Anfang der Siebziger Jahre kurz nach mir aus Amerika zurückgekommen. Von da an waren wir bis zu Deiner Rückkehr in die USA ein Jahrzehnt im
gleichen Institut an der TU in München. Ich kann mich noch gut und gerne an
die schönen Gespräche bei der täglichen Kaffeerunde erinnern. Dabei konnte ich
lernen, daß Du von anderem Holz geschnitzt warst als die vielen “Bäuerlinge”. Die
Übernahme des Namens Deiner Frau in Deinen Namen anläßlich Eurer Hochzeit
mußte im damaligen Bayern als regelrechter Kulturschock empfunden werden.
Mir hat dies als sichtbares Zeichen Deines aufgeschlossenen und unabhängigen
Geistes außerordentlich imponiert.
Die Verhältnisse waren in München leider so, daß wir fachlich so gut wie
keinen Kontakt pflegen konnten. Dabei hätten wir aufgrund unserer gemeinsamen Interessenlage beispielsweise an formalen bzw. logischen Methoden der
Programmierung, schon damals fruchtbar zusammenarbeiten können. Durch den
Kleingeist pseudomächtiger Institutsdirektoren wurde dies ausgeschlossen. Ich
war mir aber immer sicher, daß Du die mir seinerzeit von Bauer und seinen
Bäuerlingen zuteil gewordene tyrannische Behandlung in Deinem Inneren nie
gebilligt hast. Schon das war mir damals Hilfe und gab mir Stärke zum Durchhalten, wofür ich Dir noch heute dankbar bin.
Inzwischen hast Du Dich durch Dein Engagement im DFKI selbst offen zu der
Fachrichtung bekannt, derentwegen ich seinerzeit so unterdrückt wurde, nämlich
der KI. So sind wir zwar jeweils eigene, im Grunde aber doch analoge Wege
gegangen. Umso mehr freue ich mich, daß Dein Weg einen so außerordentlich erfolgreichen Verlauf genommen hat, ohne daß Du Dich dafür je verbiegen mußtest.
Dafür bezeuge ich Dir meinen höchsten Respekt.
15
16
Wolfgang Bibel ist derzeit Professor im Ruhestand für Intellektik am Fachbereich für Informatik
der Technischen Universität Darmstadt. Zugleich ist
er Adjungierter Professor an der University of British
Columbia in Vancouver, Kanada. Seine über 250 Publikationen, darunter etwa 20 Bücher, befassen sich mit
einer Vielfalt von Themen der Künstlichen Intelligenz bzw. Intellektik, mit deren Anwendungen und mit
Technologieperspektiven. Er gilt als einer der Mitbegründer der KI in Deutschland und Europa und wurde
für seine wissenschaftlichen Leistungen mit zahlreichen
Ehrungen bedacht. Er war 1972-1982 an der Technischen Universität München Kollege von Bernd KriegBrückner.
Dear BKB,
Ever since Lutz won my heart at the Freimarkt on April 1st, 2003, I tried
to be a source of inspiration for you, like Erwin (Sr) was and is for your group.
Unfortunately, without my consent Lutz handed me over to Markus and sent me
cold heartedly into exile in the backwaters of Wales. Frankly, I must say Markus
is not playing in the same league as you when it comes to humour, beards, and
generosity. For example, at a recent conference in Norway (where Markus only
got a paper in thanks to my overwhelming reasoning power) he would not even
let me sip on his Aquavit. God bless Holger, that in South Africa I got a little
glass of wine at the conference dinner.
Now I heard, dear BKB, that you are having your birthday. I persuaded
Markus to take me to your celebrations. Please, don’t let him take me back with
him! I have already earned scientific acknowledgements for enhancing Markus’
and Holger’s knowledge on product lines, for pointing out the possibility of onthe-fly evaluation, for contributing deep insights into the nature of fixed point
induction, for advice on semantical questions about Csp, and for being an entertainment to the children in the Candy-Puzzle. I was also engaged in the supervision of doctoral students in Swansea, whom I watched over their shoulders
whilst preventing them from procrastinating too much, and who unboundedly
appreciate me for my ongoing pastoral support to Markus.
I am flexible and can work in various areas, catching bees in full flight during
their navigation, constructing autonomous wheelchairs for French frogs who lost
their legs, and even standardizing these strange formal methods which you love
so much. Best of all, my services are cheap compared to those of Erwin (Sr).
In deep affection, Erwin R. Catesbaiana (Jr)
17
herzlichen glückwunsch, bernd! martin
2008-12-13 google bilder: bernd krieg brückner
Martin Gogolla is professor for Computer Science at
University of Bremen, Germany, heading the Database
group and a colleague of Bernd Krieg-Brückner since
1994. He knows Bernd Krieg-Brückner since 1984
where he met him in Bremen at the Workshop on
Abstract Data Types. His interests include software
development with object-oriented approaches, formal
methods in system design, semantics of languages, formal specification and international computer science
standards (e.g., OCL 2.0 as part of UML 2.0). Both
have collaborated in international (e.g., COMPASS)
and national projects (e.g., MMISS).
18
Grußwort
Modar Ibraheem
Universität Bremen
Lieber BKB,
ich erinnere mich jetzt spontan an den ersten Treff mit dir. Damals hatte ich
alle deutsche Wörter und Sätze vorbereitet, die man wahrscheinlich in der ersten
Rede brauchen könnte.
Du hattest angefangen, mit mir in Deutsch zu reden, hatte ich aber kaum nichts
verstanden. Nach ’Switching into English’ hatte ich nochmal zu wenig gekriegt.
In diesem Moment hätte ich dich fragen können, ob du Arabisch sprechen könntest, die Sprache deines Gesichts hatte mich aber das nicht brauchen lassen.
Diese Sprache konnte man ja einfach verstehen und schnell lernen und war jedoch keine schreckliche deutsche Sprache.
Mein Deutsch hattest du möglicherweise ein ’Unikat’ gesehen, welches neben
Mark Twain’s Unikat in Heidelberger Schloss aufgenommen werden sollte. Diese
Kunst hatte mich aber nicht zu viel wie er gekostet und kann dem Besucher
sogar kostenlos verkauft werden.
Nach diesem Treff sollte der neue 26-järige Doktorand seinen 58-järigen Betreuer
duzen und sogar ohne Titel anreden. Das war ja zu viel und sehr überraschend
für einen orientalischen Mensch.
Bevor ich nach Deutschland gekommen war, hatte ich versucht, das Alphabet
dieser Gesichtssprache durch dein Foto im Netz kennenzulernen. Damals hatten
der weiße Bart und die schwarze Brille keinen klaren Eindruck bei mir hinterlassen. Nur drei Wörter von dir hatten verdeutlicht, was sich hinter der Brille
versteckt. ’Nur kein Stress’ hattest du mir nach einer Verspätungsgeschichte gesagt. Dieser und nur dieser Zauberstab konnte mich damals beruhigen.
Nun, nach ca. zwei Jahre an deiner Arbeitsgruppe kann man viele schöne
Bilder, Erinnerungen, Momente der Schwäche, anderen der Hoffnung und Stärke
und viel mehr erzählen. Diese zwei Lebensjahre waren so lang wie ein deutscher
Satz in einer Zeitungsspalte und so kurz wie Augenschlag.
Na ja, ansonsten gibt es eigentlich viel mehr zu sagen, ich will aber am Ende
dir und deiner Familie alles Gute - vor allem jedoch Gesundheit wünschen.
Ich hoffe, dass deine Wünsche für das neue Lebensjahr in Erfüllung gehen.
Modar Ibraheem, has received his bachelor degree
from the Department of Electrical Engineering, AlBaath University-Syria in 2003. After graduation he
worked as assistant teacher in Albaath University. He
then got a scholarship from the ministry of Higher
Education in Syria in order to get his PhD degree.
Currently he is a PhD student in Bernd-Krieg Brückners group. His proposal focus on image processing and
sensor fusion for robot navigation.
19
Es war einmal ein junger PhD-Student, dessen Doktorarbeit in der rauchigen Stadt neigte sich seinem Ende zu, und Geld war auch keines mehr
da. So ging er traurig seines Weges, und wußte nicht wohin, als er einen
Professor am Wegesrand traf. “Warum so traurig, junger Freund?” fragte ihn der freundliche Graubart. “Ach, mein PhD ist nun bald am Ende,
und nun bin ich auf der Suche nach neuen Abenteuern”, antwortete unser junger Doktorand. “Dann komm mit mir nach Bremen. Ich habe
dort eine offene Stelle, auf der kannst du forschen und lehren nach deines Herzens Lust. Etwas besseres als den Tod findest du überall.”
So oder ähnlich1 war es, als ich mich damals um eine Stelle im Forschungsprojekt UniForM an der Universität Bremen bewarb. Aus Bremen war mir im
wesentlichen nur der SV Werder geläufig, aber die Universität kannte ich nur
vom Hörensagen. Dreizehn Jahre später bin ich immer noch in Bremen, und das
hat natürlich wesentlich mit der Person von Bernd zu tun. Wie wenige andere
versteht er es, in seiner Arbeitsgruppe ein Klima zu schaffen, in dem ein selbstbestimmtes, und damit wissenschaftlich befriedigendes Arbeiten ermöglicht wird.
Er hat in seinen Mitarbeitern immer sowohl den Menschen als auch den eigenständigen Wissenschaftler gesehen, nie nur bloße Erfüllungsgehilfen für seine
wissenschaftlichen Ambitionen. Besonders beeindruckend finde ich, dass er immer noch vor Ideen sprüht, und neue Forschungsrichtungen in Angriff nimmt,
anstatt wie es andere an seiner Stelle vielleicht tun würden, mit immer denselben
Arbeiten in neuem Aufguss der Pension entgegenzuforschen.
Ich wünsche Dir noch viele weitere Jahre erfüllten und erfolgreichen Schaffens!
1
Schriftstellerische Freiheiten eingeschlossen.
20
21
Christoph Lüth obtained his Diplom in Computer
Science from the Technical University of Berlin, and
his Ph.D. from the University of Edinburgh. After his
Ph.D. he went to Bremen, where he received his Habilitatition. Since 2006, he is the vice director of newly
established Safe and Secure Cognitive Systems research lab of the German Research Center for Artificial
Intelligence (DFKI).
He met Bernd in one of the COMPASS meetings, in
which group both his supervisor from Edinburgh, Don
Sannella, and Bernd were active. This lead to his applying for the UniForM position in Bremen, and the
rest, as they say, is history.
Dear Bernd,
I have joined your group in 2002 as a developer for the Casl Tool Set (Cats)
and the Heterogeneous Casl Tool Set (Hets) within the Multiple project. From
the very beginning you offered me to attend various summer schools, workshops
and conferences, where I interdisciplinarily experienced topics like algebraic specification, spatial cognition and formal methods. During my rehearsals of presentations at the Extra Meeting you gave me valuable hints on the structuring of
my presentations and on my speeches.
In 2003 the transregional collaborative research center “Spatial Cognition”
opened in Bremen and Freiburg and you gave me the chance to work in Till’s
and your project SPIN, where we developed formal representations of spatial
relations based on your Route Graph concepts within Casl. During this four year
period in which I worked within the research center I had various opportunities
to develop my own way of contributing different aspects of using Casl for the
structured development of ontologies.
In 2006 I decided to leave your group for working in the industry. Still, you
encouraged me to finish my dissertation in the first half of 2007 by reading,
discussing and commenting my drafts.
For your upcoming birthday and the following years at the University of
Bremen I wish you all the best.
Yours sincerly, Klaus
Klaus Lüttich was a researcher / Ph.D. student at
the research group of Bernd Krieg-Brückner from 2002
until 2007. Since 2007 he is working as a product manager for bremen online service (bos) GmbH & Co. KG
and since 2008 he is also the head of team of the team
Professional Services. Since 1999 bos is a software enterprise that develops and implements e-government
solutions. The main focus is on secure data transport
and digital signatures.
22
Warmest greetings, Bernd, on the occasion of your 60th birthday!
If my memory serves me well, I first met you at a WADT, about 25 years
ago: the 3rd WADT in Bremen (1984)? At the time, my main research area
was the semantics of programming languages. Actually, our paths crossed in
that connection even earlier, in 1979, while you were involved in the design and
formal definition of Ada, since my semantics implementation system (SIS) was
initially being used in connection with its denotational semantics.
It wasn’t until the end of the 1980s that I started to meet you more frequently:
you were coordinator for ESPRIT WG COMPASS, and you kindly invited me to
join. I guess we share many good memories from the joint WADT/COMPASS
meetings in the early 1990s – Dourdan, Caldes de Malavella, S. Margherita,
Sintra, and Oslo are hopefully all still as fresh in your mind as in mine?
Although it was good to meet you at workshops so often, I was fortunate
to develop a much closer relationship with you from 1995, when we both became deeply involved in starting CoFI, the Common Framework Initiative. Your
insight and energy as coordinator of the CoFI Language Design Task Group
were crucial to the success of the initiative, leading to the design of Casl, the
Common Algebraic Specification Language. When the time came to choose the
name for the new language, however, you wrote “I don’t like [the name] Casl
very much as it sounds like ‘little cheese’ in German”! (characteristically adding
“of course I will bend to overwhelming majority, though”). Fortunately, the accepted pronunciation of Casl is now ‘castle’, so I hope that the outcome of all
those Language Design meetings doesn’t sound so cheesy to you any more. . .
Best wishes for the years to come!
Peter Mosses, Swansea
Peter Mosses is professor at the Department of Computer Science, Swansea University.
23
Pencil-sharpener
Frieder Nake
Department of Mathematics and Computer Science, Universität Bremen
We don’t really know what makes us connect a specific set of habits or
characteristics with a person. In trivial cases, it is easy, like: He is even shorter
than I am. Or: He looks as if he was heavier than me. His hair is definitely a
hundred times more than mine, let alone his beard. Of a different kind is: He
dares getting up on stage to sing a song.
I have never seen a person whose leg is constantly vibrating up and down at
the speed and persistence this man is forcing his leg to go. As if it was torture
to him to just sit on a chair for, say, three minutes without walking. But then,
why does he sit when he is lecturing instead of standing up and moving back
and forth in front of class? Should someone tell him what his leg is doing? But
what for? He may not be aware of the habit. Telling him might be attacking his
identity. The typical BKB leg is just one of those things.
Prior to his arrival at our progressive computer science department, I learned
that his scientific background was the Ada project in Munich. What a difficult
situation, I thought, would this create! Ada was about the most horrible development in programming languages that you could think of by the time. Tony
Hoare had convincingly warned the software world. I was doing a seminar on
Ada, where we tried to understand the discrepancy between the DoD’s financing
this project and their open attitude in publicizing each step and decision.
I have never told BKB how important in my life his appearance was. His job
description was ”programming languages and compiler”, or similar. This fact
helped me to drop this aspect from my job’s definition and fully concentrate
on computer graphics. Soon after, BKB moved into software engineering, from
where he jumped to cognitive science. This was a change of great importance in
his academic career, I believe. I thought he was crazy. However, I learned from
his example that academia can be a really fast moving field. All it needs is that
you are young, daring, and ambitious enough, and that not too many are around
to contest you.
I also learned from him that you must, at the university of the late 20th
century, be able to raise lots of money. BKB became an outstanding figure for
me because he did everything exactly the opposite way I would do it.
There was a time, dear friendly face, when we shared a number of students
between us. I enjoyed that time because it brought me close to an area of which
I would otherwise have had a vague idea only. As so many processes in a life’s
time, this came to an end when fresh young professors arrived, and BKB switched
again, to robotics and something they call soccer.
I could go on like this, chatting about the memories I have of this remarkable
— and controversial — appearance who comes under the sign, “BKB”. The
whole world of programming probably knows him by this sign. I may be the
24
25
only one in possession of a pencil sharpener from him. Through this precious
little gift he told me a surprising lesson: Es gibt sie noch, die guten Dinge. I wish
you can keep up this belief in times ahead of us.
Frieder Nake is a professor and a colleague of Bernd
Krieg-Brückner at the computer science department,
University of Bremen. He is interested in computer
graphics, computer art, digital media, theory of computer science, and has contributed to all these fields.
Since 2005 he has also been teaching at University
of the arts, Bremen. His academic career took him
from Stuttgart to Toronto, Vancouver, and Bremen,
and from mathematics to art and semiotics.
#""
!""
26
27
Zhenyu Qian received his Ph.D. in Computer Science from Bremen University in 1991 and his Habilitation degree in 1996 under supervision of Bernd-Krieg
Brückner. After that, he worked as Privat Dozent for
a short period, before he moved to Kestrel Institute
in California to work as Staff Computer Scientist. In
2000, he joined a startup company XDegrees working on peer-to-peer storage software system. After the
company was acquired by Microsoft, he moved on cofounding a startup company EPIN Technologies Ltd.
in Silicon Valley and China and served as Vice President, first focusing on WiFi technologies and business, then transitioned to computer media and advertisement technologies and business. In 2005, Zhenyu
Qian joined Microsoft. He currently serves as Principal Group Manager responsible for product strategic
planning, development and deployment in Mobile, Local and Map Search areas in China, Japan and Korea.
Pedro de la Cruz Ramos
Dear Bernd,
It is a pleasure and a honour for me to give you my congratulations on the
occasion of your 60th birthday.
I had the luck to participate with you in the PROSPECTRA Project, as
part of the ALCATEL spanish team, and I still remember that period as one
of the best times in my life, not only because it gave me the possibility to
participate in an exciting and advanced research work, but also because I had
the opportunity to meet extraordinary people and make excellent friends; among
them, and specially, you.
I still remember how your presence, labour, personality, intelligence and wisdom, impregnated and inspired all the aspects of the project. In particular for
me, a young and inexpert engineer by that time, the experience, and notably
your influence, was decisive not only for my professional career, but also for my
personal life. Our collaboration has deeply and positively influenced all aspects
of my further life.
Me and my family want to join your colleagues, friends and relatives, to
transmit you our best wishes for the occasion as well as for your current and
future work and personal life. We also want to make them extensive to all your
family and friends. We wish you peace, joy and happiness in all aspects of your
life. You deserve it.
Your disciple and friend,
Pedro
Pedro de la Cruz Ramos obtained his degree in
Telecommunications Engineering by the Polytechnical
University of Madrid (UPM) in 1982. Then he worked
for some years in the UPM, first as granted researcher
and then as assistant professor. In 1987 he joined Alcatel Standard Eléctrica (then renamed Alcatel España),
where he worked in several positions related to Software Engineering, his first job being as Local Team
Leader in the SPRIT Prospectra Project, in collaboration with the University of Bremen under the direction and leadership of Bernd Krieg-Brückner. He left
Alcatel in 2003 to stablish its own company, Cuántalis
Europa S.A., dedicated to Telecom Expenses Management (TEM). He left Cuántalis in 2008 to join ALTENTIC, where he currently works as Consultant on Information and Communication Technologies. He is currently assigned to the Multiplattform Software Management System (SGS) project in ISBAN, the Banking Software Engineering company of the spanish Santander Bank Group.
28
Sylvie Rauer
University of Bremen, Email: [email protected]
Aujourd’hui est un grand jour de fête en ton honneur, Bernd, pour tes 60
ans ! C’est un grand plaisir pour moi de participer à cette grande fête lors de
laquelle tous pourront te temoigner tout le respect et l’affection qu’ils eprouvent
pour toi. Permets-moi de t’adresser mes voeux les plus sincères, une sante de fer
et une longue, longue. longue vie !
Cordialement
Sylvie
Sylvie Rauer depuis 15 ans j’ai la chance de travailler
dans ton équipe et j’en suis fière et très heureuse.
Responsable du secrétariat je m’efforce de remplir les
tâches administratives le mieux possible, d’organiser
tous les voyages d’affaires du groupe, ainsi que tous
les congrès. J’ai grand plaisir à aider et à accueillir
les chercheurs étrangers qui rejoignent notre équipe.
Ce travail est terriblement polyvalent et permet dêtre
autonome ce qui me plâit beaucoup, j’apprécie aussi
la confiance que tu me témoignes. En somme j’aime
beaucoup mon travail.
29
Dear Bernd,
I remember meeting you 25 years ago at a workshop in Brussels on formal
methods in Ada, and then a year later at WADT in Bremen. I guess these
encounters were what prompted you to suggest that I apply for a chair in Bremen
a few years later. I finally arrived in Bremen 19 years ago.
Despite the passage of time, I remember very well how helpful you were
during that period of my life and how energetic you were in smoothing the way
for me. In the end I decided to return to Edinburgh, and I think that was the
right decision for me, but still it was an interesting and important experience
and I don’t regret it at all. I am grateful for your suggestion that I apply for the
position and for all your help along the way.
I also enjoyed our collaboration at that time and since, on Prospectra,
Spectral, Compass and CoFI/Casl. The little I witnessed of your coordination of the Prospectra project was my first introduction to European project
management. I particularly remember an impressive project plan on your office
wall, with perhaps 100 tasks and dependencies between them. I thought it was
a joke, but you told me with a straight face: no, all of the tasks up to this line
here had been completed, and the project team was now working on these tasks
over here. I remembered that diagram when I coordinated my own European
project, Mobile Resource Guarantees, in 2002–2005. Here was my project plan
for this small (two-site) project, not including the administrative tasks:
1a
1f
1c
1b
1d
10a
1e
2a/b
7d
10b
7a
2c
2d
7c
7e
2f
2e
3e
3a
3b
3c
3d
5a/b
5c
8a
8b
8c
8d
3f
8e
6b
4d
4a
6a
4b/c
4e/f
6g
9b/c
4g
This turned out to be slightly too many tasks in the end for two partners, and
too many deliverables, but with the focus on tasks we did end up achieving a
lot more than we would have done otherwise.
I wish you all the best on your 60th birthday!
Don Sannella, Edinburgh
30
31
Don Sannella did his PhD at Edinburgh University
in 1982 where he remained until 1990 when he took
a position as Professor at the University of Bremen.
He decided to return to Edinburgh in late 1990, however, and has remained there ever since. He was briefly
involved with Bernd’s PROSPECTRA project in the
late 1980s, and during his time in Bremen he collaborated with Bernd on the design of the SPECTRAL
specification language. He was a member of the ECfunded COMPASS working group, which Bernd coordinated, and collaborated with Bernd and many others
on the design of the CASL specification language. His
research interests include functional languages, mechanised reasoning, and foundations for algebraic specification and formal software development including correctness of modular systems. His recent research has
focused on security and static analysis. He is editor-inchief of the journal Theoretical Computer Science, vice
president of the European Association for Theoretical Computer Science, and served as ETAPS steering
committee chairman from its inception until 2001. He
is currently Head of the Informatics Graduate School
in Edinburgh with overall responsibility for PhD education.
To Bernd Krieg-Brückner
on the Occasion of his 60th Birthday
Andrzej Tarlecki1,2
1
2
Institute of Informatics, University of Warsaw
Institute of Computer Science, Polish Academy of Sciences
Lieber Herr Professor Krieg-Brückner,
Dear Bernd,
Among many abilities I have always admired you for are your extensive linguistic skills — unfortunately, mine are so much below that these greetings just
have to be in my (un)Polished English, sorry!
It used to be that only our old professors were reaching the age when one
organises festschrift events for them. These days though, it so happens that
younger and younger colleagues celebrate their impressive anniversaries. I am
not quite sure why, when and how this happened, we all are still very young!
After all it has been less than 40 years since you started your scientific work,
and only about a quarter of century since we met. I wondered when exactly the
latter happened, and I recalled the 3rd International Workshop on Theory and
Applications of Abstract Data Types in Bremen in cold November 1984. It must
have been then. And then there have been numerous further ADT workshops,
other conferences and events, seminars, visits, project. Two projects have to
be mentioned: work within COMPASS and on CASL was marked with so many
meetings, presentations, heated scientific discussions, and even occasional good
quarrels. They generated a great deal of ideas, some of them to materialise in
good research, results and publications. All these owe so much to you! Not to
worry, I refrain from spelling out the technical achievements of the past, even if
they still influence what we do today.
Instead, let me hasten to add that all our meetings came not only with serious
joint work, but also with lighter (even if sometimes heavy) dinners, drinks, chats,
holiday stories, discussions about everything. This social side is necessary to
make our live of travelling academics bearable. Your good humour, optimism
and excellent taste always marked such occasions as well, and made them so
enjoyable and memorable!
My congratulations to you for many years of excellent work and excellent
humour, and very best wishes for many more even better ideas to come! And, in
the best of my selfish interest, for many more occasions to meet, work and have
good time together!
Mit herzlichen Grüßen,
Andrzej Tarlecki
32
33
Andrzej Tarlecki received his PhD in 1982 and DSc
(habilitation) in 1987 from the Institute of Computer
Science of the Polish Academy of Sciences, where he
has been employed at various levels from Researcher
to Professor (since 1997: part-time employment). Since
1992 his main affiliation is with the Institute of Informatics of the University of Warsaw (since 2000 as
Professor ordinarius; Director of the Institute in 19962005). He received the scietific title of Professor of
mathematical sciences from the President of Poland
in 1998.
His research addresses various aspects of the theory
of software specification, verification and development,
and related topics in logic, universal algebra and category theory. This includes generalization, clarification
and further elaboration of some fundamental ideas of
algebraic specification and program development, design of high-level specification languages Extended ML
and CASL, and work on the theory of institutions and
its applications.
Grußwort zum 60. Geburtstag
Mattias Werner
Fachbereich Mathematik und Informatik, Universität Bremen
[email protected], http://www.informatik.uni-bremen.de/˜mawe/
Lieber Bernd,
als einer der dienstältesten Mitarbeiter Deiner Arbeitsgruppe möchte auch ich Dir alles
Gute zu Deinem 60. Geburtstag wünschen. Im Jahr 1988 begann mein Hauptstudium
bei Dir im Projekt PILS, dann folgte unsere Diplomarbeit GIOTTO und im Jahr 1991
wurden Michael Fröhlich und ich WIMI in Deinem Forschungsprojekt KORSO. Dabei
folgtest Du wie wir uns sagen ließen ganz der Tradition, Mitarbeiter immer zu zweit einzustellen. Es entstand unsere Software daVinci, die mittlerweile uDraw(Graph) heißt.
In der damals ca. 10 Mitarbeiter umfassenden AG übernahm ich bald die Betreuung
des Rechnernetzes, zunächst mit zwei Kollegen und dann als Wissenschaftlich Technischer Angestellter eigenverantwortlich. In dieser langen Zeit hat sich viel verändert.
War es zu Beginn selbstverständlich, nur Rechner und Server der Firma Sun mit deren
eigenem Betriebssystem zu kaufen, sind heute die für uns damals undenkbaren Laptops auf PC Basis mit Linux und sogar Windows selbstverständlich. Sun findet man nur
noch im Serverraum. Während damals unser Server gigantische 500 MB Plattenplatz
hatte, haben wir heute das TB längst überschritten. Einzig konstant dabei bliebst Du
mit Deinem Sekretariat, wo immer nur Apple Macintosch in Frage kam. Angefangen
in der 5. Ebene des MZH zog die Arbeitsgruppe bald in die 8. Ebene, doch für Dich
war es immer das Ziel ein eigenes Gebäude zu konzipieren und so haben wir heute
ein technisch sehr innovatives Gebäude Cartesium, dessen teilweise von Dir erdachte Technik so manchem Handwerker nur ein ungläubiges Staunen entlockt hat. Heute
umfasst die AG fast 40 Personen an der Universität Bremen und am DFKI. Die Liste
der Ehemaligen ist fast genauso lang. Ich habe stets bewundert, mit welchem Einsatz
und Geschick Du es immer wieder geschafft hast, neue Finanzierungen für die Stellen
aller Mitarbeiter aufzutreiben, so dass die AG wachsen und viele promovieren oder sogar habilitieren konnten. Ich danke Dir, dass ich diese Entwicklung als Dein Techniker
mitmachen durfte und hoffe, dies noch viele weitere Jahre tun zu dürfen.
Mattias Werner
Mattias Werner received his Dipl. Inf. in computer science from
Universität Bremen, Germany in 1991. In the same year he started to work as a researcher in the group of Prof. Krieg-Brückner.
Since 1996 he is a researcher and technical assistent and therefor
responsible for all the computer hardware in Bernd Krieg-Brückner’s lab at Universität Bremen and in the German Research Center
for Artificial Intelligence (DFKI) department SKS. He is one of the
developers of the graph drawing software uDraw(Graph).
34
36
2
Appassionato Der Anfang
Die genauen Umstände, wie wir zum ersten Mal mit Bernd in Kontakt kamen,
lassen sich wohl nie mehr im Detail klären. Man schreibt das Jahr 1998, die Entwicklungen am alten INKA-Beweissystem waren seit längerer Zeit in eine Sättigungsphase übergegangen und der Sinn stand nach der Bewältigung gröÿerer
Aufgaben: die Behandlung komplexer Verikationsaufgaben mit ihren dazugehörigen Verwaltungsmechanismen wurde zum zentralen Thema in einem DFKIinternen BMBF-Projekt. Es war die Geburtsstunde der Developmentgraphen 1 .
Auf der Suche nach einem goutierenden Publikum kam das Interesse schnell auf
ein EU-Projekt namens CoFI, das sich die Entwicklung einer Common Algebraic
Specication Language, kurz CASL, zum Ziel gesetzt hatte. Das Projekt bot einige Vorzüge: Eine Überarbeitung der Eingabesprache von INKA war längst
überfällig und dieses CASL schien zumindest ein guter Kandidat dafür zu sein,
die beteiligten Personen stammten nicht aus dem typischerweise introvertierten
Umfeld des Automated Reasoning, und das Projekt hatte sogar Geld, um Reisen zu bezahlen! Ein erster Testballon wurde ugs nach Amsterdam auf eines
dieser CoFi-Meetings geschickt (und kam auch bald unversehrt wieder zurück),
danach ging es daran, das Eintrittsgeschenk in Form einer Umsetzung von CASLStrukturen in Developmentgraphen zu produzieren. Das Resultat kam auf dem
Workshop for Algebraic Development Techniques auf den Tisch. Der Veranstaltungsort, das etwas abgelegenes Schlöÿchen Bonas im Süden Frankreichs, bot genügend andere kulinarische Attraktivitäten, um auch die seelischen Qualen eines
Developmentgraphenvortrags bewältigen zu könnnen. Jedenfalls waren danach
INKA und der Developmentgraph feste Gröÿen in der Welt der CASL-Tools. Im
Umfeld dieser Aktivitäten kam der erste persönliche Kontakt mit Bernd zustande; die CASL-Toolsetgemeinde hatte sich in Bremen verabredet und als frisch
gebackene Gemeindemitglieder waren wir natürlich eingeladen, unsere Werkzeuge zu zeigen. Der Name des Veranstaltungsorts Mehrzweck-Hochhaus versprach
zwar eine Atmosphäre, die von dem Betonzweckbau aus den 70er Jahren auch in
vollstem Maÿe erfüllt wurde, aber ganz oben, auf der Chefetage gleichsam, war
man über solche Dinge weit erhaben. Bernds Interesse an Developmentgraphen
war nahezu unerschöpich. Es sollte der Begin einer inzwischen jahrzehnte
langen Zusammenarbeit werden.
3
Furioso MMISS
Im Herbst 2000 gab es eine Ausschreibung des BMBF zu neuen Medien in der
Bildung und damit auch die Gelegenheit, innerhalb eines gemeinsamen Projekts, quasi oziell, zusammenzuarbeiten. Das BMBF wollte in einer groÿ angelegten Aktion die Entwicklung, Erprobung sowie Einführung innovativer multimedialer Lehr- und Lernformen unterstützen. Nachdem so manche Partnerunverträglichkeit im Vorfeld ausgeräumt wurde, stand das Konsortium. Wegen
1
Der Begri des Developmentgraphen gab es im KIV-System bereits früher, war aber
in seiner Semantik stark eingeschränkt.
37
den Ausschreibemodalitäten war die Universität des Saarlandes Projektpartner,
das DFKI war ein Subcontractor gespeist aus Unterverträgen der Unis des
Saarlandes und Bremens. Der saarländische Hauptbeitrag bildete die Weiterentwicklung des ActiveMath-Systems. Wie bei vielen Projekten, in denen gleichzeitig ein Werkzeug und dessen Anwendung entwickelt werden soll, begann schon
bald der Diskurs, ob denn das Werkzeug auch rechtzeitig zur Vorbereitung des
Lehrinhaltes zur Verfügung stehen würde, oder man nicht vielleicht zwischenzeitlich eine andere Lösung parallel fahren sollte. Sichere Systeme war schon
immer ein Leitthema für Bernd und so wurde das Sicherheitsnetz in Form von
MMISS-LaTeX geknüpft: die Verwendung von LaTeX-Macros als explizite Strukturierungsmittel und XML-Ersatz. Die Idee war einfach: Dokumente sollten in
MMISS-LaTeX geschrieben werden, die Druckform erhält man durch LATEXen,
die semantischen Informationen durch die konsequente Verwendung der Macros.
Die Lösung stieÿ naturgemäÿ nicht bei allen Projektpartnern auf helle Begeisterung und schon bald machte das böse Wort von MMISS-Ergebnissen die Runde.
Bernds Diplomatiekünste waren daher in manchen Projektsitzungen gefragt,
um den angeregten Diskussionen Herr zu werden. Und apropos Projektsitzungen: schmücken sich heutzutage viele World-Wide-Manager mit Projektmeetings
in der VIP-Longue mancher internationaler Fluggesellschaft in Frankfurt, New
York oder London, so war der Hauptbahnhof in Fulda das wichtigste Projektziel
in MMISS. Saarbrücken wollte nicht nach Bremen reisen und umgekehrt: für
eine eintägige Veranstaltung waren die Fahrzeiten von mehr als sechs Stunden
pro Strecke einfach zuviel. Bernd kam daher auf die Idee, die Fahrzeiten dadurch
zu halbieren, dass man sich in der Mitte traf. Ein erster Versuch mit Frankfurt
war mangels vernünftiger Räumlichkeiten kein Erfolg, aber der nächste Versuch
mit Fulda Hbf erwies sich als tragfähig, und so traf man sich von Zeit zu Zeit im
ersten Stock eines Bahnhofgebäudes, das den Charme einer altgedienten Polizeidienststelle ausstrahlte, zur gemeinschaftlich begangenen MMISS-Handlung.
Bernd hatte seine Liebe zu Ontologien entdeckt und präsentierte sein formales Konzept für MMISS-Ontologien für Lehrinhalte - am Beispiel seines eigenen
Dokuments über Ontologien. Hier endlich konnte sich der wahre Theoretiker bewähren, der selbstverständlich noch im Traum die verschiedenen Objekt- und
Metaebenen auseinander halten kann, oder aber wie hier der Didaktiker
Bernd, der es dann doch zum guten Schluÿ schate, die Anwesenden von seinem
Konzept zu überzeugen.
Ein Resultat war, dass das Projektteam ins Kloster bzw. ins Wasser ging,
oder, um es genau zu sagen, in die Benediktinerinnenabtei Frauenwörth am
Chiemsee, in der nicht ganz zufälligerweise zur gleichen Zeit der
tagte, dem man nur allzu gerne die MMISSErgebnisse servierte. Das Projekt endete mit einer groÿen (um dem MMISSBrauch nicht Tür und Tor zu önen) Demonstration an der International University of Bremen und der Hauptbahnhof in Fulda kehrte zu jener Bedeutungslosigkeit zurück, die ihm auch schon vor den MMISS-Zeiten beschieden war.
Seien es Ontologien für strukturierte, natürlichsprachliche Dokumente oder
Konsistenzbedingungen für solche Dokumentsammlungen; in der Endphase des
Algebraic Development Techniques
Workshop for
38
Projektes wurden viele Ideen geboren, die sich aber in der verbliebenen Projektlaufzeit nicht mehr verwirklichen lieÿen. Sie sollten als Samen einer späteren,
zukünftigen Zusammenarbeit einige Zeit vor sich hin schlummern müssen. Der
Grund war, dass sich die Sicherheit (im Sinne einer Security) zunehmend als
zentrales Forschungsgebiet am DFKI im Fachbereich DMAS herauskristallisierte. Die Teilbereiche Multiagentensysteme und Deduktion sollten BMBF- und
SAB-technisch enger kollaborieren und das Ergebnis waren Sichere Agenten
und anschlieÿend Sichere Webservices. Das Gebiet der Developmentgraphen
überwinterte und gedieh derweil im typischerweise frostfreien Bremen.
4
Ostinato DFKI in Bremen
Ende 2004 keimte langsam ein neuer Sproÿ der Zusammenarbeit. Das DFKI war
nach manchen turbolenten Jahren Ende des Milleniums wieder auf Expansionskurs gegangen und langsam machte die Rede von neuen Standorten die Runde.
Erhaben über den üblich undurchdringlichen DFKI-Nebelschwaden war plötzlich die Spitze eines Leuchtturms zu sehen. Bernd hatte die günstige Gelegenheit
genutzt und brachte Bremen als neuen zukünftigen Standort ins Rennen, die Robotik ein für jedes KI-Institut gern gesehenes medienwirksames Aushängeschild
mit im Gepäck. Es folgte für Bremen die mühsame Zeit der Einarbeitung in die
Gepogenheiten im DFKI und in die wundersame Welt der DFKI-Ontologien bestehend aus LAV, T46a und anderen Mirakeln des Projektcontrollings. Der neue
Standort, oder wie er am Anfang heiÿen würde, das neue DFKI-Labor Bremen
sollte gut projektbestückt in seine zweijährige Probezeit gehen.
Mittlerweile neigte sich in Saarbrücken eine andere Epoche dem Ende zu.
Jörg Siekmann war langsam aber unentweichlich dabei, dem berufstätigen Beamtenalter zu entwachsen. Seine Nachfolgerfrage war lange Zeit umnebelt, bis
dann mal kurz die mexikanische Sonne herauskam, um dann aber nach einer nicht
enden wollenden Dämmerung zu versinken und das Problem wieder in absolute
Finsternis zu tauchen. Es war unklar, in welcher Konstellation der Forschungsbereich DMAS weitergeführt würde, zudem gab es jetzt mit Bernds Eintritt in
das DFKI einen zweiten Forschungsbereich, der sich Sicherheit auf die Fahnen
geschrieben hatte. Der Bremer Rolland jedenfalls winkte aus der Ferne und bot
seinen Schutz an2 . Es reifte das Modell zweier Sicherheitsabteilungen, die eine
in Bremen und die andere in Saarbrücken, die miteinander eng verknüpft die
Sicherheitsbrückenköpfe des DFKI bilden sollten.
5
Pieno Conclusio
Der anfänglich nicht so schlechte Eindruck von Bernd bestätigte sich und man
lernte Bernd im Lauf der zurückliegenden 10 Jahre als herzlich und groÿzügig
2
die Tatsache, dass es sich hierbei nicht um den allseits bekannten streitbaren Roland
(vryheit do ik ju openbar) sondern um einen Rollstuhl handeln sollte, war nicht
allen Beteiligten stets bewusst.
39
kennen, im Gegensatz zu seiner eher spartanischen Büroeinrichtung. Hervorzuheben hierbei ist unter anderem, dass, wenn man als Gast in Bremen war, Bernd,
trotz seines engen Terminkalenders und ausstaert mit der letzten kalifornischen Apfel-Technik, es sich oft nicht nehmen lies, einen persönlich am Hotel
abzuholen oder zum Bahnhof zu bringen. Bei den vielen gemeinsamen Essen,
am liebsten natürlich chinesisch, entpuppte sich sein Schalk und derart vielseitiges Interesse, so dass er sich dafür sogar eine zweite Persönlichkeit in Person
des Drehbuchautors, Regisseurs und Schauspielers Bernd Bellum hatte zulegen
müssen.
Wir wünschen Dir, lieber Bernd alles Gute zu Deinem 60. Geburtstag.
Serge Autexier
received his doctoral degree from
Saarland University under the supervision of Jörg
Siekmann. Together with Christoph Benzmüller and
he was leading Jörg Siekmann's OMEGA group at
Saarland University, which integrated into the DFKI Safe and Secure Cognitive Systems Lab of Bernd
Krieg-Brückner, whom he knows since the end of the
1990s from the European project CoFi. His research
focuses on management of change in general and proof
assistance systems with the objective to oer to users
who are not experts in formal logic support that only
formal logic based techniques can reliably provide.
Dominik Dietrich, born in 1980, received his Diploma in Computer Science PhD from the Saarland University in 2006 under supervision of J. Siekmann. Since
2008 he is a PhD student in Bernd Krieg-Brückner's
group. His research interests include automated reasoning with focus on proof planning and deep inference,
and formal software development.
Dieter Hutter
is a principle researcher at the German Research Center of Articial Intelligence. He started his career in the area of deduction systems developing strategies for inductive theorem provers. After
receiving his Ph.D in 1991 his interest focused on formal methods working, rst on the deductive support
for formal methods and later on evolutionary formal
program development and change management. Since 2000 he is also engaged in developing information
ow control techniques for multi-agent systems and
web services. In 2008 he took the opportunity to join
Bernd Krieg-Brückner's group and moved to the newly founded DFKI site in Bremen. Currently he heads
the department for Safe and Secure Systems - Development Technologies at the DFKI.
On Inconsistency and Unsatisfiability
Erwin R. Catesbeiana
University of Bremen, Email: [email protected]
Abstract. We study inconsistency and unsatisfiability and their relation to soundness, completeness, paraconsistency and conservative extension in an arbitrary logical system (formalised as institution equipped
with an entailment system).
1
Introduction
The study of logical inconsistencies has a long tradition that goes back to at least
Aristotle. Indeed, Aristotle examined contemporary philosophical arguments and
revealed inconsistencies. For example, Anaxagoras imagined the mind to be the
initiating principle of all things, and suspending on its axis the balance of the
universe; affirming, moreover, that the mind is a simple principle, unmixed, and
incapable of admixture, he mainly on this very consideration separates it from all
amalgamation with the soul; and yet in another passage he actually incorporates
it with the soul. This inconsistency was pointed out by Aristotle, but it remains
unclear whether he meant his criticism to be constructive, and to fill up a system
of his own, rather than destructive of the principles of others.1
Aristotle himself created a rich source of what perhaps not can be called
inconsistencies, but false theorems: a number of his Syllogisms exhibit the existential fallacy, i.e. have implicit existential assumptions, which means that they
are unsound when read literally. For example, the Fesapo syllogism:
No humans are perfect.
All perfect creatures are mythical.
Some mythical creatures are not human.
After Aristotle came a long period of logical desert, with only Scholastic
arguments, which only was left in the 19th century with the discoveries of Boole,
Frege and others. Actually, Frege created a rich and powerful logical system
called “Begriffsschrift”. It was discovered to be inconsistent by Russell in the
early 20th century, by a proof that resembles the barber paradox: assume that
there is a town with a barber that shaves all people who do not shave themselves.
Then the barber shaves himself iff he does not — a contradiction.
The origin of this inconsistency is the power of self-application, i.e. the possibility to apply predicates to themselves. For example, monosyllabic is an adjective that does not hold of itself, where as polysyllabic does. Now let non-selfreferential be the adjective expressing that an adjective does not hold for itself.
1
See http://www.ccel.org/ccel/schaff/anf03.iv.xi.xii.html.
40
41
That is, monosyllabic is non-self-referential (and polysyllabic isn’t). Is non-selfreferential non-self-referential? It is the merit of the modern web ontology language OWL-full to have provided, in the early 21st century and more than 120
years after Frege, a logic where again predicates can be applied to themselves.
In the modern use of logic in theoretical computer science, the notion of
inconsistency (and its companion unsatisfiability) gains importance in the area
for formal specification and software development with formal methods — our
jubilarian Bernd Krieg-Brückner has been quite active in this research area. A
central idea is that in an early phase of the development process, initial requirements are first formulated informally and then are formalised, such that intended
logical consequences can be checked, and inconsistencies (that prevent the development of a correct implementation) can be found. And indeed, not only
programs are notoriously buggy, but also their specifications tend to be incorrect and inconsistent. Modern specification languages like CafeOBJ and CASL
come with libraries of specifications that also provide examples of such inconsistencies. Indeed, these languages are feature-rich and complex, which eases
the development of non-trivial2 inconsistent theories. In this context, we should
also mention the research on upper ontologies, which are usually quite large and
inconsistent first-order theories.3 Indeed, in the SUMO ontology, several inconsistencies have been found [5]. The SUMO $100 Challenges4 explicitly call for
demonstrating the consistency or inconsistency of (parts of) SUMO. Needless so
say that so far, only an inconsistency has been found.
2
Institutions and Logics
The study of inconsistency and unsatisfiability can be carried out largely independently of the nature of the underlying logical system. We use the notion
of institution introduced by Goguen and Burstall [3] in the late 1970ies. It approaches the notion of logical system from a relativistic view: rather than treating
the concept of logic as eternal and given, it accepts the need for a large variety
of different logical systems, and ask for the central principles of logical systems.
They arrived at a mathematical formalisation of that notion which is really ingenious and lead to rich and flourishing theory that is laid out in hundreds of
research papers, and most recently also in a book [1].
While the notion of institution takes a model.theoretic perspective, it was
later complemented by the more proof-theoretic notion of entailment system
(also called Π-institution) [8, 2].
What are the essential ingredients of a logical system? It surely has a notion of sentence, and derivability relation on sentences that allows to derive
conclusions from a given set of assumptions. Moreover, at the model-theoretic
2
3
4
Of course, the set of logical consequences of an inconsistent theory is always trivial.
However, the axioms of the theory itself may have varying complexity and subtlety
of interaction in order to lead to the inconsistency.
Although there are more efficient ways of writing up such theories . . .
See http://www.cs.miami.edu/~tptp/Challenges/SUMOChallenge/.
42
side, there is a notion of model and a satisfaction relation between models and
sentences, denoted by |=. The latter leads to the relation of logical consequence
(all models satisfying the premises also satisfy the conclusion); this relation is
also denoted by |=. A logic is sound, if Γ ϕ implies Γ |= ϕ, and complete, if
the converse implication holds.
Moreover, an important observation is that all this structure depends on the
context, i.e. the set of non-logical (or user-defined) symbols around. These contexts are called signatures and formalised just as object of an abstract category.
The reader not familiar with category theory or not interested in the formal
details can safely skip the rest of this section and just keep in mind the above
informal motivations.
Definition 1. An entailment system consists of
– a category SignE of signatures and signature morphisms,
– a functor SenE : SignE −→ Set giving, for each signature Σ, the set of sentences SenE (Σ), and for each signature morphism σ : Σ −→ Σ , the sentence
translation map SenE (σ) : SenE (Σ) −→ SenE (Σ ), where often SenE (σ)(ϕ) is
written as σ(ϕ),
Σ ⊆ P(Sen(Σ)) × Sen(Σ) for each Σ ∈ |Sign|, such that the following
properties are satisfied:
1. reflexivity: for any ϕ ∈ Sen(Σ), {ϕ} Σ ϕ,
2. monotonicity: if Ψ Σ ϕ and Ψ ⊇ Ψ then Ψ Σ ϕ,
3. transitivity: if Ψ Σ ϕi for i ∈ I and Ψ ∪ {ϕi | i ∈ I} Σ ψ, then Ψ Σ ψ,
4. -translation: if Ψ Σ ϕ, then for any σ : Σ −→ Σ in Sign, σ(Ψ ) Σ σ(ϕ).
– for each signature Σ ∈ |SignE |, an entailment relation (Sen(Σ), EΣ ),
– for each signature morphism σ : Σ1 −→ Σ2 , a sentence translation map
SenE (Σ1 ) → SenE (Σ2 ) preserving E . By abuse of notation, we will also
denote this map with σ.
A theory is a pair (Σ, Γ ) where Γ is a set of Σ-sentences. An entailment
theory morphism 5 (Σ, Γ ) −→ (Σ , Γ ) is a signature morphism σ : Σ −→ Σ
such that Γ Σ σ(Γ ).
Let Cat be the category of categories and functors.6
Definition 2. An institution I = (SignI , SenI , ModI , |=I ) consists of
– a category SignI of signatures,
– a functor SenI : SignI −→ Set (as for entailment systems),
5
6
also called interpretation of theories.
Strictly speaking, Cat is not a category but only a so-called quasicategory, which is
a category that lives in a higher set-theoretic universe [4]. However, we do not care
about this here. Indeed, these foundational questions are ignored by most mathematicians and computer scientists, since ignoring them provides a rich source of
inconsistencies.
43
– a functor ModI : (SignI )op −→ Cat giving, for each signature Σ, the category
of models ModI (Σ), and for each signature morphism σ : Σ −→ Σ , the
reduct functor ModI (σ) : ModI (Σ ) −→ ModI (Σ), where often ModI (σ)(M )
is written as M |σ ,
– a satisfaction relation |=IΣ ⊆ |ModI (Σ)| × SenI (Σ) for each Σ ∈ SignI ,
such that for each σ : Σ −→ Σ in SignI the following satisfaction condition
holds:
M |=IΣ σ(ϕ) ↔ M |σ |=IΣ ϕ
for each M ∈ ModI (Σ ) and ϕ ∈ SenI (Σ).
We will omit the index I when it is clear from the context.
A logic is an entailment system paired with an institution agreeing on their
signature and sentence parts. Usually, a logic is required to be sound, that is,
Γ Σ ϕ implies Γ |=Σ ϕ. If also the converse holds, the logic is complete.
A theory is defined as for entailment systems. An institution theory morphism
(Σ, Γ ) −→ (Σ , Γ ) is a signature morphism σ : Σ −→ Σ such that Γ |=Σ σ(Γ ). Let Th(I) denote this category. Each theory (Σ, Γ ) inherits sentences
from SenI (Σ), while the models are restricted to those models in ModI (Σ) that
satisfy all sentences in Γ . It is easy to see that I maps theory morphisms to
corridors in this way. By taking Th(I) as “signature” category, we arrive at the
institution I T h of theories.
Example 3. Classical propositional logic (CPL) has the category Set of sets
and functions as its signature category. Σ-sentences are given by the following
grammar
ϕ ::= p | ¬ϕ | ϕ1 ∧ ϕ2 | ϕ1 ∨ ϕ2 | ϕ1 → ϕ2 | | ⊥
where p denotes propositional variables from Σ. Sentence translation is the obvious replacement of propositional variables.
Σ-models are functions from Σ to {T, F }. A (unique) model morphism between two models exists if the values of variables change only from F to T , but
not vice versa. Given a signature morphism σ : Σ1 −→ Σ2 , the σ-reduct of a
Σ2 -model M2 : Σ −→ {T, F } is given by composition: M2 ◦ σ. This obviously
extends to reducts of morphisms.
Satisfaction is inductively defined by the usual truth table semantics. Since
reduct is composition, it is straightforward to prove the satisfaction condition.7
The entailment relation is the minimal relation satisfying the properties listed
in Table 1, plus the property ¬¬ϕ ϕ.
This logic is sound and complete.
7
A precise argument is as follows: the Boolean operations form a signature such that
Sen is the free algebra functor for algebras over that signature (these algebras are
Boolean algebras without laws). {T, F } also is such an algebra, denoted by Bool,
and for a model M , sentence evaluation is εBool ◦ Sen(M ), where ε is the counit.
Then the satisfaction condition is εBool ◦Sen(M )◦Sen(σ) = εBool ◦Sen(M ◦σ), which
is just functoriality of Sen.
44
A Heyting algebra H is a partial order (H, ≤) with a greatest element and
a least one ⊥ and such that any two elements a, b ∈ H
– have a greatest lower bound a ∧ b and a least upper bound a ∨ b, and
– there exists a greatest element x such that a ∧ x ≤ b; this element is denoted
a ⇒ b.
In a Heyting algebra, we can define a derived operation ¬ by ¬a := a ⇒ ⊥. A
Heyting algebra morphism h : H1 −→ H2 is a map preserving all the operations.
Example 4. Heyting-algebra based intuitionistic propositional logic (IPL-HA)
inherits the signature category and sentences from CPL.
A Σ-model (ν, H) consist of a Heyting algebra H together with a valuation
function ν : Σ −→ |H| into the underlying set |H| of H. Σ-model morphisms
h : (ν1 , H1 ) −→ (ν2 , H2 ) are Heyting algebra morphisms h : H1 −→ H2 such that
h1 ◦ν1 (p) ≤ ν2 (p) for each p ∈ Σ. Again, model reduct is defined by composition,
and this easily extends to morphisms.
Using the Heyting algebra operations, it is straightforward to extend the
valuation ν of a Σ-model (ν, H) from propositional variables to all sentences:
ν # : Sen(Σ) −→ |H|. (ν, H) satisfies a sentence ϕ iff ν # (ϕ) = . The satisfaction
condition follows similarly as for CPL.
The entailment relation is the minimal relation satisfying the properties listed
in Table 1. This turns IPL-HA into a sound and complete logic.
3
Logical connectives
Within an abstract logic, it is possible to define logical connectives purely be
their proof-theoretic and model-theoretic properties. We start with proof theory
and adapt the standard definitions from [7].
connective
defining property
proof-theoretic conjunction ∧
Γ ϕ ∧ ψ iff Γ ϕ and Γ ψ
proof-theoretic disjunction ∨ ϕ ∨ ψ, Γ χ iff ϕ, Γ χ and ψ, Γ χ
proof-theoretic implication →
Γ, ϕ ψ iff Γ ϕ → ψ
proof-theoretic truth Γ proof-theoretic falsum ⊥
⊥ϕ
proof-theoretic negation ¬
Γ, ϕ ⊥ iff Γ ¬ϕ
Table 1. Properties of proof-theoretic connectives
Note that these properties characterise connective essentially by their prooftheoretic behaviour; they mostly even directly correspond to proof rules. Below,
45
we will also introduce semantic connectives. A logic is said to have a prooftheoretic connective if it is possible to define an operation on sentences with the
properties specified in Table 1. For example, both IPL-HA and CPL have all
proof-theoretic connectives.
We also can define internal connectives at the semantic level. A logic is said
to have a semantic connective if it is possible to define an operation on sentences
with the specified properties.
connective
defining property
semantic disjunction ∨
M |= ϕ ∨ ψ iff M |= ϕ or M |= ψ
semantic conjunction ∧ M |= ϕ ∧ ψ iff M |= ϕ and M |= ψ
semantic implication → M |= ϕ → ψ iff M |= ϕ implies M |= ψ
semantic negation ¬
M |= ¬ϕ iff M |= ϕ
semantic truth M |= semantic falsum ⊥
M |= ⊥
Table 2. Properties of semantic connectives
While CPL has all semantic connectives (indeed, they coincide with the
proof-theoretic ones), IPL-HA only has semantic conjunction, truth and falsum.
4
Inconsistency and Unsatisfiability
In the sequel, let us fix a logic L (in the above sense), which a priori need neither
be sound nor complete.
The notion of unsatisfiability is quite clear:
Definition 5. A theory is unsatisfiable if it has no models.
Indeed, the notion of inconsistency is more interesting than unsatisfiability8 ,
since it enjoyed some historical development. According to Aristotle, inconsistency means that both some sentence as well as its negation can be proved.
Definition 6. Assume that L has a negation connective (not necessarily being a
proof-theoretic or semantic negation9 ), and let T be a theory in L. T is Aristotle
inconsistent, if there is some sentence ϕ with T ϕ and T ¬ϕ.
This notion has several disadvantages. Firstly, it presupposes the notion of
negation, which is not available in all logics. More importantly, it classifies paraconsistent logics as inconsistent, which can only be apologised by the fact that
paraconsistent logics were not known at Aristotle’s time.
8
9
Although the Rolling Stones devoted a song to unsatisfiability.
However, let us assume that if there is a proof-theoretic negation, then this is used.
Otherwise, the notion of inconsistency of course depends on the chosen connective.
46
A modern definition of inconsistency overcoming this problem was coined by
David Hilbert. Hilbert was the founder of the famous “Hilbert programme”, the
aim of which was to prove the consistency of all of mathematics by reducing
it to the consistency of a small number of finitary principles, for which there
is enough faith into their consistency. Hilbert’s programme greatly failed, as
was shown by Gödel’s second incompleteness theorem (actually, the name is
misleading: it should be called Gödel’s Great Inconsistency Theorem):
Theorem 7 (Gödel). There is a first-order theory T of zero, successor, addition and ordering on the natural numbers (which is actually quite simple and
weak), such that for any extension T of T , if T can prove its own consistency
(encoded as a statement on natural numbers), then T is inconsistent.10
Hence, although Hilbert’s programme was a powerful and striking idea, in the
end it could not be successful. As a result, the question whether the theories like
ZF C (that are used as a foundation of mathematics and theoretical computer
science!) are consistent or inconsistent is open. Indeed, the only way to firmly
resolve this open question would be to prove the inconsistency of ZF C.11 All
what we have so far are relative results, such as the famous result by Gödel:
ZF C is inconsistent iff ZF is inconsistent.12
which means that when looking for an inconsistency proof for ZF , we equally
well may use the stronger (and hence easier to prove inconsistent) system ZF C.
But even though Hilbert’s programme failed, Hilbert left us with a modern
definition of inconsistency:
Definition 8 (Hilbert). Assume that L has a falsum constant (not necessarily
being a proof-theoretic or semantic falsum13 ). T is ⊥-inconsistent, if T ⊥.
Still, this definition does not work with logics that do not have a falsum,
for example positive or equational logic. Hilbert therefore also found a notion of
inconsistency that has the least prerequisites (no logical connectives are needed)
and that simultaneously is most powerful (in terms of logical strength of inconsistent theories):
Definition 9 (Hilbert). T is absolutely inconsistent, if T ϕ for any sentence
ϕ of same signature as T . This is also known as the principle ex contradictione
quodlibet (from a contradiction, one can follow everything).
10
11
12
13
For first-order logic, the various notions of inconsistency discussed so far are equivalent; hence we can be unspecific here.
For an initial attempt in this direction, see [6].
Actually, Gödel proved the corresponding statement about unsatisfiability, but by
Gödel’s completeness theorem for first-order logic, inconsistency and unsatisfiability
are equivalent, see our Prop. 12 below.
Again, let us assume that if there is a proof-theoretic falsum, then this is used.
Otherwise, the notion of inconsistency depends on the chosen constant.
47
We should also mention a notion of inconsistency introduced by Emil Post:
a propositional theory T is Post-inconsistent, if it can derive a propositional
variable (not occurring in the axioms of T ; the signature possibly needs to be
enlarged to get such a variable). Unfortunately, this notion is too much tied to
a specific logical system to be of interest here.
The different notions of inconsistency14 are related as follows:
Proposition 10. 1. absolute inconsistency implies ⊥-inconsistency and Aristotle inconsistency.
2. In presence of proof-theoretic falsum, absolutely inconsistency and ⊥-inconsistency
are equivalent.
3. In presence of proof-theoretic falsum and negation, all three notions of inconsistency re equivalent.
Proof. 1. obvious
2. Directly from the definition of proof-theoretic falsum.
3. By 1. and 2., it remains to show that Aristotle inconsistency.implies absolutely inconsistency. By definition of proof-theoretic negation, from Γ ¬ϕ we
get and Γ ∪ {ϕ} ⊥. Together with Γ ϕ, this leads to and Γ ⊥.
5
Soundness and Completeness, with an Application to
Paraconsistency
Inconsistency and unsatisfiability also play a great role in determining whether
a logic is sound or complete.
We begin with a simple lemma showing that falsum and truth are two sides
of the same coin:
Lemma 11. In presence of proof-theoretic negation, falsum and truth,
¬⊥ and ¬ ⊥.
Soundness and completeness, while defined in terms of entailment, can be
characterised completely in terms of inconsistency and unsatisfiability.
Proposition 12. Let L be a logic with both proof-theoretic and semantic negation, truth and falsum. (Then, by Prop. 10, all notions of inconsistency are
equivalent, hence we can be unspecific in the sequel.)
1. A logic is sound iff every inconsistent theory is unsatisfiable.
2. A logic is complete iff every unsatisfiable theory is inconsistent.
14
We
credit
http://home.utah.edu/~nahaj/logic/structures/systems/
inconsistent.html for an excellent overview of these notions.
48
Proof. (1), “⇒” Let T be inconsistent, i.e. T ⊥. By soundness, T |= ⊥, hence
T is unsatisfiable.
(1), “⇐” Let T ϕ, then T ∪ {¬ϕ} is inconsistent, hence, by the assumption,
also unsatisfiable. But this means that T |= ϕ.
(2), “⇒” Let T |= ϕ. Then T ∪ {¬ϕ} is not satisfiable, hence by the assumption, it is inconsistent. From T ∪ {¬ϕ} ⊥, we obtain T ∪ {¬ϕ} ¬¬⊥ and
hence T ∪ {¬⊥} ϕ. By Lemma 11, T ∪ {} ϕ. By the properties of and
transitivity, we get T ϕ.
(2), “⇐” Let T be unsatisfiable, i.e. T |= ⊥. By completeness, T ⊥, hence
T is inconsistent.
It should be stressed that these proofs become less elegant when reformulated
in terms of consistency and satisfiability, as some over-cautious logicians do —
they tend to be easily frightened by inconsistencies15 . The more natural relation
is indeed that between inconsistency and unsatisfiability.
Definition 13. A logic is paraconsistent if it admits a theory that is Aristotle
inconsistent but absolutely consistent.
Proposition 14. 1. A paraconsistent logic cannot have proof-theoretic negation.
2. A sound and complete paraconsistent logic cannot have model-theoretic negation.
Proof. 1. By Prop. 10.
2. Let T be an Aristotle inconsistent theory, i.e. T ψ and T ¬ψ for some
sentence ψ. By soundness, T |= ψ and T |= ¬ψ. By the definition of modeltheoretic negation, T is unsatisfiable. But then for any sentence ϕ, T |= ϕ, thus
by completeness, also T ϕ. Hence, T is absolutely inconsistent.
6
Conservative Extensions
Definition 15. A theory morphism σ : T1 −→ T2 is model-theoretically conservative, if each T1 -model has a σ-expansion to a T2 -model. It is consequencetheoretically conservative, if for each sentence ϕ of same signature as T1 ,
T2 |= σ(ϕ) implies T1 |= ϕ.
It is proof-theoretically conservative, if the same holds for , i.e. for each sentence ϕ of same signature as T1 ,
T2 σ(ϕ) implies T1 ϕ.
The relation between these notions is as follows:
15
This goes as far as the Wikipedia website for “Inconsistency” being redirected to
“Consistency”!
49
Proposition 16. Model-theoretic conservativity implies consequence-theoretic
conservativity (but not vice versa). In a sound and complete logic, consequencetheoretic and proof-theoretic conservativity are equivalent.
Proof. Concerning the first statement, let T2 |= σ(ϕ). We need to show T1 |= ϕ.
Let M1 be a T1 -model. By model-theoretic conservativity, it has an expansion
M2 to a T2 -model; hence also M2 |= σ(ϕ). By the satisfaction condition, M1 =
M2 |σ |= ϕ.
The second statement is obvious.
The importance of conservativity is that it allows to reduce inconsistency
(resp. unsatisfiability) of a smaller theory to that of larger one, using the following obvious result:
Proposition 17. 1. Model-theoretically conservative theory morphisms reflect
unsatisfiability.
2. Proof-theoretically conservative theory morphisms reflect inconsistency.
Typically, the target of a conservative theory morphism is larger than (or even
an extension of) the source. At first sight, it may sound strange to reduce the
goal of showing inconsistency of a given theory to that of showing it for a larger
one. However, note that generally, larger theories have more axioms, and hence
it is more likely to prove that they are inconsistent. (Indeed, the largest theory
consisting of all sentences is always inconsistent.) Moreover, small inconsistent
theories can be boring: who is interested in the smallest inconsistent theory
consisting just of falsum? For example, Frege’s Begriffsschrift (which is a conservative extension of falsum) is much more interesting.
7
Conclusion
We have summarised and discussed recent research aimed at proving inconsistency of specifications in a structured and institution-independent way. These
efforts form part of a larger program aimed at proving the inconsistency of the
empty specification, i.e. inconsistency of the meta-framework underlying most
of the formalisms under consideration, Zermelo-Fraenkel set theory with choice
(ZFC) and hence, by the independence of the axiom of choice as mentioned
above, of ZF set theory without choice. In particular, recent research in pure
mathematics has been concerned with finding sufficient conditions for the inconsistency of ZF. E.g. it has been shown that the inconsistency of ZF can be
reduced to (and is therefore equivalent to) provability in ZF of
– the existence of natural numbers a, b, c and n ≥ 3 such that an + bn = cn
(Andrew Wiles [13])
– the existence of natural numbers (a, b, x, y) = (2, 3, 3, 1) such that xa −y b = 1
(Preda Mihăilescu [9])
– the existence of a simply connected closed 3-manifold not homeomorphic to
S 3 (Grigori Perelman [10–12])
50
Moreover, ongoing work in automated theorem proving, Thomas Hales’ FlysPecK project (http://code.google.com/p/flyspeck/), is directed at reducing the
inconsistency of ZF to existence of a sphere packing of average density strictly
less than π/18. In summary, there is good hope that the paradise of mathematics,
a widely accepted inconsistent set of foundations, will soon be re-opened.
References
1. R. Diaconescu. Institution-Independent Model Theory. Birkhäuser, 2008.
2. J. Fiadeiro and A. Sernadas. Structuring theories on consequence. In D. Sannella
and A. Tarlecki, eds., 5th WADT, vol. 332 of LNCS, pp. 44–72. Springer Verlag,
1988.
3. J. A. Goguen and R. M. Burstall. Institutions: Abstract model theory for specification and programming. Journal of the Association for Computing Machinery,
39:95–146, 1992. Predecessor in: LNCS 164, 221–256, 1984.
4. H. Herrlich and G. Strecker. Category Theory. Allyn and Bacon, Boston, 1973.
5. I. Horrocks and A. Voronkov. Reasoning support for expressive ontology languages
using a theorem prover. In Proceedings of the Fourth International Symposium
on Foundations of Information and Knowledge Systems (FoIKS), number 3861 in
Lecture Notes in Computer Science, pp. 201–218. Springer, 2006.
6. R. E. Kamouna. A paradox ⇒ SAT is (NOT) NP-complete & ZFC is inconsistent.
http://arxiv.org/abs/0806.2947.
7. J. Lambek and P. J. Scott. Introduction to Higher Order Categorical Logic. Cambridge University Press, 1986.
8. J. Meseguer. General logics. In Logic Colloquium 87, pp. 275–329. North Holland,
1989.
9. P. Mihăilescu. Primary cyclotomic units and a proof of catalan’s conjecture. J.
Reine angew. Math., 572:167–195, 2004.
10. G. Perelman. The entropy formula for the ricci flow and its geometric applications.
arXiv:math.DG/, 2002.
11. G. Perelman. Ricci flow with surgery on three-manifolds. arXiv:math.DG/0303109,
2002.
12. G. Perelman. Finite extinction time for the solutions to the ricci flow on certain
three-manifolds. arXiv:math.DG/0211159, 2003.
13. A. Wiles. Modular elliptic curves and fermat’s last theorem. Ann. Math., 141:443–
551, 1995.
The VSE Refinement Method in Hets
Mihai Codescu1 , Bruno Langenstein2 , Christian Maeder1 , and Till
Mossakowski1,3,4
1
1
German Research Center for Artificial Intelligence (DFKI GmbH), Bremen,
Germany
2
DFKI GmbH, Saarbrücken, Germany
3
SFB/TR 8 Spatial Cognition, Bremen and Freiburg, Germany
4
Albert-Ludwigs-Universität Freiburg, Germany
Introduction
Bernd Krieg-Brückner’s research group originally (in the early 1980ies) carried
the title “programming languages and compiler construction”. Indeed, Bernd
had been a co-designer of the programming language Ada. But already in this
phase, he interests extended to a further topic: formal specification. He worked
on Anna [12], a language allowing for annotating Ada programs with formal
specifications.
More than ten years later, Bernd was founding a member of the “Common
Framework Initiative for Algebraic Specification and Development” (CoFI),
which originated in the COMPASS project and the IFIP working group 1.3.
The aim was to reach an agreement, among approximately 50 researchers from
all over Europe, on a common specification language called Casl (the Common
Algebraic Specification Language) [1, 3, 7]. Indeed, Bernd has played an essential role in the language design process, especially in the critical phase where
everything was unclear and everybody put his or her favourite features on the
table.5
Again nearly ten years later, Bernd has initiated the founding of a new branch
of the German research center for artificial intelligence (DFKI) in Bremen. Some
four years later, this is now a great success, with many different research areas
and industry cooperations. The present work arose in the context of the project
“FormalSafe” funded by the German ministry of research. Indeed, part of this
project consists of a cooperation of DFKI Saarbrücken with DFKI Bremen in
the area of formal methods. While DFKI Saarbrücken has long experiences with
the development and use of the industrial-strength verification tool VSE, DFKI
Bremen has developed the experimental and flexible verification tool Hets. We
want to combine the best of both worlds by establishing a connection between the
5
After three years of design discussion, in Spring 1998, another one of dozens meetings of the CoFI language design group took place. At a time point when nearly
everything was settled and everyone was completely exhausted, Bernd announced:
“unfortunately, I just have realised that we have to revert a design decision and go
back to non-linear visibility.” Everyone sighed. But then Bernd revealed that it was
an April joke — and the following laughter was very refreshing indeed.
51
52
VSE prover and the Hets proof management. For VSE, this brings additionally
flexibility: VSE specifications can now be verified not only with the VSE prover,
but also with provers like SPASS and Isabelle which are interfaced with Hets.
On the other hand, Hets benefits from VSE’s industrial experience, including
a practical relation between specification and programming languages.
2
Presentation of Hets
Hets is a multi-logic proof management tool that heterogeneously integrates
many languages, logics and tools on a strong semantic basis.
The origins of Hets date back to 1997, when a tool for the newly developed
specification language Casl was needed. Cats, the Casl tool set (see Fig. 1),
provided a parser and a static analysis for Casl specifications, as well as an
encoding of Casl into the theorem prover Isabelle/HOL. With Cats, simple
consequences of specifications could be verified, but proof support for e.g. refinements between specifications was lacking — although a first connection to the
“development graph manager” built in Saarbrücken was set up.
Fig. 1. Architecture of the Casl tool set
53
2.1
CASL
Casl had been designed from the outset as the central language in a family of
languages. Hence, soon sublanguages and extension of Casl, like the higherorder extension HasCasl, the coalgebraic extension CoCasl, the modal logic
extension ModalCasl the reactive extensions Casl-Ltl and Csp-Casl and
others emerged. Then, naturally the question arose: how to extend Cats to
these languages? Luckily, the Casl follows a separation of concerns — it has
been designed in four different layers [1, 3, 7]:
basic specifications are unstructured collections of symbols, axioms and theorems, serving the specification of individual software modules. The specific
logic chosen for Casl here is first-order logic with partial functions, subsorting and induction principles for datatypes;
structured specifications organise large specifications in a structured way, by
allowing their translation, union, parameterisation. restriction to an export
interface and more. Still, structured specifications only cover the specification
of individual software modules;
architectural specifications allow for prescribing the structure of implementations, thereby also determining the degree of parallelism that is possible
in letting different programmers independently develop implementations of
different subparts;
specification libraries allow the storage and retrieval of collections of specifications, distributed over the Internet.
2.2
Institutions
A crucial point in the design of these layers is that the syntax and semantics
of each layer is orthogonal to that of the other layers. In particular, the layer
of basic specifications can be changed to a different language and logic (e.g. an
extension of Casl, or even a logic completely unrelated to Casl), while retaining
the other layers. The central abstraction principle to achieve this separation of
layers is the formalisation of the notion of logical system as institutions [9], a
notion that arose in the late 1970ies when Goguen and Burstall developed a
semantics for the modular specification language Clear [5].
We recall this central notion here, assuming some basic knowledge of category theory. The reader not familiar with category theory can safely skip this
definition, and just remember that an institution provides
– a notion of signature, carrying the context of user-defined (i.e. non-logical)
symbols, and a notion of signature morphisms (translations between signatures);
– for each signature, notions of sentence and model, and a satisfaction relation
between these;
– for each signature morphism, a sentence translation and a model reduction
(the direction of the latter being opposite to the signature morphism), such
that satisfaction is invariant under translation resp. reduction along signature morphisms.
54
More formally, an institution I consists of:
– a category SignI of signatures;
– a functor SenI : SignI → Set,6 giving a set Sen(Σ) of Σ-sentences for
each signature Σ ∈ |SignI |, and a function Sen(σ) : Sen(Σ) → Sen(Σ ),
denoted by σ, that yields σ-translation of Σ-sentences to Σ -sentences for
each signature morphism σ : Σ → Σ ;
7
– a functor ModI : Signop
I → Cat, giving a category Mod(Σ) of Σ-models
for each signature Σ ∈ |SignI |, and a functor Mod(σ) : Mod(Σ ) → Mod(Σ),
denoted by |σ , that yields σ-reducts of Σ -models for each signature morphism σ : Σ → Σ ; and
– for each Σ ∈ |SignI |, a satisfaction relation |=I,Σ ⊆ ModI (Σ) × SenI (Σ)
such that for any signature morphism σ : Σ → Σ , Σ-sentence ϕ ∈ SenI (Σ) and
Σ -model M ∈ |ModI (Σ )|:
M |=I,Σ σ(ϕ) ⇐⇒ M |σ |=I,Σ ϕ
[Satisfaction condition]
A very prominent example is the institution F OL= of many-sorted firstorder logic with equality. Signatures are many-sorted first-order signatures, i.e.
many-sorted algebraic signatures enriched with predicate symbols. Models are
many-sorted first-order structures, and model reduction is done by translating a
symbol that needs to be interpreted along the signature morphism before looking
up its interpretation in the model that is being reduced. Sentences are first-order
formulas, and sentence translation means replacement of the translated symbols.
Satisfaction is the usual satisfaction of a first-order sentence in a first-order
structure.
The institution CF OL= adds sort generation constraints to F OL= . These
express that some of the carriers sets are generated by some of the operations
(and possibly the other carrier sets). SubP CF OL= , the Casl institution, further
equips CF OL= with subsorting and partial functions (which, however, will not
play a role in this paper).
With the notion of institution providing the abstraction barrier between the
layer of basic specifications on the one hand and the other layers on the other
hand, it was quite natural (though also a great challenge) to realise this abstraction barrier also at the level of tools. The necessary re-implementation of Cats
lead to the Heterogeneous Tool Set Hets. Hets provides an object-oriented
interface for institutions, realised through a Haskell type class. New institutions (together with their concrete syntax, parsers, static analysis tools) can be
plugged in with moderate effort. Moreover, Hets provides generic interfaces that
ease the connection of individual institutions to proof tools. There are interfaces
capturing both interactive and automatic tools.
Hets allows for relating specifications written in different logics, e.g. Casl
specifications can be imported for Casl extensions, or refinements can occur
across different logics. In order to support this, Hets treats logic translations,
6
7
The category Set has all sets as objects and all functions as morphisms.
Cat is the quasi-category of categories and functors.
55
formalised as institution comorphisms (and morphisms), as first-class citizens.
An institution comorphism captures the idea of encoding or embedding between
two institutions. It provides
– a translation of signatures (and signature morphisms),
– a translation of sentences,
– a translation of models (going backwards to the direction of the comorphism),
such that satisfaction is invariant under translation of sentences resp. models.
More formally, an institution comorphism ρ : I → I consists of:
– a functor ρSign : Sign → Sign ;
– a natural transformation ρSen : Sen → ρSign ; Sen , that is, a family of func Sign
(Σ)), natural in Σ ∈ |Sign|; and
tions ρSen
Σ : Sen(Σ) → Sen (ρ
Mod
: (ρSign )op ; Mod → Mod, that is, a family
– a natural transformation ρ
: Mod (ρSign (Σ)) → Mod(Σ), natural in Σ ∈ |Sign|,
of functions ρMod
Σ
Sign
such that for any Σ ∈ |Sign|, the translations ρSen
(Σ))
Σ : Sen(Σ) → Sen (ρ
Sign
Mod
(Σ)) → Mod(Σ) of models preserve the
of sentences and ρΣ : Mod (ρ
satisfaction relation, i.e., for any ϕ ∈ Sen(Σ) and M ∈ Mod (ρSign (Σ)):
Mod
M |=ρSign (Σ) ρSen
Σ (ϕ) ⇐⇒ ρΣ (M ) |=Σ ϕ
[Satisfaction condition]
Hets is based on a logic graph of institutions and comorphisms, which is a
parameter to the tools acting at the structured, architectural and library layers.
The logic graph can be changed and extended without the need even to recompile
those logic independent analysis tools. The architecture of Hets is shown in
Fig. 2. Hets’ development graph component, inspired by the tool Maya (a
cousin of VSE, also developed in Saarbrücken) provides a proof management
for heterogeneous specifications, relying on proof tools for the individual logics
involved.
2.3
Refinement
The methodology of formal software development by stepwise refinement describes the ideal process (which in practice is more a loop with continuous feedback) as follows: starting from initial informal requirements, these are translated
to a formal requirement specification, which is then further refined to a formal
design specification and then to an executable program.
Simple refinements between specifications can be expressed as so-called views
in Casl, which are just theory morphisms. The degree of looseness diminishes
along a refinement (technically, the model class shrinks). For more complex refinements involving architectural decompositions (i.e. branching points in the
emerging refinement tree), a refinement language has been designed [13]. Sometimes (e.g. when refining arrays to stacks with pointers), an observational interpretation of specification is needed. This has been developed in theory to some
degree [4], but not implemented in Hets yet.
56
Architecture of the heterogeneous tool set Hets
Text
Tools for heterogeneous
specifications
Logic graph
Tools for specific logics
Haskell
Isabelle
Parser
Text
Abstract syntax
HasCASL
Parser
Static Analysis
VSE
CspCASL
CoCASL
Static Analysis
(Signature, Sentences)
SoftFOL
ModalCASL
Interfaces
CASL
XML, Aterm
Conservativity checkers
Model finders
Model checkers
RelScheme
Global Environment
Interfaces
CASL-DL
Theorem provers
Rewriters
Abstract syntax
XML, Aterms
WWW, GUI
DL
Propositional OWL-DL
Grothendieck logic
(Flattened logic graph)
Heterogeneous
development graphs
Heterogeneous inference engine
Decomposition of proof obligations
Management of proofs
Heterogeneous proof trees
Fig. 2. Architecture of Hets
3
Presentation of VSE
The V erification S upport E nvironment (VSE) is a tool that supports the formal development of complex large scale software systems from abstract high
level specifications down to the code level. It provides both an administration
system to manage structured formal specifications and a deductive component
to maintain correctness on the various abstraction levels (see Fig. ??). Taken
together these components guarantee the overall correctness of the complete
development. The structured approach allows the developer to combine specifications in an algebraic functional style with state based formal descriptions of
concurrent systems.
3.1
History of VSE
VSE has been developed in two phases on behalf the German Bundesamt für
Sicherheit in der Informationstechnik (BSI) to satisfy the needs in software developments according to the upcoming standards ITSEC and Common Criteria.
Since then VSE has been successfully applied in several industrial and research
projects, many of them being related to software evaluation [10, 2, 11, 6].
57
;"
;#
!"
"
;
Fig. 3. Architecture of VSE
3.2
The VSE Methodology
VSE supports a development process that starts with a modular formal description of the system model and possibly together with separate requirements or
security objectives. Logically the requirements have to be derivable from the system model. Therefore, the requirements lead to proof obligations that must be
discharged by using the integrated deductive component of VSE.
In a refinement process the abstract system model can be related to more
concrete models. This is in correspondence with a software development that
starts from a high level design and then descends to the lower software layers
such that in a sense higher layers are implemented based on lower layers. Each
such step can be reflected by a refinement step in VSE. These steps involve
programming notions in the form of abstract implementations, that can later be
exploited to generate executable code. Each refinement step gives rise to proof
obligations showing the correctness of the implementations. Refinements also
can be used to prove consistency of specifications, because they describe a way
how to construct a model. This plays a major role for the formal specifications
required for Common Criteria, which only need to cover higher abstraction levels.
In addition to the vertical structure given by refinement steps, VSE also allows the specification to be structured horizontally to organise the specifications
on one abstraction level. Each single (sub)specification can be refined vertically
or further decomposed horizontally, such that the complete development is rep-
58
resented by a development graph. The deductive component is aware of this
structure. And this is an important aspect for the interactive proof approach, as
the structure helps the user to prove lemmas or prove obligations that require
properties from various parts of the specification.
3.3
Specification Formalisms in VSE
Formal development in VSE is currently based on two formalisms: Abstract data
types are used to describe data structures and functional computations on them.
A variant of temporal logic similar to TLA is provided as a specification language for state based systems that consist of concurrent components. While
abstract data types can already be used in the temporal logic specifications, a
new formalism will be available in the near future, that allows the developer to
combine state based aspects with abstract data types. This new formalism aims
at formalising implementations of abstract datatypes with the help of pointer
structures.
Functional Modelling Abstract data type specifications in VSE are essentially
first-order theories (often simply called theories) extended with sort generation
constraints. They are similar to CASL specifications and a similar set of structuring operations is provided that consists of enrichment, union, parametrisation
and instantiation.
In addition, the VSE specification language supports a refinement approach
based on explicit submodels and congruences [14]. (This is in contrast to Casl’s
observational refinement [4], where the congruence is generated implicitly.) Refinements associate an abstract data type specification, called the export specification of the refinement, with an implementation. The implementation is based
on another theory, called the import specification and contains several functional
procedures written in an imperative language. These procedures use the functions and predicates of the import specifications. A so called mapping relates
each sort of the export specification to a sort of the import specification, while
the functions and procedures are mapped to procedures in the import specification.
A refinement describes the construction of a model for the signature of the
export specification (export model) from a model of the import specification
(import model). The functions and predicates are interpreted by the computations of the procedures. The elements of the carrier sets of the export model are
constructed from the carrier sets of the import model. The implementations are
allowed to represent a single value in the export specification by several values
of the import specifications. For example, when implementing sets by lists, a
set might be represented by any list containing all elements of the set in any
order. Furthermore, we do not require that all values of a sort in the import
specification really represent a value of the export specification. In the example
below where we will implement natural numbers by binary words, we will exclude
words with leading zeroes. In order to describe the construction of the carrier
59
sets, the refinement contains two additional procedures for each sort: procedure
defining a congruence relation and a procedure defining a restriction. The restriction terminates on all elements, that represent export specification values.
The congruence relation determines the equivalence classes that represent the
elements of the export model.
A refinement is correct, if for every import model the export model constructed according to the refinement is actually a model of the export theory.
The VSE system generates proof obligations that are sufficient for the correctness.
3.4
The Deductive Component of VSE
Corresponding to the two specification formalisms, the deductive component of
VSE supports two logics: dynamic logic for abstract data type refinement and a
slightly adapted form of TLA for the state based systems. For both logics VSE
uses a sequent calculus in the style of [8].
Integration of automation with user interactions is achieved by the so called
heuristic loop. A set of heuristics is tried in turn. In case of success the heuristic
will apply a rule and increase the proof tree. In case of failure the next heuristic
is called. Heuristics can be interactive and ask the user for help, for example to
select an appropriate instantiation. There is also a last resort heuristic which
allows the user to select a rule from the set of all applicable rules.
4
Institution of Dynamic Logic
VSE provides an interactive prover, which supports a Gentzen style natural deduction calculus for dynamic logic. This logic is an extension of first-order logic
with two additional kinds of formulas that allow for reasoning about programs.
One of them is the box formula [α]ϕ, where α is a program written in an imperative language, and ϕ is a dynamic logic formula. The meaning of [α]ϕ can
be roughly put as “After every terminating execution of α, ϕ holds.”. The other
new kind of formulas is the diamond formula αϕ, which is the dual counter
part of a box formula. The meaning of αϕ can be circumscribed as “After some
terminating execution of α, which satisfies ϕ”.
4.1
Signatures
The starting point for dynamic logic signatures are the signatures of first-order
logic with equality (FOL= ) that have the form ΣFOL= = (S, F, P ) consisting of
a set S of sorts, a family F of function symbols and a family P of predicate
symbols. Because we need to name procedures, we add a S ∗ × S ∗ -sorted family
P R = (P Rv,w )v,w∈S ∗ of procedure symbols, leading to signatures of the form
Σ = (S, F, P, P R). We have two separate lists v and w of the argument sorts of
the procedure symbols in P Rv,w , in order to distinguish the sorts of the input
parameters (v) from those of the output parameters (w).
60
A signature morphism between two signatures maps sorts, operation symbols,
predicate symbols and procedure symbols in a way such that argument and result
sorts are preserved.
4.2
Sentences
Let Σ = (S, F, P, P R) be a dynamic logic signature with P R = (P Rv,w )v,w∈S ∗ .
First we define the syntax of the programs that may appear in dynamic logic
formulas. The set PΣ of Σ-programs is the smallest set containing:
–
–
–
–
–
–
–
–
abort, skip
x := τ
α; β
declare x : s = τ in α
declare x : s =? in α
if ε then α else β fi
while ε do α od
p(x1 , x2 , ..., xn , y1 , y2 , . . . , ym ) ,
where x, x1 , x2 , . . . , xn , y1 , y2 , . . . , ym are variables, τ a Σ-term of the same sort s
as x, ε a boolean Σ-formula (i.e. a Σ-Formula without quantifiers, boxes and diamonds), α, β ∈ PΣ , p a procedure symbol, such that the sorts of x1 , . . . , xn , y1 , . . . , ym
match the argument and result sorts of p.
These kinds of program statements can be explained informally as follows:
abort is a program that never terminates. skip is a program that does nothing. x := τ is the assignment. declare x = σ in α is the deterministic form
of a variable declaration which sets x to the value of σ. Its nondeterministic
form var x =? in α sets x to an arbitrary value before executing α. α; β is the
composition of the programs α and β, such that α is executed before β. The
conditional if ε then α else β fi means that α is executed if ε holds, otherwise β
is computed. Finally, the loop while ε do α od checks the condition ε, in case of
validity executes α and repeats the loop.
There are three kinds of sentences that may occur in a Σ-dynamic logic
specification.
1. Dynamic logic formulas are
– any (S, F, P )-first-order formula ϕ
– for any Σ-formula ϕ and a Σ-program α the formulas [α]ϕ and αϕ.
2. Procedure definitions are expressions of the form:
defprocs
procedure p(x1 , . . . , xn , y1 , . . . , ym )α
...
defprocsend
where p ∈ P Rv,w for some v, w ∈ S ∗ , x1 , . . . , xn , y1 , . . . , ym are variables
of the corresponding sorts in v, w, and α ∈ PΣ is a Σ-program with free
variables from the set {x1 , . . . , xn , y1 , . . . , ym }.
61
3. Restricted sort generation constraints express that a set of values defined by
restriction procedure can be generated by the given set of procedures, the
constructors. Sort generation constraints in the export specification will give
rise to proof obligations stated as a sentence of this kind. Syntactically a
restricted sort generation constraints takes the form
generated type s ::= p1 (. . . )|p2 (. . . )| . . . |pn (. . . ) restricted by r ,
(1)
where s is a sort symbol, p1 ,. . . ,pn are functional procedure symbols, the
dots in p1 (. . . ) etc. has to be replaced by a list of the argument sorts, and r
is a procedure symbol taking one argument of sort s. The meaning is, that
all elements of sort s that fulfil the restriction r, i.e. for which r terminates,
can be generated by the constructors p1 , p2 , . . . , pn .
4.3
Models
Let Σ = (S, F, P, P R) be a dynamic logic signature. A (dynamic logic) Σ-model
M is obtained from any FOL= (S, F, P )-model M F OL by supplementing it with
a mapping from the family P R of procedure symbols to relations, such that for
P R = (P Rv,w )v,w∈S ∗
–
–
–
–
Ms = MsF OL ,
(fw,s )M = (fw,s )M F OL ,
(pw )M = (pw )M F OL and
for all pr ∈ P Rv,w : prM ⊆ Mv × Mw is a relation.
4.4
Satisfaction of Dynamic Logic Formulas
Semantics is defined in a Kripke-like manner. For a given signature Σ and a
Σ-model M the (program) states are variable valuations, i.e. functions from the
fixed infinite set of variables to M , such that for each sort s variables of sort s
are mapped to values of Ms . The semantics of a program α with respect to a
model M is a predicate .αM . on two program states. qαr can be read as: If
α is started in state q it may terminate after having changed the state to r.
– qskipM q
– not qabortM r
–
–
–
–
–
–
–
σ q,M if v ≡ x
qx := σM r ⇔ r = λv.
q(v) otherwise
M
qα; β r ⇔ for some state s : qαs and sβr
qdeclare x : s = σ; αM r ⇔ qx := σ; αr
qdeclare x : s =?; αM r ⇔ for some a ∈ sM : q[x ← a]αr
qif ε then α else β fir⇔ (q |= ε and qαr) or (q |= ¬ε and qβr)
qwhile ε do α odr ⇔ q(if ε then α else skip fiM )∗ r and r |= ¬ε
qpr(x1 , . . . , xn ; y1 , . . . , ym )M r ⇔ prM (q(x1 ), . . . , q(xn ); r(y1 ), . . . , r(ym ))
62
where for any program α, (αM )∗ is the reflexive transitive closure of the relation
q(y) if y = x
, and
αM , and the state q[x ← a] is defined as q[x ← a](y) =
a
if y = x
σ q,M is the evaluation of the term σ with respect to the model M and state q.
The satisfaction relation for first-order logic (FOL= ) operators is as usual.
For the box operator we stipulate
M, q |= [α]ϕ iff for all program states r : qαM r implies M, r |= ϕ
and αϕ is defined as ¬[α]¬ϕ.
4.5
Satisfaction of Procedure Definitions
The procedures in our model will not have any side effects (except for modifying
the output parameters).
Unwinding a procedure call by replacing it by the body of the procedure and
substituting the formal parameter variables by the actual parameters should not
change the result of a program. Therefore, for a signature Σ a Σ-model M is a
model of a procedure declaration
defprocs procedure pr(x1 , . . . , xn , y1 , . . . , ym )α defprocsend
without recursion if
M |= ∀x1 , . . . , xn , r1 , . . . , rm :
(pr(x1 , . . . , xn ; y1 , . . . , ym )y1 = r1 ∧ · · · ∧ ym = rm )
⇔ αy1 = r1 ∧ · · · ∧ ym = rm
holds. We then write M ||= π. It is obvious how this can be extended to declarations of more than one procedure.
In the presence of recursion this is not sufficient to make the procedure definitions non-ambiguous and adequate to conventional semantics of programming
languages. Therefore, from several models complying with the definitions the
minimal model with respect to some order will be chosen. The order compares the
interpretations of the procedures symbols, such that the order relation M1 ≤ M2
holds for two models M1 and M2 for the same signature Σ = (S, F, P, P R), iff
prM1 ⊆ prM2 for all pr ∈ P R and the interpretations of sort, function and
predicate symbols are identical. Then we define that M satisfies π (M |= π), iff
M ||= π, and for all other models M only differing in the interpretations of the
procedure symbols pr1 , . . . , prk , M ||= π implies M ≤ M .
4.6
Satisfaction of restricted sort generation constraints
A restricted sort generation constraint as in (1) is said to hold, iff
r(a)true → uniform restr s (a)true
63
holds, where uniform restr s is a uniform restriction procedure constructed from
the procedures enumerated in the constraint. This is a procedure that nondeterministically constructs a value using the constructors. Only if the result equals
the input, the procedure terminates. According to the semantics, the existence
of one sequence of nondeterministic decisions that leads to the termination of
the program, is sufficient to consider the program to terminate.
As an example consider the constraint
generated type nat ::= i 0|i suc(nat) restricted by restr nat .
The corresponding uniform restriction uniform restr nat is defined by:
procedure uniform restr nat (x)
declare y : nat :=?;
y := i 0();
if x = y then
declare z : nat :=?;
uniform restr nat (z);
y = i succ(z);
if x = y then abort fi
fi
4.7
Calculus
The common rules for first-order logic are used in VSE. In addition, rules for the
modal operators are required. We only present some examples of the rules used
in VSE for dynamic logic to give an impression how the dynamic logic proofs
can be done:
In most cases, assignments on the right hand side are dealt with by this rule:
Γ, x = τ ϕx←x , Δ
Γ x := τ ϕ, Δ
where ϕx←x is the formula obtained by renaming all occurrences of the variable
x with the fresh variable x .
The following rule for conditionals introduces a distinction between the case
when the condition ε holds and its opposite case.
Γ, ε α ϕ, Δ
Γ, ¬ε β ϕ, Δ
Γ if ε then α else β fi ϕ, Δ
For all other kinds of statement there are also rules in VSE. Applying the
appropriate rules is in many respects like executing a the program except that
instead of computing a value in the program variables, equations between program variables and terms are derived. Therefore, this approach is called symbolic
execution. The heuristics in VSE support the selection of rules according to this
approach.
64
5
VSE Refinement as an Institution Comorphism
When integrating VSE and its notion of refinement into Hets, we want to avoid
the need for a special treatment of this refinement (one could consider adding
special refinement links in development graphs). Rather, our central idea is to
encode VSE refinement as ordinary refinement in Hets, with the help of a
comorphism that does the actually restriction-quotient construction. With this
approach, only the Hets’ logic graph needs to be extended by a logic and a
comorphism (actually, we will see that two comorphisms are necessary), while
the logic-independent part of Hets can be left untouched.
We define a comorphism CASL2V SERef ine from CF OL= to CDyn= . Each
signature (S, F, P ) is mapped to the CDyn= theory ((S, ∅, ∅, P R), E). P R contains (1) for each sort s, a symbol restr s ∈ P R[s],[] for the restriction formula on
the sort (the restriction predicate is then obtained as the set of values for which
restr s terminates) and a symbol eqs ∈ P R[s,s],[bool] for the equality on the sort
and (2) for each function symbol f : w → s ∈ Fw,s , a symbol f : w → s ∈ P Rw,[s]
and for each predicate symbol p : w ∈ Pw , a symbol p : w → [bool] ∈ P Rw,[bool] .
The procedure symbols are not assigned any definition and thus any implementation can be later plugged in using a CDyn= signature morphism.
The set of axioms E contains sentences saying that for each sort s, (1) eqs is
a congruence and it terminates for inputs satisfying the restriction and (2) the
procedures that implement functions/predicates terminate for inputs satisfying
the restriction and their results also satisfy the restriction. These properties are
to be proven when providing an actual implementation.
Given a signature (S, F, P ) and a model M of the theory ((S, ∅, ∅, P R), E),
we define the translation of M to a (S, F, P )-model, denoted M . The interpretation of a sort s in M is constructed in two steps. First we take the subset
Mrestr s ⊆ Ms of elements, for which the restriction terminates. Then we take the
quotient Mrestr s /≡ according to the congruence relation ≡ defined by eqs , such
that for all x1 , x2 ∈ sM x1 ≡ x2 is equivalent to M |= eqs (x1 , x2 ; y)y = true.
For each function symbol f , we define the value of Mf in the arguments x1 , ..., xn
to be the value returned by the call of procedure Nf on inputs x1 , .., xn , that is
Mf (x1 , ..., xn ) = b if and only if Nf (x1 , ..., xn ; y)y = b. The axioms (2) in E
ensure that Mf is total and well-defined. Similarly, for each predicate symbol p,
Mp (x1 , ..., xn ) holds iff Np (x1 , .., xn ; y)y = true.
Sentence translation is based on translation of terms into programs that
compute the representation of the term. Basically, each function application is
translated to a procedure call of the implementing procedure, and new output
variables are introduced.
– a variable x is mapped to x := x, where the left-hand side x is the output
variable and the right-hand side x is the logical variable;
– an constant c is mapped to c(; y), where c is now the procedure implementing
the constant and y is a new output variable;
– a term f (t1 , .., tn ) is mapped to α1 ; ..αn ; a := f (y1 , .., yn ), where αi is the
translation of ti with the output variable yi and a is a new output variable
65
Then
– a equation t1 = t2 is translated to
α1 ; α2 ; eqs (y1 , y2 ; y)y = true
where αi is the translation of the term ti , with the output variable yi (one
could also eliminate the assignments through substitutions);
– a predicate p(t1 , .., tn ) is translated to
α1 ..αn p(y1 , .., yn ; y)y = true
where αi is the translation of the term ti , with the output variable yi (one
could again eliminate the assignments through substitutions);
– Boolean connections of formulas are translated into the same connections of
their translated formulas;
– for universally and existentially qualified formulas one also has to make sure
that the bound variables are assigned a value that satisfies the restriction.
The complete translation is obtained by adding as hypotheses to the translation of the formula a list of formulas of type s(y)true, where y is a free variable
of the formula, of sort s.
Sort generation constraints are translated to sort generation restricted constraints over implementing procedures.
Assume we have in the abstract specification of natural numbers a sort generation constraint:
generated type nat ::= 0 | suc (nat)
Then in the resulting VSE theory, the sort generation restricted constraint
generated type nat ::= 0 | suc(nat) restricted by restr nat .
is introduced, where 0 and suc are now the procedures implementing the constructors and restr nat is the restriction procedure on sort nat.
If we implement natural numbers as binary words and have a restriction
procedure on sort bin that terminates for words with non-leading zeros:
type bin ::= b zero | b one | s0(pop :?bin)| s1(pop :?bin);
PROCEDURE nlz : IN x;
FUNCTION i zero :→ bin;
FUNCTION i succ : IN bin → bin;
then in HETS we get as proof obligation
generated type bin ::= i zero|i succ(nat) restricted by nlz
which will be discarded using the VSE mechanism of introducing a uniform
restriction which checks if an element of bin can be generated using the procedures i zero, i suc of the sort generation constraint and then proving that
whenever the restriction nlz holds on a bin value, the uniform restriction also
holds.
66
5.1
Structuring in Context of Refinement
Consider the situation where a theory of a library (e.g. the natural numbers) or
a parameter theory that will be instantiated later occurs both in the abstract
and the refined specification. Such common import specifications should not be
refined, but rather kept identically — and this is indeed the case in VSE.8
To handle this situation in the present context, the import of a first-order
specification into a dynamic logic one is not done along the trivial inclusion
comorphism from CF OL= to CDyn= — this would mean that the operations
of the import need to be implemented as procedures. Instead, we define a comorphism, CASL2V SEImport, which, besides keeping the first-order part, will
introduce for the symbols of import specification new procedure symbols, similarly to CASL2V SERef ine. The difference is that this time the interpretation
of procedure symbols is not loosely specified, but definitions are introduced in
such a way that semantics is not affected: procedures are defined to be implementing each symbol as itself. In the case of sorts, this requires that no element
is restricted and equality procedure on the sort is the set-theoretic equality. In
the case of functions, the procedure simply returns the value of the function in
the given arguments, while in the case of predicates, the procedure returns true
if the corresponding predicate holds. Sentences are translated identically, and
the model reduct keeps interpretation of symbols.
For example, let us consider the situation in the diagram below, where the
natural numbers are imported both in the abstract and the concrete specification
(and the (heterogeneous) refinement link is represented by the double arrow):
N atG
GG
ww
w
GG
ww
GG
w
w
G#
w
{w
+3 Impl
Abstr
Assume there is a sort nat in N at. When translating Abstr to CDyn= along
the refinement comorphism, we make no distinction between the sorts defined in
Abstr and the imported ones, so in the resulting translated theory we will have
symbols for the restriction on sort nat and for equality. The second comorphism,
used for imports, introduces the same symbols in Impl and imposes that they
are implemented as themselves: no element of nat is restricted and the identification is implemented as the strong equality. Thus, the expected behaviour is
obtained: Impl provides default implementations, introduced by the importing
comorphism, such that the sort nat can be regarded as kept, modulo wrapping
into procedures.
8
This resembles a bit the notion of imports of parameterised specifications in Casl,
where the import is shared between formal and actual parameter and is kept identically.
67
6
Example: Implementing natural numbers by binary
words
As an example, we present the implementation of natural numbers as lists of
binary digits, abridged from [14]. The abstract specification, nats, is the specification of natural numbers with 0, successor and addition with the usual Peano
axioms. The implementation, nats-impl, provides procedures for implementation of natural numbers as binary words, which are imported as data part along
CASL2V SEImport from the first-order specification bin (omitted here). We
illustrate the way the procedures are written with the example of the restriction
procedure, nlz, which terminates whenever the given argument has no leading
zeros. Notice that the equality is in this case simply the equality on binary words.
The view binary arith expresses the fact that binary words, restricted to
those with non-leading zeros, represent a refinement of natural numbers, where
each symbol of nats is implemented by the corresponding procedure in the symbol mapping of the view. In figure 5, we present the proof obligations generated
by the refinement comorphism and the translation of a sentence of the abstract
specification along the same comorphism.
7
Conclusions
We have integrated VSE’s mechanism of refining abstract specifications into
procedural implementations into Hets. Via a new logic and two comorphisms,
one of them doing the usual restriction-quotient construction, we could avoid
entirely the introduction of new types of “refinement links” into Hets, but
rather could re-use the present machinery of heterogeneous development graphs
and thus demomstrate its flexibility.
Recently, the interface between VSE and Hets has been extended from flat
basic specifications to structured specifications, such that Hets sends whole
development graphs to VSE. Of course, certain structuring mechanisms like heterogeneous translation and hiding are transformed away before the graph is sent
to VSE.
Future work will try to exploit synergy effects between VSE and Hets e.g.
by integrating automatic provers like SPASS into the refinement proof process.
The refinement method could also be extended from first-order logic to the richer
language Casl, which also features subsorting and partial functions.
Acknowledgements
Work on this paper has been supported by the German Federal Ministry of
Education and Research (Project 01 IW 07002 FormalSafe). We thank Werner
Stephan for conceptual discussions and Erwin R. Catesbeiana for pointing out
a class of specifications particularly easily usable as targets of refinements.
68
spec nats =
free type nats ::= zero n | succ n(nats)
op
add n : nats × nats → nats
%% axioms ...
end
logic VSE
spec nats impl =
bin with logic CASL2VSEImport
then PROCEDURES
hnlz : IN bin; nlz : IN bin; i badd : IN bin, IN bin, OUT bin, OUT bin;
i add : IN bin, IN bin → bin; i succ : IN bin → bin; i zero : → bin;
eq : IN bin, IN bin → Boolean
• DEFPROCS
PROCEDURE hnlz (x )
BEGIN
IF x = b zero THEN ABORT
ELSE IF x = b one THEN SKIP ELSE hnlz (pop(x )) FI
FI
END;
PROCEDURE nlz (x )
BEGIN IF x = b zero THEN SKIP ELSE hnlz (x ) FI END
DEFPROCSEND
%% PROCEDURE i zero(x) ...
end
view binary arith :
{ nats with logic CASL2VSERefine } to nats impl =
nats → bin, gn restr nats → nlz, gn eq nats → eq,
gn zero n → i zero, gn succ n → i succ, gn add n → i add
end
Fig. 4. Natural numbers as binary words.
69
%% Proof obligations introduced by the view
%% i-Termination
%% for i zero
∀ gn x : bin • <:i zero(gn x : bin):> <:nlz (gn x ):> true
%% ii-Congruence
%% termination
∀ gn x, gn y : bin; gn b : Boolean
• <:nlz (gn x ):> true ∧ <:nlz (gn y):> true
⇒ <:eq(gn x : bin, gn y : bin, gn b : Boolean):> true
%% reflexivity
∀ gn x : bin; gn b : Boolean • <:nlz (gn x ):> true
⇒ <:eq(gn x : bin, gn x : bin, gn b : Boolean):>
gn b = (op True : Boolean)
%% compatibility with i add
∀ gn x, gn y : bin; gn b, gn b1, gn b2 : Boolean; gn x1, gn y1, gn x2, gn y2 : bin
• ((<:nlz (gn x1 ):> true) ∧ (<:nlz (gn y1 ):> true)
∧ (<:eq(gn x1 : bin, gn y1 : bin, gn b1 : Boolean):> gn b1 = (op True : Boolean))
∧ (<:nlz (gn x2 ):> true) ∧ (<:nlz (gn y2 ):> true)
∧ <:eq(gn x2 : bin, gn y2 : bin, gn b2 : Boolean):> gn b2 = (op True : Boolean))
⇒ <:i add (gn x1 : bin, gn x2 : bin, gn x : bin):>
<:i add (gn y1 : bin, gn y2 : bin, gn y : bin):>
<:eq(gn x : bin, gn y : bin, gn b : Boolean):>
gn b = (op True : Boolean)
%% iii-export axioms
%% translation of : forall m : nats . add n(m, zero n) = m
∀ gn x0, gn x1, gn x2, gn x3 : bin; gn x4 : Boolean;
m : bin
• <:nlz (m):> true
⇒ <:gn x1 := m ;
gn x2 := i zero;
gn x0 := i add (gn x1, gn x2 );
gn x3 := m;
gn x4 := eq(gn x0, gn x3 ):>
gn x4 = (op True : Boolean)
Fig. 5. Generated proof obligations
70
References
1. E. Astesiano, M. Bidoit, B. Krieg-Brückner, H. Kirchner, P. D. Mosses, D. Sannella,
and A. Tarlecki. Casl - the common algebraic specification language. Theoretical
Computer Science, 286:153–196, 2002.
2. S. Autexier, D. Hutter, B. Langenstein, H. Mantel, G. Rock, A. Schairer,
W. Stephan, R. Vogt, and A. Wolpers. VSE: Formal methods meet industrial
needs. International Journal on Software Tools for Technology Transfer, Special
issue on Mechanized Theorem Proving for Technology, 3(1), september 2000.
3. Michel Bidoit and Peter D. Mosses. Casl User Manual, volume 2900 of LNCS
(IFIP Series). Springer, 2004. With chapters by T. Mossakowski, D. Sannella, and
A. Tarlecki.
4. Michel Bidoit, Donald Sannella, and Andrzej Tarlecki. Observational interpretation
of casl specifications. Mathematical Structures in Computer Science, 18(2):325–371,
2008.
5. R.M. Burstall and J.A. Goguen. The semantics of CLEAR, a specification language. In Proceedings of the Abstract Software Specifications, 1979 Copenhagen
Winter School, LNCS 86, pages 292–332. Springer, 1980.
6. Lassaad Cheikhrouhou, Georg Rock, Werner Stephan, Matthias Schwan, and
Gunter Lassmann. Verifying a chipcard-based biometric identification protocol
in VSE. In Janusz Górski, editor, Computer Safety, Reliability, and Security, 25th
International Conference, SAFECOMP 2006, Gdansk, Poland, September 27-29,
2006, Proceedings, volume 4166 of Lecture Notes in Computer Science, pages 42–56.
Springer, 2006.
7. CoFI (The Common Framework Initiative). Casl Reference Manual, volume 2960
of LNCS (IFIP Series). Springer, 2004.
8. Gerhard Gentzen. Untersuchungen über das logische Schließen I & II. Mathematische Zeitschrift, 39:176–210, 572–595, 1935.
9. J.A. Goguen and R.M. Burstall. Institutions: Abstract model theory for specification and programming. Journal of the ACM, 39(1):95–146, 1992.
10. Dieter Hutter, Bruno Langenstein, Georg Rock, Jörg Siekmann, Werner Stephan,
and Roland Vogt. Formal software development in the verification support environment. Journal of Experimental and Theoretical Artificial Intelligence, 12(4):383–
406, December 2000.
11. Bruno Langenstein, Roland Vogt, and Markus Ullmann. The use of formal methods
for trusted digital signature devices. In James N. Etheredge and Bill Z. Manaris,
editors, FLAIRS Conference, pages 336–340. AAAI Press, 2000.
12. D.C. Luckham, F.W. von Henke, B. Krieg-Brückner, and O. Owe. Anna, a Language for Annotating Ada Programs: Reference Manual. Number 260 in Lecture Notes in Computer Science. Springer-Verlag; D-69121 Heidelberg, Germany;
http://www.springer.de, 1987.
13. T. Mossakowski, D. Sannella, and A.Tarlecki. A simple refinement language for
Casl. In Jose Luiz Fiadeiro, editor, WADT 2004, volume 3423 of Lecture Notes in
Computer Science, pages 162–185. Springer; Berlin; http://www.springer.de, 2005.
14. Wolfgang Reif. Correctness of generic modules. In Anil Nerode and Michael A.
Taitslin, editors, LFCS, volume 620 of Lecture Notes in Computer Science, pages
406–417. Springer, 1992.
71
Mihai Codescu obtained his M.Sc at Şcoala Normală
Superioară Bucureşti, Romania, under supervision of
R. Diaconescu. Since November 2006 he joined BerndKrieg Brückner’s group as a PhD candidate. His research interests are formal methods and mathematical
logic.
Bruno Langenstein received his Diploma from the
University of Karlsruhe (TH) in 1991. After a short intermezzo as a software developer, he joined the Formal
Methods group at DFKI Saarbrücken lead by Werner
Stephan and Dieter Hutter. This research group now
has strong ties to Bernd Krieg-Brückner’s research
group in Bremen. Until today Bruno Langenstein has
participated significantly in the development of VSE
and of formal models for various industrial projects.
His research interests are formal methods with a focus
on their practical application.
Christian Maeder received his PhD from the Technical University Berlin in 2001 under supervision of Peter Pepper. In 2002 he joined Bernd Krieg-Brückner’s
group to work on Casl, HasCasl and Hets in varying
project contexts, initially employed at the University
of Bremen and since May 2006 at DFKI Bremen. His
main interests are functional programs and their verification.
Till Mossakowski met Bernd Krieg-Brückner the
first time in 1986 when he started to study computer
science, and later took part in the student’s project
PILS about implementation of a functional programming language, supervised by Bernd. He received his
PhD from the University of Bremen in 1996 under supervision of Hans-Jörg Kreowski. He then moved to
Bernd’s group (with different employment statuses),
where he also obtained his Habilitation degree with a
work on heterogeneous formal specification. Currently,
he is senior researcher in Bernd’s group at DFKI Bremen and on leave for a substitute professorship in artificial intelligence at the Albert-Ludwigs-University,
Freiburg im Breisgau. His research focus is formal
methods and heterogeneous specification, applied both
to formal software development and to qualitative spatial reasoning (the latter within the special research
center “Spatial Cognition”).
Encapsulating Lazy Behavioral Subtyping Johan Dovland, Einar Broch Johnsen, Olaf Owe, and Martin Steffen
Dept. of Informatics, University of Oslo, Norway
{johand,einarj,olaf,msteffen}@ifi.uio.no
Abstract. Object-orientation supports incremental program development by extending the class hierarchy gradually. Subclassing and late bound method calls
allow very flexible reuse of code, thereby avoiding code duplication. Combined
with incremental program development, this flexibility poses a challenge for program analysis. The dominant solution to this problem is behavioral subtyping,
which avoids re-verification of verified code but requires that all properties of
a method are preserved in subclasses. Program analysis becomes incremental,
but behavioral subtyping severely restricts code reuse. Lazy behavioral subtyping relaxes this restriction to the preservation of properties that are required by
the call-site usage of methods. Previous work developed corresponding inference
systems for languages with single- and multiple-inheritance hierarchies, but although incremental the approach could make it necessary to revisit previously
analyzed classes in order to establish new properties. This paper combines the
proof system for lazy behavioral subtyping with behavioral interfaces to support
incremental reasoning in a modular way. A class may be fully analyzed at development time by relying on interface information for external method calls. Separating classes and interfaces encapsulates the objects in a cleaner way and leads to
a simplification of the formal reasoning system. The approach is presented using
a simple object-oriented language (based on Featherweight Java) extended with
interfaces and illustrated by an example using a Hoare-style proof system.
1
Introduction
Object-orientation supports an incremental style of program development in which new
classes and subclasses may gradually be added to previously developed class hierarchies; these new subclasses typically extend and specialize existing code from superclasses, potentially overriding existing methods. In that way, the code associated with a
late bound method call depends on the run-time class of the callee object, and its effects
are not statically decidable. Subclassing combined with late binding lead to a very flexible mechanism for code reuse, as illustrated through a plethora of design patterns [15],
but pose a challenge for program analysis. The intricacies of late-binding, inheritance,
encapsulation, and other advanced features of object-oriented languages spawned a lot
of research to clarify the semantical foundations, and especially to capture such features
in a type-safe manner. This led to the development of expressive type systems and calculi. In this line of research, Qian and Krieg-Brückner developed a language combining
This research is partially funded by the EU project IST-33826 CREDO: Modeling and analysis
of evolutionary structures for distributed services (http://credo.cwi.nl).
73
features of object-oriented and functional languages in a unified, typed calculus [26],
where much of the expressive power is needed to capture late-binding and overloading.
The static, type-theoretic foundations of mainstream object-oriented languages are
now largely understood, but verification support for object-oriented languages still poses
significant challenges. There are two main approaches in the literature to the verification
of class hierarchies with late bound method calls. Behavioral subtyping was originally
proposed by America [2] and Liskov and Wing [20] and later used in, e.g., Spec [19].
This is an open world approach: it facilitates reasoning about programs in an incremental way which fits well with the incremental development style of object-oriented
programs. Roughly speaking, the basic idea of behavioral subtyping is that any property
of a supertype should also hold for all subtypes. The approach focuses on the declared
properties of a type, and applied to the object-oriented setting, any property of a superclass should also hold for all subclasses. The idea is appealing, as it provides substitutability not just for the static signatures of objects, as in standard subtyping, but also
at the behavioral level. Behavioral subtyping, however, imposes severe restrictions on
subclassing, limiting how code may be reused in ways which break with programming
practice [27]. For example, the class hierarchies of Java libraries do not obey the behavioral subtyping discipline. Alternatively, one can accept the practice of unrestricted
code reuse and overriding, and capture that in a reasoning system. For instance, Pierik
and de Boer [25] have proposed a complete proof system for object-oriented programs
which is able to address much more flexible code reuse. However, it is a closed world
approach: it requires the full class hierarchy to be available at analysis time to ensure
that any binding meets the requirements imposed by the usage of values of the type.
The approach focuses on the required properties of a type. It overcomes the limitations
of behavioral subtyping, but breaks with incremental reasoning.
Lazy behavioral subtyping has been proposed by the authors to preserve the appealing features of behavioral subtyping, i.e., incremental reasoning, but allow more flexible
code reuse in a controlled way. Lazy behavioral subtyping balances the required properties reflecting the call-site use of a method with its provided properties, and the basic
insight is that the properties that need to be preserved depend on the use of a method
rather than on its declared contract. Previous use, therefore, imposes restrictions on future redefinitions in order to maintain the incremental reasoning property. The approach
is supported by an inference system which tracks declaration site specifications and call
site requirements for methods in an extensible class hierarchy [13]. This inference system, which is independent from the underlying specific program logic of a given reasoning system, ensures that proven properties are not violated due to method redefinition in
new subclasses, and that required properties of a method always hold for redefinitions.
The approach has later been extended to deal with multiple inheritance [14].
These previous papers present a slightly simplistic version of lazy behavioral subtyping, concentrating on the core mechanisms for flexible code reuse without breaking
the principle of incremental reasoning. In particular, we considered a language without
interfaces i.e., classes played the roles of types for objects and of generators of objects
instances at the same time. As a consequence, external method calls could recursively
lead to new proof obligations in previously analyzed classes. In this paper, we aim to
combine lazy behavioral subtyping with a notion of modularity for external calls, so
74
P ::= K L {t}
MS ::= m (x) : (p, q)
M ::= m (x) {t}
K
L
e
t
::=
::=
::=
::=
interface I extends I {MS}
class C extends C implements I { f M MS inv p}
f | this | b | new C() | e.m(e) | m(e)
f := e | return e | skip | if b then t else t fi | t;t
Fig. 1. The language syntax, where I, C and, m are interface, class, and method names
(of types Iid, Cid, and Mid, respectively), and p and q are assertions. Vector notation
denotes lists, as in the expression list e.
that lazy behavioral subtyping applies to internal code reuse and additional proof obligations are avoided for classes which have already been analyzed. For this purpose, the
type hierarchy will be separated from the class hierarchy, and behavioral interfaces are
introduced to type object variables and references. Thus, a class which inherits a superclass need not inherit the type of the superclass, and may thereby reuse code more freely.
As this approach can be encoded in the general method of lazy behavioral subtyping,
the soundness of the proof system of this paper follows directly from the soundness of
the pure lazy behavioral subtyping method (see [13]).
The remainder of the paper is structured as follows. Section 2 presents a variant of
Featherweight Java as the language we use for our development. Section 3 presents the
data structures needed to keep track of the different proof obligations, and Section 4 the
inference system to analyse a class hierarchy. The method is illustrated on an example
in Section 5, and Section 6 discusses the context of this work.
2
The Programming Language
We consider a programming language based on Featherweight Java [18], but extended
with (behavioral) interfaces. A program P consists of a set K of interfaces, a set L of
classes, and an initial statement t. The syntax is given in Fig. 1 and explained below.
2.1
Behavioral Interfaces
A behavioral interface consists of a set of method names with signatures and semantic
constraints on the use of these methods. In Fig. 1, an interface I may extend a list I of superinterfaces, and declare a set MS of method signatures, where behavioral constraints
are given as specifications (pre, post) of pre- and postconditions to the signatures. An
interface may declare signatures of new methods not found in its superinterfaces, and
it may declare additional specifications of methods declared in the superinterfaces. The
relationship between interfaces is restricted to a form of behavioral subtyping. An interface may extend several interfaces, adding to its superinterfaces new syntactic and
semantic constraints. We assume that the interface hierarchy conforms with these requirements. The interfaces thus form a type hierarchy: if I extends I, then I is a subtype
of I and I is a supertype of I . Let denote the reflexive and transitive subtype relation,
75
which is given by the nominal extends-relation over interfaces. Thus, I I if I equals
I or if I extends (directly or indirectly) I.
An object supports an interface I if the object provides the methods declared in I
and adheres to the specifications imposed by I on these methods. Fields are typed by
interfaces; if an object supports I then the object may be referenced by a field typed by I.
A class implements an interface if its code is such that all instances support the interface.
The analysis of the class must ensure that this requirement holds. Objects of different
classes may support the same interface, corresponding to different implementations of
its behavior. Only the methods declared by I are available for external invocations on
references typed by I. The class may implement additional auxiliary methods.
The substitution principle for objects applies to the level of interfaces: an object
supporting an interface I may be replaced by another object supporting I or a subtype
of I. A subclass C of C need not satisfy the interface I of the superclass. If I is not
implemented by C , the substitution principle ensures that an instance of C cannot be
used where an object of type I is expected. If a field x is declared with interface I,
the actual object referenced by x at run-time will satisfy the behavioral specification of
I. However, as object references are typed by interface, the run-time class of a called
object is hidden by the behavioral interface of that object. Consequently, all external
method calls are late bound.
2.2
The Imperative Language
The imperative part of the language consists of classes which may implement an interface, inherit from a superclass, and define fields f , methods M, and method specifications MS (see Fig. 1). The superclass is given by the extends clause in the
class header, and the interface supported by instances of the class is given by the
implements clause. The syntactic parts of a class are referred to by the functions
inh, att, mtds, and impl, returning the superclass name, attributes, methods, and interface, respectively. Let ≤ denote the reflexive and transitive subclass relation, such that
C ≤ C if C equals C, or C extends (directly or indirectly) C.
For analysis purposes, a class may specify an invariant inv p, where p is a predicate
over the fields of the class (implemented directly or inherited). As the interface of a class
C hides the implementation details of C, the class invariant is also hidden. An external
call x.m(), where x refers to an instance of class C, cannot assume that the invariant
of x holds when the method starts execution. In the imperative language, expressions e
include program variables f and Boolean expressions b, external calls e.m(e), and self
calls m(e). If m does not return a value, or if the returned value is of no concern, we
may use e.m(e) and m(e) directly as statements for simplicity (ignoring the assignment
of the return value to a program variable). Note that the list of actual parameter values
may be empty and that the formal parameters x and the reserved variable this (for self
reference) are read-only variables. Statements include assignment f := e, return e
which returns an expression e to the caller, conditionals, and sequential composition.
For simplicity, all method calls in this language are late bound. Method binding searches
the class hierarchy from the actual class of the called object in the usual way.
A user given specification m (x) : (p, q) applies to the definition of m in C or to
the inherited definition of m (without redefinition) from the superclass of C. Thus, a
76
subclass may provide additional specifications of methods inherited from superclasses.
For convenience, we use the syntax m (x) : (p, q){t} to both define and specify m.
3
Class Analysis
An essential part of the verification of a class is to ensure that the methods defined by
the class support the behavioral specification of the interface implemented by the class.
Methods are verified in terms of proof outlines [23]; i.e., m(x) : (p, q){t} such that t is a
method body decorated with pre/post requirements on method calls and PL {p} t {q}
is derivable in the given program logic PL if the requirements hold for the method calls
in t. Let body(C, m) denote the decorated method body of m in C. The body is either
found in a definition of m in C, or inherited from a superclass of C.
Notation. Given assertions p and q, the type APair ranges over assertion pairs (p, q). If
p is the pre- and q the postcondition to some method, the pair (p, q) is a specification
of that method. For an interface I, public(I) denotes the set of method identifiers supported by I; thus, m ∈ public(I) if m is declared by I or by a supertype of I. A subtype
cannot remove methods declared by a supertype, so public(I) ⊆ public(I ) if I I. If
m ∈ public(I), the function spec(I, m) returns a set of type Set[APair] with the behavioral
specifications supported by m in I, as declared in I or in a supertype of I. The function
returns a set since a subinterface may provide additional specifications of methods inherited from superinterfaces; if m ∈ public(I) and I I, then spec(I, m) ⊆ spec(I , m).
Consequently, entailment (denoted ) is defined between sets of assertion pairs.
Definition 1 (Entailment). Assume given assertion pairs (p, q) and (r, s), and sets U =
{(pi , qi ) | 1 ≤ i ≤ n} and V = {(ri , si ) | 1 ≤ i ≤ m}, and let p be the assertion p with all
fields f substituted by f , avoiding name capture. Entailment is defined by
i) (p, q) (r, s) (∀z . p ⇒ q ) ⇒ (∀y . r ⇒ s ),
where z and y are the logical variables in (p, q) and (r, s), respectively
V
ii) U (r, s) ( 1≤i≤n (∀zi . pi ⇒ qi )) ⇒ (∀y . r ⇒ s ) .
V
iii) U V 1≤i≤m U (ri , si ) .
The relation U (r, s) corresponds to Hoare-style reasoning, proving {r} t {s} from
{pi } t {qi } for all 1 ≤ i ≤ n, by means of the adaptation and conjunction rules [3].
Entailment is reflexive and transitive, and V ⊆ U implies U V .
Lazy behavioral subtyping is a method for reasoning about redefined methods and late
binding which may be explained as follows. Let m be a method defined in a class C.
The declared behavior of this method definition is given by the specification set S(C, m),
where S is the specification mapping taking a class and a method name as arguments.
We assume that for each (p, q) ∈ S(C, m) there is a proof outline for body(C, m) such
that PL {p} body(C, m) {q}. For a self call {r} n(e) {s} in the proof outline, the assertion pair (r, s) is a requirement imposed by C on possible implementations of n to
which the call can bind. (Requirements made by external calls are considered below.)
Each such requirement, collected during the analysis of C, is included in the set R(C, n),
77
where R is the requirement mapping. Lazy behavioral subtyping ensures that all requirements in the set R(C, n) follow from our knowledge of the definition of method n in C;
i.e., S(C, n) R(C, n). If n is later overridden by some subclass D of C, the same requirements apply to the new version of n; i.e., S(D, n) R(C, n) must be proved. This
yields an incremental reasoning strategy.
In general, we let S↑(C, m) return the accumulated specification set of m in C. If m
is defined in C, this is the set S(C, m). If m is inherited, the set is S(C, m) ∪ S↑(C.inh, m),
which captures that a subclass may provide additional specifications of methods that
are inherited from superclasses. The requirements toward m that are recorded during
the analysis of superclasses are returned by the set R↑(C, m) such that R↑(C, m) =
R(C, m) ∪ R↑(C.inh, m).1 For each class C and method m defined in C, the lazy behavioral subtyping calculus (see Sec. 4) maintains the relation S↑(C, m) R↑(C, m).
If C implements an interface I, the class defines (or inherits) an implementation of
each m ∈ public(I). For each such method, the behavioral specification declared by I
must follow from the method specification, i.e., S↑(C, m) spec(I, m). Now consider
the analysis of a requirement stemming from the analysis of an external call in some
proof outline. In the following, we denote by x : I.m the external call x.m where x is
declared with static type I. As the interface hides the actual class of the object referenced
by x, the call is analyzed based on the interface specification of m. For an external call
{r} x : I.m() {s}, the requirement (r, s) must follow from the specification of m given
by type I, expressed by spec(I, m) (r, s). Soundness in this setting is given by the
following argument. Assume that the call to x.m can bind to m on an instance of class
C, and let I = C.impl. Type analysis then ensures that I I. During the analysis of C,
the relation S↑(C, m) spec(I , m) is established. The desired S↑(C, m) spec(I, m)
then follows since spec(I, m) ⊆ spec(I , m) when I I.
The invariant p of a class C is taken as a pre/post specification of each method
visible through the supported interface of C. Thus, the invariant is analyzed by proving
the specification (p, p) for each such method m. In this manner, the invariant analysis
is covered by the general approach of lazy behavioral subtyping as the declaration of
an invariant can be considered as an abbreviation for a pre/post specification of each
method. Note that this approach does not require that the invariant holds whenever m
starts execution; the specification expresses that if the invariant holds prior to method
execution, then it will also hold upon termination. This approach to invariants works
when analyzing extensible class hierarchies [27], even if the invariant of a subclass is
different from the superclass invariant. The superclass invariant need not hold in the
subclass, but methods defined in the superclass can be inherited by the subclass.
4
The Inference System
Classes and interfaces are analyzed in the context a proof environment. The proof environment tracks the specifications and requirements for the different classes, and the
1
Note that for the language considered in this paper, the set of requirements could be made
more fine-grained by removing requirements stemming from redefined method definitions.
However, in a language with static calls, this simplification would no longer apply.
78
A ::= P | C : O · P
P ::= K | L | P · P
O ::= ε | anReq(M) | anSpec(MS) | verify(m, R)
| anOutln(t) | intSpec(m) | inv(p, m) | O · O
Fig. 2. Syntax for analysis operations.
interface specifications to which each class must adhere. Let Cid, Iid, and Mid denote
the types of class, interface, and method names, respectively.
Definition 2 (Proof environments). A proof environment E of type Env is a tuple
L, K, S, R where L : Cid → Class, K : Iid → Interface are partial mappings and S, R :
Cid × Mid → Set[APair] are total mappings.
Subscripts are used to refer to a specific environment; e.g., SE is the S-mapping
of E . Now, environment soundness is defined. The definition is adapted from [13] by
taking interfaces into account. Condition 3 below captures interface implementations,
requiring that each method satisfies the behavioral specification given by the interface.
Definition 3 (Sound environments). A proof environment E is sound if it satisfies the
following conditions for each C : Cid and m : Mid.
1. ∀(p, q) ∈ SE (C, m) . ∃bodyE (C, m) . PL {p} body(C, m) {q}
∧ ∀{r} n {s} ∈ bodyE (C, m) . RE (C, n) (r, s)
∧ ∀{r} x : I.n {s} ∈ bodyE (C, m) . spec(I, n) (r, s)
2. S↑E(C, m) R↑E(C, m)
3. ∀n ∈ public(I) . S↑E(C, n) spec(I, n), where I = C.impl.
There are four operations to update a proof environment; these load a new class L
or interface K, and extend the commitment and requirement mappings with a pair (p, q)
for a given method m and class C. We define an operator ⊕ : Env × Update → Env,
where the first argument is the current proof environment and the second argument is
the environment update, as follows:
E ⊕ extL(C, D, I, f , M) = LE [C "→ D, I, f , M], KE , SE , RE E ⊕ extK(I, I, MS) = LE , KE [I "→ I, MS], SE , RE E ⊕ extS(C, m, (p, q)) = KE , KE , SE [(C, m) "→ SE (C, m) ∪ {(p, q)}], RE E ⊕ extR(C, m, (p, q)) = LE , KE SE , RE [(C, m) "→ RE (C, m) ∪ {(p, q)}]
In the calculus, judgments have the form E A , where E is the proof environment
and A is a sequence of analysis operations (see Fig. 2). The main inference rules are
given in Fig. 3. The operations and the calculus are discussed below. We emphasize
the differences wrt. the calculus in [13], which reflect the introduction of interfaces and
class invariants. In the rules, I ∈ E and C ∈ E denote that KE (I) and LE (C) are defined, respectively. For brevity, we elide a few straightforward rules which formalize a
lifting from single elements to sets or sequences of elements. For example, the rule for
anReq(M) (see the premise of (N EW C LASS )) generalizes the analysis of a single method
done in (N EW M TD ). The omitted rules may be found in [13] and are similar to the decomposition rules (D ECOMP I NT ) and (D ECOMP I NV ) for interface and invariant requirements.
79
I∈
/E
I = nil ⇒ I ∈ E
E ⊕ extK(I, I, MS) P
E (interface I extends I {MS}) · P
I∈E
(N EW I NT )
C∈
/E
D = nil ⇒ D ∈ E
E ⊕ extL(C, D, I, f , M) C : anSpec(MS) · anReq(M) · inv(p, publicE (I)) · intSpec(publicE (I)) · P
E (class C extends D implements I { f M MS inv p}) · P
E C : verify(m, R↑E(PE (C).inh, m)) · O · P
(N EW M TD )
E C : anReq(m(x){t}) · O · P
E C : verify(m, (p, q)) · O · P
E C : anSpec(m(x) : (p, q)) · O · P
(N EW C LASS )
(N EW S PEC )
S↑E(C, m) (p, q)
E C : O · P
E C : verify(m, (p, q)) · O · P
(R EQ D ER )
PL {p} bodyE (C, m) {q}
E ⊕ extS(C, m, (p, q)) C : anOutln(bodyE (C, m)) · O · P
E C : verify(m, (p, q)) · O · P
E ⊕ extR(C, m, (p, q)) C : verify(m, (p, q)) · O · P
E C : anOutln({p} m {q}) · O · P
I∈E
(R EQ N OT D ER )
(L ATE C ALL )
specE (I, m) (p, q)
E C : O · P
E C : anOutln({p} x : I.m {q}) · O · P
(E XT C ALL )
S↑E(C, m) specE (C.impl, m)
E C : O · P
E C : intSpec(m) · O · P
(I NT S PEC )
E C : verify(m, (p, p)) · O · P
E C : inv(p, m) · O · P
E C : inv(p, m1 ) · inv(p, m2 ) · O · P
E C : inv(p, m1 ∪ m2 ) · O · P
(I NV )
(D ECOMP I NV )
E C : intSpec(m1 ) · intSpec(m2 ) · O · P
E C : intSpec(m1 ∪ m2 ) · O · P
E P
E C : ε · P
(D ECOMP I NT )
(E MP C LASS )
Fig. 3. The inference system, where P is a (possibly empty) sequence of classes and
interfaces. For simplicity, m denotes a method call including actual parameters and nil
the empty list.
A program is analyzed as a sequence of interfaces and classes. For simplicity, we
require that superclasses appear before subclasses and that interfaces appear before they
80
are used. This ordering ensures that requirements imposed by superclasses are verified
in an incremental manner on subclass overridings. Rules (N EW I NT ) and (N EW C LASS ) extend
the environment with new interfaces and classes; the introduction of a new class leads
to an analysis of the class. The specification and requirement mappings are extended
based on the internal analysis of each class. We assume that programs are well-typed.
Especially, if a field x is declared with type I and there is a call to a method m on x, then
m is assumed to be supported by I. Rule (N EW C LASS ) generates an operation of the form
C : O , where O is a sequence of analysis operations to be performed for class C. Note
that (N EW I NT ) and (N EW C LASS ) cannot be applied while a C : O operation is analyzed,
which ensures that C : O is analysed before a new class is analyzed. A successful
analysis of C yields an operation C : ε which is discarded by (E MP C LASS ).
For a class C implementing an interface I, (N EW C LASS ) generates four initial operations anSpec, anReq, inv, and intSpec. Operation anSpec is analyzed by (N EW S PEC ),
initiating a verification of the user given specifications with regard to their respective
implementations. For each method m defined in C, anReq collects the inherited requirements toward m which are analyzed in (N EW M TD ). Rule (I NV ) analyses the class invariant
as a pre/post specification of each externally visible method in C. Finally, (I NT S PEC ),
ensures that the implementation of C satisfies the behavioral specification of I.
Specifications are verified by (R EQ D ER ) or (R EQ N OT D ER ). If a method specification follows from previously proven specifications of the method, the specification is discarded
by (R EQ D ER ). Otherwise, (R EQ N OT D ER ) leads to the analysis of a new proof outline for the
method. In such proof outlines, external calls {r} x : I.m() {s} are handled by (E XT C ALL ),
which ensures that (p, q) follows from the specification spec(I, m) of m in I. Internal
calls are handled by (L ATE C ALL ), which ensures that the method definitions to which the
call may be bound satisfy the requirement (p, q).
5
Example
In this section we illustrate our approach by a small account system implemented by
two classes: PosAccount and a subclass FeeAccount. The example illustrates how interface encapsulation and the separation of class inheritance and subtyping facilitate
code reuse. Class FeeAccount reuses the implementation of PosAccount, but the type
of PosAccount is not supported by FeeAccount. Thus FeeAccount does not represent
a behavioral subtype of PosAccount.
A system of communicating components can be specified in terms of the observable
interaction between the different components [7, 17]. In an object-oriented setting with
interface encapsulation, the observable interaction of an object may be described by the
communication history, which is a sequence of invocation and completion messages of
the methods declared by the interface. At any point in time, the communication history
abstractly captures the system state. Previous work [12] illustrates how the observable
interaction and the internal implementation of an object can be connected. Expressing
pre- and postconditions to methods declared by an interface in terms of the communication history allows abstract specifications of objects supporting the interface. For this
purpose, we assume an auxiliary variable h of type Seq[Msg] (sometimes called a mythical program variable [11]), where Msg ranges over invocation and completion (return)
81
messages to the methods declared by the interface. However, for the example below it
suffices to consider only completion messages. Thus, a history h will be constructed as
a sequence of completion messages by the empty (ε) and right append (·) constructor.
We write completion messages on the form o, m(x, r) where m is a method completed
on object o, x is the actual parameter values for this method execution, and r is the return value. For reasoning purposes, such a completion message is implicitly appended
to the history at each method termination, as a side effect of the return statement.
5.1
Class PosAccount
Interface IPosAccount supports the three methods deposit, withdraw, and getBalance.
The current balance of the account is abstractly captured by the function Val(h) defined
below, and the three methods maintain Val(h) ≥ 0. Method deposit deposits an amount
as specified by the parameter value and returns the current balance after the deposit,
and method getBalance returns the current balance. Method withdraw returns true if
the withdrawal succeeded, and false otherwise. A withdrawal succeeds only if it leads
to a non-negative balance. In postconditions we let return denote the returned value.
interface IPosAccount {
int deposit(nat x) : (Val(h) ≥ 0, return = Val(h) ∧ return ≥ 0)
bool withdraw(nat x) : (Val(h) ≥ 0 ∧ h = h0 , return = Val(h0 ) ≥ x ∧ Val(h) ≥ 0)
int getBalance() : (Val(h) ≥ 0, return = Val(h) ∧ return ≥ 0)
}
where
Val(ε)
0
Val(h · o, deposit(x, r)) Val(h) + x
Val(h · o, withdraw(x, r)) if r then Val(h) − x else Val(h) fi
Val(h · others)
Val(h)
This interface is implemented by a class PosAccount, given below. The balance is
maintained by a variable bal , and the corresponding invariant expresses that the balance
equals Val(h) and remains non-negative. Notice that the invariant bal = Val(h) connects
the internal state of PosAccount objects to their observable behavior, and is needed in
order to ensure the postconditions declared in the interface.
class PosAccount implements IPosAccount {
int bal = 0;
int deposit(nat x) : (true, return = bal) {update(x); return bal}
bool withdraw(nat x) : (bal = b0 , return = b0 ≥ x) {
if (bal >= x) then update(-x);
return true else return false fi}
int getBalance() : (true, return = bal) {return bal}
void update(int v) : (bal = b0 ∧ h = h0 , bal = b0 + v ∧ h = h0 ) {
bal := bal + v}
inv bal = Val(h) ∧ bal ≥ 0 }
Notice that the method update is hidden by the interface. This means that the method
is not available to the environment, but only for internal use. The following simple definition of withdraw maintains the invariant of the class as it preserves bal = Val(h):
bool withdraw(int x) {return false}
82
However, this implementation does not meet the interface specification which requires
that the method must return true if the withdrawal can be performed without resulting
in a non-negative balance. Next we consider the verification of class PosAccount.
Pre- and postconditions. The pre- and postconditions in the definition of PosAccount
lead to the following extensions of the S mapping:
(true, return = bal) ∈ S(PosAccount, deposit)
(bal = b0 , return = b0 ≥ x) ∈ S(PosAccount, withdraw)
(true, return = bal) ∈ S(PosAccount, getBalance)
(bal = b0 ∧ h = h0 , bal = b0 + v ∧ h = h0 ) ∈ S(PosAccount, update)
(1)
(2)
(3)
(4)
These specifications are easily verified over their respective method bodies.
Invariant analysis. Rule (I NV ) of Fig.3 initiates the analysis of the class invariant wrt.
the methods deposit, withdraw, and getBalance. By (R EQ N OT D ER ), the invariant is treated
as a specification of these methods:
(bal = Val(h) ∧ bal ≥ 0, bal = Val(h) ∧ bal ≥ 0) ∈ S(PosAccount, m),
(5)
for m ∈ {deposit, withdraw, getBalance}. Methods deposit and withdraw perform self
calls to update, which result in the following two requirements:
R(PosAccount, update) =
{ (bal = Val(h) ∧ bal ≥ 0 ∧ v ≥ 0, bal = Val(h) + v ∧ bal ≥ 0),
(bal = Val(h) ∧ v ≤ 0 ∧ bal + v ≥ 0, bal = Val(h) + v ∧ bal ≥ 0) }
(6)
These requirements are proved by entailment from Equation 4.
Interface specifications. At last, we must verify that the implementation of each method
defined by interface IPosAccount satisfies the corresponding interface specification,
according to (I NT S PEC ). For getBalance, it can be proved that the method specification,
as given by Specifications 3 and 5, entails the interface specification
(Val(h) ≥ 0, return = Val(h) ∧ return ≥ 0)
Verification of the other two methods follows the same outline, and this concludes the
verification of class PosAccount.
5.2
Class FeeAccount
Interface IFeeAccount resembles IPosAccount, as the same methods are supported.
However, IFeeAccount takes an additional fee for each successful withdrawal, and the
balance is not guaranteed to be non-negative. For simplicity we take fee as a (read-only)
parameter of the interface and of the class (which means that it can be used directly in
the definition of Fval below).
83
interface IFeeAccount(nat fee) {
int deposit(nat x) : (N(h), return = Fval(h) ∧ N(h))
bool withdraw(nat x) : (N(h) ∧ h = h0 , return = Fval(h0 ) ≥ x ∧ N(h))
int getBalance() : (N(h), return = Fval(h) ∧ N(h)) }
where
N(h)
Fval(h) ≥ ¬fee
Fval(ε)
0
Fval(h · o, deposit(x, r)) Fval(h) + x
Fval(h · o, withdraw(x, r)) if r then Fval(h) − x − fee else Fval(h) fi
Fval(h · others)
Fval(h)
Note that IFeeAccount is not a behavioral subtype of IPosAccount: a class that implements IFeeAccount will not implement IPosAccount. Informally, this can be seen
from the postcondition of withdraw. For both interfaces, withdraw returns true if the
parameter value is less or equal to the current balance, but IFeeAccount takes an additional fee in this case, which possibly decreases the balance to ¬fee.
Given that the implementation provided by class PosAccount is available, it might
be feasible to reuse the code of this class when implementing IFeeAccount. In fact, only
withdraw needs reimplementation, which is illustrated by the class FeeAccount below.
This class implements IFeeAccount and extends the implementation of PosAccount,
which means that the interface supported by the superclass is not supported by the
subclass. Typing restrictions prohibit that methods on an instance of FeeAccount are
called through the superclass interface IPosAccount.
class FeeAccount(int fee)
extends PosAccount implements IFeeAccount {
bool withdraw(nat x) : (bal = b0 , return = b0 ≥ x) {
if (bal >= x) then update(-(x+fee));
return true else return false fi }
inv bal = Fval(h) ∧ bal ≥ −fee
}
Pre- and postconditions. As the methods deposit and getBalance are inherited without
redefinition, the specifications of these methods can be relied on in the context of the
subclass. Especially, Specifications 1, 3, and 4 above remain valid. For withdraw, the
declared specification can be proved:
(bal = b0 , return = b0 ≥ x) ∈ S(FeeAccount, withdraw)
(7)
Invariant verification. The subclass invariant can be proved over the inherited methods
deposit and getBalance in addition to the overridden method withdraw. For deposit,
the following requirement on update is included in the requirement mapping:
(bal = Fval(h) ∧ bal ≥ −fee ∧ v ≥ 0, bal = Fval(h) + v ∧ bal ≥ −fee)
∈ R(FeeAccount, update)
84
This requirement follows from Specification (4) of update. Analysis of withdraw gives
the following requirement, which also follows from Specification 4:
(bal = Fval(h) ∧ bal ≥ x ∧ x ≥ 0 ∧ v = −(x + fee),
bal = Fval(h) − (x + fee) ∧ bal ≥ −fee)
∈ R(FeeAccount, update)
Interface specification. Now reconsider getBalance. After analyzing the subclass invariant, the specification of getBalance is given by S↑(FeeAccount, getBalance), defined as follows:
S↑(FeeAccount, getBalance) =
{(bal = Val(h) ∧ bal ≥ 0, bal = Val(h) ∧ bal ≥ 0),
(true, return = bal),
(bal = Fval(h) ∧ bal ≥ −fee, bal = Fval(h) ∧ bal ≥ −fee)}
(8)
which is the specification set that can be assumed to prove the interface specification
(Fval(h) ≥ −fee, return = Fval(h) ∧ Fval(h) ≥ −fee)
(9)
Specification 9 can be proved by entailment from Specification 8 using (I NT S PEC ). Note
that the superclass invariant is not established by the precondition of Specification 9,
which means that the inherited invariant cannot be assumed when establishing the postcondition of Specification 9. The other inherited specification is however needed, expressing that return equals bal. Verification of the interface specifications for deposit
and withdraw then follows the same outline.
6
Discussion
The notion of behavioral subtyping, i.e., to require substitutability not just for static
object signatures but also for object behavior, goes back to America [2] and Liskov and
Wing [20]. Since then, it has both been adopted and developed further, and at the same
time criticised for being too restrictive to reflect the situation in actual class hierarchies.
For example, Wehrheim has studied variations of behavioral subtyping characterized
by different notions of testing in the context of CSP processes [29]. Recent advances in
program development platforms [5, 8] and in theorem proving technology for program
verification [6] make the development of more fine-grained systems for incremental
reasoning interesting, as a tool is able to collect and often automatically discharge proof
obligations during program development.
Related to the concept of behavioral subtyping is the notion of refinement. In an
object-oriented setting, Back, Mikhajlova, and von Wright propose class refinement
and use the refinement calculus to reason about substitutability for object-oriented programs [4]. Similarly, Utting [28] and Mikhajlova and Sekerinsky [22] deal with modular
reasoning for object-oriented programs using the refinement calculus.
Putting the emphasis not on how to avoid reverification on the client side of a
method call, but for the designer of a derived class, Soundarajan and Fridella [27] separate two specifications for each class, an abstract specification for the clients of the class
85
and a concrete specification for the derived subclasses. Like the current work, they aim
at relaxing behavioral subtyping and especially separating subclassing/inheritance from
behavioral subtyping to gain flexibility while maintaining proof reuse. Lately, incremental reasoning, both for single and multiple inheritance, has been considered in the
setting of separation logic [21, 9, 24]. These approaches support a distinction between
static specifications, given for each method implementation, and dynamic specifications
used to verify late bound calls. The dynamic specifications are given at the declaration
site, in contrast to our work where late bound calls are verified based on call-site requirements. Ideas from behavioral subtyping have also been used to support modular
reasoning for aspect-oriented programming [10] and for active objects [16].
We propose lazy behavioral subtyping to relax some of the restrictions for incremental development of classes imposed by behavioral subtyping, namely to require
preservation only for those properties actually needed for client side verification of
methods. The system is syntax-driven and should be possible to integrate in program
development platforms. In this paper, lazy behavioral subtyping has been combined
with interface encapsulation, allowing code reuse internally while relying on behavioral
interfaces for external method calls. This combination ensures not only that analysis is
incremental but also that analysis is modular; i.e., a class need only be considered once
during the program analysis. Revisiting a class for further analysis, which could happen
in previous work on lazy behavioral subtyping, is thereby avoided.
To use our system, the programmer is expected to provide specifications of interfaces and classes (by means of invariants or pre/post pairs), such that for each class the
specifications given entail the corresponding interface specifications. All other specifications and requirements are generated by the system. We have illustrated the approach
by a simple bank example where the subclass does not represent a behavioral subtype
of the superclass. Both the reuse of code and the reuse of proofs are demonstrated in
the example. At the same time, client-side reasoning may fully exploit the different
properties of the two classes, due to the presence of behavioral interfaces.
In future work we plan to investigate the possibilities of letting interfaces also influence the reasoning of self calls in a more fine-grained manner, with the aim of obtaining
even weaker requirements to redefinitions. Furthermore, the extension of the work presented in this paper to multiple inheritance could follow the approach of [14].
Acknowledgements. The authors are indebted to Bernd Krieg-Brückner, Ole-Johan
Dahl, and Neelam Soundarajan for cooperation and inspiration, the two latter for their
insight in object-orientation and formal methods, and Bernd Krieg-Brückner for his
insight in program specification and program verification, and for championing the importance of formal methods in computer science.
References
1. ACM. 37th Annual Symposium on Principles of Programming Languages (POPL), Jan.
2008.
2. P. America. Designing an object-oriented programming language with behavioural subtyping. In J. W. de Bakker, W.-P. de Roever, and G. Rozenberg, editors, Foundations of
Object-Oriented Languages, pages 60–90. Springer, 1991.
86
3. K. R. Apt. Ten years of Hoare’s logic: A survey — Part I. ACM Transactions on Programming Languages and Systems, 3(4):431–483, Oct. 1981.
4. R.-J. Back, A. Mikhajlova, and J. von Wright. Class refinement as semantics of correct object
substitutability. Formal Aspects of Computing, 12(1):18–40, 2000. Also as Turku Center of
Computer Science, TUCS Technical Report No. 333, March 2000.
5. M. Barnett, K. R. M. Leino, and W. Schulte. The Spec# programming system: An overview.
In G. Barthe, L. Burdy, M. Huisman, J.-L. Lanet, and T. Muntean, editors, Intl. Workshop on
Construction and Analysis of Safe, Secure, and Interoperable Smart Devices (CASSIS’04),
volume 3362 of LNCS, pages 49–69. Springer, 2005.
6. B. Beckert, R. Hähnle, and P. H. Schmitt, editors. Verification of Object-Oriented Software.
The KeY Approach, volume 4334 of LNAI. Springer, 2007.
7. M. Broy and K. Stølen. Specification and Development of Interactive Systems. Monographs
in Computer Science. Springer, 2001.
8. L. Burdy, Y. Cheon, D. R. Cok, M. D. Ernst, J. R. Kiniry, G. T. Leavens, K. R. M. Leino,
and E. Poll. An overview of JML tools and applications. International Journal on Software
Tools for Technology Transfer, 7(3):212–232, 2005.
9. W.-N. Chin, C. David, H.-H. Nguyen, and S. Qin. Enhancing modular OO verification with
separation logic. In POPL’08 [1], pages 87–99.
10. C. Clifton and G. T. Leavens. Obliviousness, modular reasoning, and the behavioral subtyping analogy. In SPLAT 2003: Software engineering Properties of Languages for Aspect
Technologies at AOSD 2003, Mar. 2003. Available as Computer Science Technical Report
TR03-01a from ftp//:ftp.cs.iastate.edu/pub/techreports/TR03-01/TR.pdf.
11. O.-J. Dahl. Verifiable Programming. International Series in Computer Science. Prentice
Hall, New York, N.Y., 1992.
12. J. Dovland, E. B. Johnsen, and O. Owe. Observable Behavior of Dynamic Systems: Component Reasoning for Concurrent Objects. Electronic Notes in Theoretical Computer Science,
203(3):19–34, 2008.
13. J. Dovland, E. B. Johnsen, O. Owe, and M. Steffen. Lazy behavioral subtyping. In J. Cuellar
and T. Maibaum, editors, Proc. 15th International Symposium on Formal Methods (FM’08),
volume 5014 of LNCS, pages 52–67. Springer, May 2008.
14. J. Dovland, E. B. Johnsen, O. Owe, and M. Steffen. Incremental reasoning for multiple
inheritance. In M. Leuschel and H. Wehrheim, editors, Proc. 7th International Conference
on Integrated Formal Methods (iFM’09), volume 5423 of LNCS, pages 215–230. Springer,
Feb. 2009.
15. E. Gamma, R. Helm, R. Johnson, and J. Vlissides. Design Patterns: Elements of Reusable
Object-Oriented Software. Addison-Wesley, Reading, Mass., 1995.
16. N. Hameurlain. Behavioural subtyping and property preservation for active objects. In
B. Jacobs and A. Rensink, editors, Proceedings of the Fifth International Conference on
Formal Methods for Open Object-Based Distributed Systems (FMOODS 2002), volume 209,
pages 95–110. Kluwer, 2002.
17. C. A. R. Hoare. Communicating Sequential Processes. International Series in Computer
Science. Prentice Hall, 1985.
18. A. Igarashi, B. C. Pierce, and P. Wadler. Featherweight Java: a minimal core calculus for
Java and GJ. ACM Transactions on Programming Languages and Systems, 23(3):396–450,
2001.
19. G. T. Leavens and D. A. Naumann. Behavioral subtyping, specification inheritance, and
modular reasoning. Technical Report 06-20a, Department of Computer Science, Iowa State
University, Ames, Iowa, 2006.
20. B. H. Liskov and J. M. Wing. A behavioral notion of subtyping. ACM Transactions on
Programming Languages and Systems, 16(6):1811–1841, Nov. 1994.
87
21. C. Luo and S. Qin. Separation logic for multiple inheritance. ENTCS, 212:27–40, 2008.
22. A. Mikhajlova and E. Sekerinski. Class refinement and interface refinement in objectoriented programs. In J. Fitzgerald, C. B. Jones, and P. Lucas, editors, FME’97. Industrial
Benefits of Formal Methods, volume 1313 of LNCS, pages 82–101. Springer, 1997.
23. S. Owicki and D. Gries. An axiomatic proof technique for parallel programs I. Acta Informatica, 6(4):319–340, 1976.
24. M. J. Parkinson and G. M. Biermann. Separation logic, abstraction, and inheritance. In
POPL’08 [1].
25. C. Pierik and F. S. de Boer. A proof outline logic for object-oriented programming. Theoretical Computer Science, 343(3):413–442, 2005.
26. Z. Qian and B. Krieg-Brückner. Typed object-oriented functional programming with late
binding. In P. Cointe, editor, Proc. 10th European Conference on Object-Oriented Programming (ECOOP’96), volume 1098 of LNCS, pages 48–72. Springer, 1996.
27. N. Soundarajan and S. Fridella. Inheritance: From code reuse to reasoning reuse. In
P. Devanbu and J. Poulin, editors, Proc. Fifth International Conference on Software Reuse
(ICSR5), pages 206–215. IEEE Computer Society Press, 1998.
28. M. Utting. An Object-oriented Refinement Calculus with Modular Reasoning. PhD thesis,
University of New South Wales, Australia, 1992.
29. H. Wehrheim. Behavioral subtyping relations for active objects. Form. Methods Syst. Des.,
23(2):143–170, 2003.
Einar Broch Johnsen had the pleasure of spending a year as a Post.
Doc. in Bernt Krieg-Brückner’s group at the University of Bremen.
This year exposed him to many new and for him slightly surprising
research activities, reflecting the diversity of the group; for example, wheel-chair avoidance in the corridor, the coffee machines of
algebraic specification (i.e., natural numbers), and the laws of inconsistency and their illustruous proponent Erwin. It was indeed a
memorable year!
A Paradox in Estimating Uncertainty
Udo Frese1,2 and Oliver Birbach2
1
Fachbereich 3 - Mathematik und Informatik, Universität Bremen,
Postfach 330 440, 28334 Bremen, Germany
E-Mail: [email protected]
2
Deutsches Forschungszentrum für Künstliche Intelligenz GmbH,
Sichere Kognitive Systeme, Enrique-Schmidt-Str. 5, 28359 Bremen, Germany
E-Mail: [email protected]
Abstract. This paper addresses the theoretical problems behind estimating the
measurement noise standard deviation σi of different measurement groups along
with the state in Gaussian least square estimation. We derive a solution in a Maximum Likelihood (ML) framework and present an iterative algorithm for computing the ML estimate. However, it turns out, that often ML gives meaningless,
paradoxical results. We analyze this phenomenon for the simple example of two
measurements. Paradoxically, the ML estimator blames one of the measurements
for all the error, pretending the other one was exact.
We then proof, that this happens, if the measurements associated with a single
σi are not over-constrained alone. This is the case for most interesting situations,
showing, that ML estimation is mostly useless here. Finally we derive a formula
for the residual covariance that allows to pin down the problems described above
and provides a starting point for a better uncertainty estimator.
1 Introduction
Least-square estimation is the task to obtain an estimate for an unknown state x based on
indirect measurements of that state corrupted by Gaussian noise z = f (x) + N (0, P ).
It has been invented by C.F. Gauss [1] for astronomy and surveying, and has found wide
spread application since then.
In robotics, (Nonlinear) Least Square (NLS) estimation is very widely used for calibration. Examples are: a camera, an inertial sensor, a laser-scanner, the kinematics
of a robot arm. Moreover, the whole areas of Simultaneous Localization and Mapping
(SLAM) [2], Multiple View Geometry [3] in computer vision, and tracking [4] are based
largely on least-square estimation.
One shortcoming of NLS is that the user has to specify the uncertainty of measurements by a covariance matrix. So in this paper, we address the problem of estimating
uncertainty parameters of the measurements along with the state.
1.1 Example Applications
Here are two applications that motivated our research and are used throughout the paper
for providing intuition.
89
2D feature based SLAM [2, §2] is concerned with building a map with a mobile
robot. A feature sensor provides the location of point features relative to the robot and
odometry provides the pose of the robot after moving a step relative to the pose before
moving. The goal would be to estimate two uncertainty parameters for the feature sensor
and for odometry.
In camera-inertial calibration internal camera parameters and the orientation of the
camera relative to the inertial sensor are to be estimated. The input are the image points
of a horizontal checkerboard pattern and gravity measured by the accelerometers in the
inertial sensor. Again, one wants to estimate the uncertainty in the image points and the
accelerometer.
How could this be achieved?
2
ML Estimation of Mean and Variance of a 1-D Gaussian
In fact, there is the well known textbook result on ML estimation of mean and variance [4, §2.5.3]. It assures that the mean and the variance of a 1-D Gaussian can be
estimated from samples z1 , . . . , zn by maximizing the likelihood
p(z• |μ, σ) =
n
p(zi |μ, σ) ∝
i=1
n 1
i=1
σ
exp
−(zi − μ)2
2σ 2
(1)
with respect to both μ and σ.
2.1 Optimal μ
There is a closed solution which we will derive in the following to illustrate the difference to classical least square estimation. The first step is to take the negative logarithm
χ2 (μ,σ)
− ln p(z• |μ, σ) = n ln σ +
= n ln σ +
n
1 (zi − μ)2
2
i=1
σ2
+ const
n
1 (zi − μ)2 + const
2σ 2 i=1
(2)
(3)
which is then to be minimized. In classical least-square estimation σ is fixed and only
the squared residual χ2 (μ; σ) remains. The optimal μ is the average of the samples
n
μmin =
being independent from σ.
1
zi .
n i=1
(4)
90
8
ln σ +
1
2σ 2
7
6
5
4
3
2
1
0
0
5
10
σ
15
20
Fig. 1. The function a ln σ + 2σb 2 plays an important role in estimating the standard deviation σ.
p
It converges to ∞ both for σ → 0 and σ → ∞ and has a minimum at σ = b/a. The plot
shows this function for a = b = 1.
2.2 Optimal σ
From the perspective of σ, (3) has the form a ln σ + 2σb 2 shown in Fig. 1. At this point
we can already catch a glimpse what happens when minimizing with respect to σ. With
σ → ∞ the χ2 error is scaled towards 0 but ln σ → ∞ . For σ → 0, ln σ → −∞, but
the χ2 error goes to +∞. The latter is stronger so the result will go to +∞. In between,
there is the minimum we are looking for. A notable exception is the situation where all
residuals are 0, hence the χ2 error is 0 and the minimum is −∞ at σ = 0. This is still
reasonable, since all samples are identical. This phenomenon will be important later,
where we will encounter paradoxical results.
The behavior discussed above is established by the following lemma (proof in the
appendix).
Lemma 1. If a, b > 0, then σ = b/a. is the minimum of
a ln σ +
b
,
2σ 2
(5)
which converges to +∞ both for σ → 0 and σ → ∞.
If b = 0, (5) converges to −∞ for σ → 0.
Combining this result with (3) we obtain
n
1 σ=
(zi − μ)2 .
n i=1
(6)
This is a well known formula for computing the standard deviation [5]. Many readers
may remember it mainly for the discussion, whether n or n − 1 should be used in the
denominator. Indeed, the ML estimate (6) would be okay, if μ was fixed. However, μ
is estimated by (4) as the average sample and hence is closer to the samples than the
actual mean of the distribution. This leads to a smaller χ2 error and an underestimated
91
σ. It can be shown [4, §2.6.3] that the corresponding sample standard deviation formula
with n − 1 in the denominator is unbiased and thus more desirable.
As a final remark, Bar-Shalom et al. [4] pointed out that many samples are needed
for a good σ estimate, for instance ≈ 800 samples for ≤ 10% error with 95% probability.
3
An ML View on Combined State and Covariance Estimation
3.1 Definition
Motivated by the successful estimator above, we will now derive a ML formulation
for NLS with estimation of covariance parameters. As for classical NLS, our goal is to
estimate a state x from vectors of stacked measurements z1 , . . . , zn . But unlike NLS
we also estimate a scalar uncertainty parameter σi for every such vector zi . Each vector
zi is build by stacking all measurements that have an uncertainty governed by the same
factor σi . So intuitively this corresponds to measurement groups rather than individual
measurements, e.g. all measurements from the same sensor.
Definition 1 (Probabilistic Model). The measurement group vector zi is assumed to
consist of a “true” measurement being a function fi (x) of the state vector x plus unknown noise ei . The noise ei is independently distributed as a Gaussian with 0 mean
and covariance matrix σi2 Pi .
zi = fi (x) + ei ,
ei ∼ N (0, σi2 Pi ),
i = 1, . . . , n
(7)
The measurement vectors zi , the functions fi and the covariance matrices Pi are known.
The state measurement errors ei , the state x, and the factors σi scaling the covariances
are unknown, the latter two shall be estimated.
3.2 Intuition
To understand the intuitive meaning of this definition, consider the 2D SLAM example
from the introduction. The state vector x would consist of all feature positions and
all robot poses stacked in a single vector. As Bar-Shalom pointed out, a lot of data
is needed to estimate an uncertainty parameter. So we can not hope to estimate an
individual uncertainty parameter for every such measurement.
As a first idea, one could assume that the noise in every component of all odometry
measurements has the same standard deviation σ1 and the noise in every component of
all feature measurements has the same standard deviation σ2 . The goal would then be to
estimate σ1 and σ2 along with the state vector x. According to Def. 1. all odometry measurements would be stacked into z1 and all feature measurements into z2 respectively.
So there are two measurement groups and P1 , P2 would be identity matrices.
92
3.3 Sensors with Varying Uncertainty
The assumption of having the same noise in all measurements of one sensor may be
too restrictive, because the noise may depend on a lot of factors. The noise of a feature
measurement depends for instance on distance, the feature itself or some signal to noise
ratio. The noise of an odometry measurement depends on the distance traveled and even
has correlations originating from the kinematic of wheeled motion. All these effects can
be incorporated into Pi , if they are known. The matrix Pi is derived as for usual NLS
assuming the sensor has some arbitrary reference uncertainty and the covariance of the
measurements group zi is then scaled by σi which is estimated by the algorithm.
There is a interesting class of sensors where raw sensor data is internally preprocessed. These sensors have a covariance which is known up to an unknown scale, so
are perfectly compatible with Def. 1. One example is odometry, where the rotation of
the wheels is integrated into a geometrical pose. Another example would be metrical
information from computer vision. In those cases one can assume σi to be the unknown
noise of the raw measurements r in each component, which then propagates through
the sensor preprocessing function zi = g(r) as
cov(zi ) = cov(g(r)) = J cov(r)J T
= J(σi I)J T = σi JJ T , with J =
(8)
∂g(r)
.
∂r
(9)
These kind of sensors can be readily incorporated into Def. 1 by setting Pi = JJ T .
4 A Practical ML Algorithm
4.1 The Likelihood function
We are now deriving an ML algorithm for the system in Def. 1. The likelihood function
p(z• |x, σ• ) =
=
n
i=1
n
p(zi |x, σi )
(10)
p(ei = zi − fi (x)|σi )
(11)
i=1
n
1 2
exp − χi (x, σ) ,
2
|σi2 Pi |
(12)
with χ2i (x, σi ) = (zi − fi (x))T (σi2 Pi )−1 (zi − fi (x))
(13)
∝
i=1
1
to be maximized is derived by using independence of the different zi , solving (7) for ei ,
and substituting the probability density function of a Gaussian [4].
Again, we take the negative logarithm
n 1
(14)
const + dim(zi ) ln σi + χ2i (x, σi ) ,
− ln p(z• |x, σ• ) =
2
i=1
93
C OMPUTE MLE STIMATE (f• , z• , P• , σ•initial )
σ• = σ•initial
repeat
compute x by NLS algorithm with fixed σ•
compute σi from (15) for all i
until convergence
return x, σ•
Fig. 2. Algorithm to compute the ML estimate of x and σ• for an input corresponding to Def. 1.
The subscript • means “for all i”.
2 dim(z )
i
using that |σi2 Pi | = σi
|Pi |. The result consists of the usual χ2 -error χ2 (x, σ) =
n
−2
2
χ
(x,
σ
)
where
the
σ
i
i
i scale the squared residuals and of a second term, namely
i=1
n
dim(z
)
ln
σ
that
favors
smaller σi . When σ• is fixed, the result is the minimum
i
i
i=1
of χ2 (x) as computed by NLS. All this is similar to the 1-D Gaussian case. The main
difference is, that there are now several σi , so the optimal x depends on σ• .
4.2 The Minimization Algorithm
One could simply use a general function minimization algorithm, such as downhillsimplex [6, §10.4] or Powell’s method [6, §10.5] to minimize (14). However, we propose
to use an NLS algorithm with fixed σi as a building block. We think of the problem
rather as an extension to NLS and try to design the algorithm as an extension too. The
advantage is that any problem specific NLS algorithm, for instance a SLAM algorithm,
could be used inside the overall ML algorithm then.
The proposed ML algorithm (Fig. 2) is iterative and related to expectation maximization (EM) [7]. It alternatingly minimizes over x with σ• fixed and over σ• with x
fixed. The essential observation for the second step is, that for a fixed x (14) decomposes into independent minimization problems for each σi . Each of them can be solved
by applying lemma 1 and using χ2i (x, σi ) = σi−2 χ2i (x, 1).
χ2i (x, 1)
χ2i (x, σiold )
= σiold
(15)
σi =
dim(zi )
dim(zi )
So, to compute an ML estimate for σi given x, one simply takes the residual of the last
NLS step (that used σiold ), computes the χ2 error, and applies (15).
4.3 Convergence
Two final remarks on the algorithm: First, each of the two steps in Fig. 2 reduces (14), so
the algorithm converges if (14) has a lower bound. When it converges, the gradient with
respect to x is 0 due to the convergence of the first step and the gradient with respect
to σ• is 0 due to the convergence of the second step. Hence, the overall gradient is 0
and the algorithm found a local minimum at least. Second, when the NLS algorithm is
94
iterative itself, for example Levenberg-Marquardt [6, §15.5], it is attractive to integrate
the σ• estimation step into the existing loop, because it does not make sense to iterate
x to convergence if σ• still is changing.
To summarize: A range of practically relevant problems involving estimation of state
and covariance parameters can be cast as an ML problem and there is a comparatively
simple algorithm for computing it. This sounds promising, but as we will analyze in the
following, in many practical cases the ML estimate is meaningless or even paradoxical.
5 An Example for a Paradoxical ML Estimate
When we had developed the algorithm above, we experienced that one of the σi always
converged to zero [8]. In words, the ML estimate blamed some of the sensors for all the
noise while pretending that the other sensor was perfect. Here is a simple example with
an analytical solution to obtain a deeper inside into this behavior.
5.1 Problem Definition
Imagine the same 1-D quantity measured by two sensors with different uncertainty as 0
and 1.
x ∈ R, f1 (x) = f2 (x) = x,
(16)
z1 = 0, z2 = 1, P1 = P2 = (1).
(17)
What estimate would you expect intuitively? It is hard to say, since two measurements
seem not enough for estimating one state and two uncertainty parameters. At least we
can say something about the sum of both uncertainties. We have one sample z2 −z1 = 1
from a Gaussian with 0 mean and variance σ12 + σ22 . So we can expect at least σ12 + σ22 to
have a well defined estimate. Apart from this, one could reasonably expect the minimum
to be in the middle (σ1 = σ2 = 21 , x = 12 ) for reasons of symmetry. Or, one could
equally reasonably expect the minimum to be indifferent with the same likelihood for
all σ1 , σ2 pairs that have the same σ12 + σ22 value.
That is what intuition says. But what is the true minimum?
5.2 Derivation of the Minimum
We plug (17) into (14), getting
ln σ1 +
(0 − x)2
(1 − x)2
+ ln σ2 +
2
2σ1
2σ22
(18)
We know from linear least squares that the optimal x is the average inversely weighted
by covariance [4]. So we substitute
x=
σ22 0 + σ12 1
σ12
=
σ12 + σ22
σ12 + σ22
(19)
95
Fig. 3. Likelihood of the example in Sec.5, (21) plotted over σ1 and σ2 . There is a saddle point at
σ1 = σ2 = 21 but the minimum is −∞ at either σ1 = 0 or σ2 = 0. The lines show the likelihood
on σ1 = σ2 (black), where the same shape as in Fig. 1 can be observed and on σ12 + σ22 = 21
(yellow) respectively.
96
into (18) getting
σ14
σ24
+
2(σ12 + σ22 )2 σ12
2(σ12 + σ22 )2 σ22
1
= ln σ1 + ln σ2 +
2(σ12 + σ22 )
= ln σ1 + ln σ2 +
(20)
(21)
At this point we can already see, that something strange is happening. If we stay on a
circle σ12 + σ22 = const the third term is constant and the overall function can be made
−∞ by either setting σ1 = 0 or σ2 = 0 (Fig. 3). Formally, we set the derivatives to 0.
1
σ1
− 2
σ1
(σ1 + σ22 )2
σ2
1
− 2
and 0 =
σ2
(σ1 + σ22 )2
0=
⇒
σ12 = σ22 = (σ12 + σ22 )2
1
⇒ σ1 = σ2 =
2
(22)
(23)
(24)
(25)
Indeed, we have a critical point at σ1 = σ2 = 12 , but it is a saddle point, not a minimum.
As Fig. 3 shows, it is a minimum in the direction of σ1 = σ2 , but a maximum in the
orthogonal direction of σ12 + σ22 = 12 . This shows, if we assume both measurements
to have the same standard deviation, the ML estimate of σ = 12 would be reasonable.
However, if we allow both measurements to have different standard deviations, ML has
a paradoxical estimate, claiming that one of the sensors is perfect and the other source
of all the error.
6 A Condition for Meaningful ML Covariance Estimation
One could criticize the above example for evaluating the ML estimate in a situation,
where even intuitively it is unclear what the result should be. It would therefor be interesting to know conditions when the ML estimate becomes paradoxical with one σi = 0
and when it is regular with all σi > 0. We can characterize this situation in the following, i.e. we give a necessary and a sufficient condition for one of the σi being 0.
6.1 Sufficient Condition
When one of the measurement groups alone can be realized without residual, the estimate will become paradoxical.
Theorem 1. If for some i, zi = fi (x0 ) for some x0 , a minimum of the likelihood function (14) is −∞ at x = x0 , σi = 0, σj > 0 for i = j.
Proof. Define a sequence of standard deviation vectors σ•k , with σik = k1 and σjk = 1
for all j = i. Then we claim, that (14) on x = x0 , σ• = σ•k converges to −∞:
97
We go through the terms in the sum in (14): χ2j (x, σjk ) as well as ln σjk is constant
for j = i. χ2 (x, σik ) is 0 by (13), since zi = fi (x0 ). Finally, ln σik = ln k1 = − ln k
converges to −∞ for k → ∞. So the whole negative log-likelihood (14), being the sum
of these terms converges to −∞ as σi approaches 0 with x = x0 and the other σj fixed.
6.2 Necessary Condition
We formulate the necessary condition in reversed form: If for all measurement groups
the residual error is lower bounded, the estimate will not be paradoxical.
Theorem 2. If for all measurement groups the residual error is lower bounded, i.e. ∃r ∈
k
converges
R : ∀x0 , i : |zi − fi (x0 )| ≥ r, then for any sequence (xk , σ•k ) where one σm
to 0 the whole negative log-likelihood (14) converges to +∞.
Proof. We first show that χ2i (x, 1) is bounded from below. As a symmetric positive
definite matrix, Pi can be decomposed as Pi = U T DU with U orthonormal and D
diagonal, consisting of the positive eigenvalues of Pi .
χ2i (x, 1) = (zi − fi (x))T (Pi )−1 (zi − fi (x))
T
T
(26)
−1
= (zi − fi (x)) U D U (zi − fi (x))
2
dim(zi )
U (zi − fi (x)) j
=
Djj
j=1
(27)
(28)
Replacing all Djj by the maximal eigenvalue Dmax of all Pi yields a lower bound
2
dim(zi )
U (zi − fi (x)) j
≥
(29)
Dmax
j=1
|U (zi − fi (x))|2
Dmax
|zi − fi (x)|2
=
Dmax
r2
≥
Dmax
=
(30)
(31)
(32)
We plug this result into a single term of (14).
1
dim(zi ) ln σi + χ2i (x, σi )
2
1
= dim(zi ) ln σi + 2 χ2i (x, 1)
2σi
1 r2
≥ dim(zi ) ln σi + 2
2σi Dmax
(33)
(34)
We apply lemma 1 to (14), setting a = dim(zi ), and b = Drmax . It follows that all
terms in the sum in (14) are bounded from below for all x and σ. For the particular
2
98
term i = m the lemma even implies convergence to +∞, because σm → 0. So overall,
(14) converges to +∞, being the sum of lower bounded terms and a term converging to
+∞.
6.3 The Cases between Necessary and Sufficient Condition
An ideal characterization would be an equivalent, i.e. a necessary and sufficient condition. So it is interesting to know, which cases fulfill the necessary but not the sufficient
condition, so we cannot tell whether they we will have a paradoxical σ = 0 estimate or
not.
In such a case, there would be no lower bound on the residual error |zi − fi (x0 )|
but still there would also be no x0 where residual error |zi − fi (x0 )| would actually
become 0. This cannot happen for a continuous measurements function in a compact
domain, where |zi − fi (x0 )| would also be continuous and hence achieve its minimum.
This can only happen, when there “is” a perfect, i.e. residual-free estimate from zi but it
lies at infinity or in a singularity. An example would be triangulation when the observed
bearings are parallel. Then we would argue, even classical NLS will fail.
So, for take-home message, we can focus on the necessary condition. For the covariance estimation to be non-paradoxical, each measurement group alone must not be
realizable by a state without leading to a residual.
6.4 Discussion
This result is very disappointing. It says, that from the perspective of whether it is
possible to estimate measurement uncertainty, sensor fusion does not make a difference.
A measurement can be expected to have a non-zero residual for all x0 , if it alone
already over-constrains the state. So, the characterization above means that for ML estimation of uncertainty to be possible, each group of measurements alone must already
be over-constrained. But if this holds, we could also estimate the measurement uncertainty much more easily. We would simply perform NLS on each measurement group
alone and apply the estimator (41) for a single σ.
This is disappointing, because by contrast the power of NLS is, that the result of
fusing several measurement groups is often dramatically better than a state obtained
from one of the measurement groups alone. Frequently, each group even does not contain enough information to estimate a state at all and only by fusing all measurement
groups a state estimate becomes available.
It is even more disappointing for robotics, because an important type of robotic sensors generally produces no over-constrained measurements. This holds for all sensors
that can be integrated over time, i.e. that can be used in the Kalman-filter’s [4] dynamic
equation. For every initial state there is one sequence of successive states in time that
perfectly corresponds to the measurements and is obtained by integrating the measurements over time. Examples for these sensors are odometry, inertial sensors, or dynamic
probabilistic models for target tracking that are mathematically treated as a sensor.
In the 2D SLAM example from the introduction, estimating the uncertainty of the
feature sensor and odometry is not possible, since measurements from a feature sensor
are usually over-constrained, but from odometry not.
99
7
An Analysis of the Inherent Difficulty
7.1 What should we expect?
Given that the ML approach to estimating measurement covariance did not succeed, an
interesting question is, whether this is due to the ML paradigm or whether the problem
itself is ill-posed. From our view, it is at least not always ill-posed. Consider again
the 2D-SLAM example. The measurement group of feature measurements is overconstrained. So one can estimate the uncertainty of feature measurements using the
single σ estimator (41). Now there are two extreme cases.
First, assume, odometry has very high uncertainty. Then, one could derive the relation between successive robot poses, i.e. what odometry measures, from the feature
result and use it more or less as ground truth to estimate the odometry uncertainty. In
that case is should indeed be possible to estimate both measurement uncertainties.
Now, assume, odometry has very low uncertainty. Then the discrepancy between
odometry and feature measurements could be completely explained by the uncertainty
in the feature measurements. Odometry alone is not over-constrained, has no inner discrepancy. So, there is no way, we could tell from this data, whether we have a very
precise, an extremely precise, or even exact odometry.
From this example we can conclude two things: First, there are cases, where ML
fails but estimation of measurement uncertainties should be possible. Second, if one of
the sensors is not over-constraint and much preciser than all others, then it is not possible rule to out that the sensor is exact. So it is not possible to estimate the measurement
uncertainty with some relative accuracy. All one can provide is an upper bound.
7.2 The Distribution of the Residual
We want to conclude the paper with a more fundamental view on estimation of measurement uncertainties.
The estimated state does not carry information on the uncertainties, because any
x could be estimated if measurements were exactly zi = fi (x). We have to estimate
the uncertainties from the residuals. So, in the following we will derive the covariance
matrix of the residuals assuming the measurements follow Def. 1 for some σ• . This
allows to pin down the core difficulty of the problem and is a starting point for the
future work of designing a better estimator.
From the perspective of the NLS algorithm the measurement function is f• , i.e. the
measurement functions from Def. 1 stacked. Let J be the corresponding Jacobian and
C = diag(σi2 Pi ) the measurement covariance used by the NLS algorithm. Then according to Frese [2, (17), (18)],
X = (J T C −1 J)−1 J T C −1
(35)
is the matrix version of least square estimation, mapping from measurement error e• to
state error x̃. So the overall matrix Y mapping from the measurement error e• to the
residual zi − f (x) is given by
Y = I − JX, with
(36)
100
The matrix Y is a projection matrix, i.e. Y Y = Y .
−1 T −1 T −1 −1 T −1
J C J J C J
J C
(JX)2 = J J T C −1 J
T −1 −1 T −1
=J J C J
J C = JX
Y = (I − JX)(I − JX) = I − 2JX + (JX) = I − JX = Y
2
2
(37)
(38)
(39)
It removes those degrees of freedom of the measurement noise that get absorbed into
the state estimate and passes the remaining degrees of freedom unmodified. If the measurement covariance is cov(e• ) = C, then the residual covariance is
cov (z• − f• (x)) = Y CY T
(40)
7.3 In a Nutshell
We are now in the position to put everything in a nutshell.
Our problem? We have obtained a sample of the residual distribution N (0; Y CY T )
and want to estimate the parameters σ• scaling the different blocks of the block-diagonal
matrix C.
ML’s fault? It treats Y as identity I.
The inherent difficulty? Y projects away some degrees of freedom which are diagonal to the subspaces corresponding to the different σi . Hence, the effects of the different
σi become coupled in a difficult way.
8 Related Work
NLS is covered in textbooks on statistical estimation [4] and numerical computation [6,
§15.5] for instance with the Levenberg-Marquardt algorithm. The first book also contains the ML estimator presented in Sec.2.
The second book presents a simple unbiased estimator for a single parameter σ
scaling the covariance C used in NLS. It is based on the observation that the χ2 error of
NLS approximately follows a χ2 distribution (hence the name) with dim(z) − dim(x)
degrees of freedom and expected value dim(z) − dim(x). This allows to estimate a σ
as
1
σ = σ old dim(z)−dim(x)
χ2 (x, σold )
(41)
This estimator is unbiased, because unlike (15) it takes the degrees of freedom into
account that are projected away by Y into the estimate. However, it works only for a
single σ, because for several σ it is unclear how these degrees of freedom are “divided”
among the different measurement groups.
The Analysis of Variance (ANOVA) [9] method introduced by Fisher is also related.
It addresses the problem of samples drawn from different classes. The method estimates
two parameters, the variance inside every class and the variance of the mean among the
classes. This has applications in psychology where the classes could for instance be
different cities.
101
9
Conclusion
In this paper we have analyzed a ML approach to estimating measurement uncertainty
along with the state as an extension to NLS. The main conclusion is, that the ML estimate becomes paradoxical, if a group of measurements governed by the same uncertainty factor σi does not over-constrain the state alone. Unfortunately this holds for
most interesting scenarios, so the ML estimate is mostly useless here.
One can also draw more general consequences from this result. Probabilistic inference is used in many branches of robotics, computer vision, and artificial intelligence.
Often, maximum-likelihood is viewed as a gold-standard there and the discussion focusses on how to compute it efficiently. This result suggests to also question the ML
solution itself, because it can sometimes be surprisingly ill-behaved.
Finally, we have analyzed the reason for this problem and provided a characterization of the residual covariance that could be a starting point for a better estimator.
References
1. Gauss, C.: Theoria combinationis observationum erroribus minimis obnoxiae. Commentationes societatis regiae scientiarum Gottingensis recentiores 5 (1821) 6–93
2. Frese, U.: A discussion of simultaneous localization and mapping. Autonomous Robots 20(1)
(2006) 25–42
3. Hartley, R., Zisserman, A.: Multiple View Geometry in Computer Vision. Cambridge University Press (2004)
4. Bar-Shalom, Y., Rong-Li, X., Kirubarajan, T.: Estimation with Applications to Tracking and
Navigation. John Wiley & Sons, Inc. (2001)
5. : Wikipedia: Standard deviation http://en.wikipedia.org/wiki/Standard_
deviation, accessed 14. January 2008.
6. Press, W., Teukolsky, S., Vetterling, W., Flannery, B.: Numerical Recipes, Second Edition.
Cambridge University Press, Cambridge (1992)
7. Dempster, A., Laird, A., Rubin, D.: Maximum likelihood from incomplete data via the EM
algorithm. Journal of the Royal Statistical Society 39 (1977) 1 – 38
8. Birbach, O.: Accuracy analysis of camera-inertial sensor based ball-trajectory prediction.
Master’s thesis, Universität Bremen, Fachbereich 3: Mathematik und Informatik (2008)
9. King, B., Edward, W.: Statistical Reasoning in Psychology and Education, Fourth Edition.
John Wiley & Sons, Inc. (2003)
102
Udo Frese received his Ph.D. from University of ErlangenNürnberg in 2004. From 2004 until 2008 he has been a postdoc researcher in Bernd Krieg-Brückners lab at University of
Bremen and in the German Research Center for Artificial Intelligence (DFKI). He is now assistant professor for real-time
computer vision at University of Bremen and further on affiliated with the DFKI. His research topics are real time computer vision, in particular simultaneous localization and mapping (SLAM) and computer vision for sport robotics.
Oliver Birbach received the diploma degree in computer science from University of Bremen in 2008. He is currently a junior researcher at the German Research Center for Artificial
Intelligence in Bremen. His main research areas are multitarget tracking, context-based computer vision and computer
vision in sports environments.
Appendix
Proof (Proof of Lemma 1). We start with the asymptotics. For σ → ∞, a ln σ → ∞
and 2σb 2 → 0. For σ → 0, it is more difficult, as a ln σ → −∞ and 2σb 2 → +∞. So we
apply L’Hospital’s rule to the fraction.
a ln σ
b
2σ2
L’Hospital
−−−−−→
b
b
a ln σ + 2 = 2
2σ
2σ
a
σ
− σb3
a ln σ
b
2σ2
aσ 2 σ→0
−−−→ 0
b
=−
+1
σ→0
−−−→ ∞1 = ∞
(42)
(43)
In the special case of b = 0, the function is simply ∝ ln σ which converges to −∞ for
σ → 0.
For a minimum, we set the derivative of (5) to 0:
0=
b
a
− 3 ⇒ σ = b/a.
σ σ
(44)
This is actually a minimum, because it is located in the interval [0, ∞) where the function goes to +∞ on both sides, so there must be at least one minimum in between.
(A) Vision for 2050
Context-Based Image Understanding for a
Human-Robot Soccer Match
Udo Frese1,2 , Tim Laue2 , Oliver Birbach2 ,
Jörg Kurlbaum1 , and Thomas Röfer2
1
Fachbereich 3 - Mathematik und Informatik, Universität Bremen,
Postfach 330 440, 28334 Bremen, Germany
E-Mail: {ufrese,jkur}@informatik.uni-bremen.de
2
Deutsches Forschungszentrum für Künstliche Intelligenz GmbH,
Sichere Kognitive Systeme, Enrique-Schmidt-Str. 5, 28359 Bremen, Germany
E-Mail: {Tim.Laue,Oliver.Birbach,Thomas.Roefer}@dfki.de
Abstract. We believe it is possible to create the visual subsystem
needed for the RoboCup 2050 challenge – a soccer match between humans and robots – within the next decade. In this position paper, we
argue, that the basic techniques are available, but the main challenge
will be to achieve the necessary robustness. We propose to address this
challenge through the use of probabilistically modeled context, so for
instance a visually indistinct circle is accepted as the ball, if it fits well
with the ball’s motion model and vice versa.
Our vision is accompanied by a sequence of (partially already conducted)
experiments for its verification. In these experiments, a human soccer
player carries a helmet with a camera and an inertial sensor and the
vision system has to extract all information from that data, a humanoid
robot would need to take the human’s place.
1
Introduction
Soon after establishing the RoboCup competition in 1997, the RoboCup Federation proclaimed an ambitious long term goal (depicted in Fig. 1).
“By mid-21st century, a team of fully autonomous humanoid robot soccer
players shall win the soccer game, comply with the official rule of the
FIFA, against the winner of the most recent World Cup.”
Hiroaki Kitano, Minoru Asada [18]
Currently, RoboCup competitions take place every year. Within a defined set of
different sub-competitions and leagues, incremental steps towards this big goal
are made [28]. Although a rapid and remarkable progress has been observed
1
This paper is an extended version of [11] presented at the celebration of the 60th
birthday of Bernd Krieg-Brückner.
104
Fig. 1. Illustration of the RoboCup vision: An autonomous humanoid robot playing
soccer.
during the first decade of these robot competitions, it is not obvious, if and how
the final goal will be reached. There exist rough roadmaps, e.g. by [7], but in
many research areas, huge gaps must be bridged within the next 40 years.
While this is obvious for several areas, e.g. actuator design and control, we
claim that the situation is surprisingly positive for vision:
Within the next decade, it will be possible to develop a vision system that
is able to provide all environmental information necessary to play soccer
on a human level.
Annual RoboCup competitions are always bound to strict rule sets (defined for
the state of the art of the competing robots) and demand competitive robot
teams. Thus only incremental progress adapting to actual rule changes (which
continuously raise the level of complexity) is fostered. By developing the aforementioned vision system independently of these competitions, we hope to set a
new landmark which could guide the incremental development.
Because a real human level soccer robot will not be available for a long time,
our vision is accompanied by a (partially already conducted) set of experiments
that verify our claim without needing a robot.
This paper is organized as follows: Section 2 roughly identifies the challenges
for playing robot soccer and compares them to the state of the art in robotics. In
Sect. 3 we explain, why the basic techniques for the vision system are available.
We argue, why the remaining challenge is robustness, for which we present our
idea of a solution in Sect. 4. Finally, a sequence of experiments to verify our
claim is described in Sect. 5.
105
Fig. 2. The Sense-Think-Act cycle roughly depicting major tasks for playing soccer
with a humanoid robot.
2
Challenges for Playing Soccer
The global task of playing soccer consists of several different, interdepending
challenges. We roughly categorize them according to the Sense-Think-Act cycle
(see Fig.2). This should be considered as a possible architecture for illustration. In the following, the challenges are described in reverted order but with
decreasing degree of difficulty.
2.1
Challenges for Actuation
The hugest obvious gap may be observed in the field of actuation. Nowadays, the
probably most advanced humanoid robot, Honda’s ASIMO, is capable of running
at a top speed of six kilometers per hour [16]. This is an impressive result, but
still more than five times slower than the top speed of a human soccer player.
A similar gap regarding kicking velocity has been pointed out by [13, 14]. They
showed that a state-of-the-art robot arm (with a configuration comparable to
a human leg) is six times slower than required to accelerate a standard soccer
ball to an adequate velocity. It is still an open issue whether today’s motor
technology could be developed further on enough, or if more efficient actuators,
e.g. artificial muscles, will be needed. Since soccer is a contact sport leading to
physical human-robot interaction [13], not only direct position control but also
approaches for compliant motion, such as impedance control, need to be taken
into account. As pointed out by [14], joint elasticity appears to be a crucial
aspect for velocity as well as for safety.
Additionally, the problems of energy efficiency and power supply need to be
solved. The ASIMO robot for example is, according to [16], capable of walking
(with a speed of less than three kilometers per hour) for 40 minutes.
2.2
Challenges for Thinking
In this area, two different levels may be distinguished: motion planning and highlevel multi-agent coordination. The latter is a research topic in the RoboCup
106
Soccer Simulation League since a while and has reached a remarkable level.
Dealing with the offside rule as well as playing one-two passes are standard
behaviors, complex group tasks as playing keep-away soccer serve as a test bed
for learning algorithms [30]. This area could be considered to be already quite
close to human capabilities.
On the other hand, when playing with real humanoid robots, sophisticated
methods for motion planning are needed. The current research frontier on humanoid motion control is balancing and dynamic foot placement for walking
robots. Algorithms for full-body motion planning exist [19], but are subject to
restrictions that make them inapplicable to tasks as playing soccer.
Here is a big gap to human level soccer. As an example consider volleykicking. The player has to hit the ball exactly at the right time, position, and
velocity, with a motion compatible to the step pattern, allowing balancing, and
considering opponents. Last but not least, all this must happen in real-time.
2.3
Challenges for Sensing
According to [18], it is evident that the robots’ sensorial capabilities should resemble the human ones. Thus, we could assume to deal with data from cameras
and inertial sensors emulating the human eyes and vestibular system. The required information are estimates of the own position and the positions of the
ball and of other players. In case of tackles or dribbling, the latter will be needed
to be recognized in more detail (e.g. the positions of the feet and limbs).
Current solutions for these tasks and our idea how to bridge the remaining
gap are presented in the following section.
3
The Vision System
Our main thesis is that the “sense” part of the RoboCup 2050 challenge can be
realized within a decade starting from the current state of the art in computer
vision. This is remarkable, since the “act” and “think” parts are apparently lightyears away from reaching human level performance and for computer vision in
general, this is also true. The reason, why we believe such a vision system can
be realized, is that unlike a household robot for instance, a soccer robot faces a
rather structured environment.
3.1
State of the Art
The objects relevant in a soccer match are the ball, the goals, the line markings,
and, of course, the players as well as the referees. Ball, goal, and line markings are
geometrical features, i. e. circles and lines. There is a large number of algorithms
for detecting them in images, from the classical Hough transform [8] up to a
range of more elaborate methods [12].
Recognizing other players is more challenging. It is particularly difficult because we will probably need not only the general position but the detailed state
107
Fig. 3. Tracking sportsmen and their limbs: Three persons which are running and
passing a ball (left image), one person performing jumping jacks (right image). Both
images are results of [27] and are used with permission of Deva Ramanan.
of motion for close range tackling and to infer the player’s action for tactical
purposes. Fortunately, people tracking is an important topic in computer vision
with a large body of literature [25]. An approach which is capable of tracking the
body as well as the limbs of sportsmen (see Fig. 3) has been presented by [27].
Furthermore, soccer scenes are lightly colored with green lawn and the players
wearing colored clothes of high contrast. In the RoboCup competition, this idea
is taken to an extreme, where most teams rely on color segmentation on a pixelper-pixel basis as their primary vision engine (see Fig. 4). This will not be
possible for real-world soccer, mainly due to changing lighting conditions. Still,
color can provide a valuable additional cue, at least when looking below the
horizon, where objects are in front of green lawn.
The background above the horizon, including the stadium and the audience
is of course also visible and unfortunately rather undefined. However, if it is
relevant for the soccer robot at all, then not for recognition, but only in the
sense of a general landmark. For this purpose there are nowadays well working
techniques, such as the Scale Invariant Feature Transform (SIFT) [24].
Overall, understanding a soccer scene from the player’s perspective seems
much easier than for instance understanding an arbitrary household, traffic, or
outdoor scene. Indeed there are already half-automatic systems in the related
area of TV soccer scene analysis, for example the ASPOGAMO system by [4,
3], proofing that soccer scene understanding in general is on the edge of being
functional.
3.2
Open Problems
So, is a vision system for the RoboCup 2050 challenge an easy task? We believe
it is not. It is surprisingly a realistic task but well beyond the current state of
the art. The first problem is that the camera is moving along with the head
of the humanoid soccer robot. To predict a flying ball, the orientation of the
108
Fig. 4. Two images taken and processed by a humanoid soccer robot on a standard
RoboCup field [22]. The colored dots and lines indicate perceived parts of goals and
field lines; the recognized ball is bordered by an orange circle.
camera must be known very precisely. It seems unrealistic that the necessary
precision can be obtained from the robot’s forward kinematic, since unlike an
industrial robot, a humanoid robot is not fixed to the ground. So our solution
is to integrate an inertial sensor with the camera and fuse the complementary
measurements of both sensors in a probabilistic least-square framework.
The second problem is the player’s perspective. It is much more difficult than
the overview perspective used in TV soccer scene analysis. In the TV perspective
the scale of an object in the image varies by a factor of about 3 [4, Fig. 5] whereas
in the player’s perspective it can vary by a factor of 250 assuming the distance to
an object ranging from 0.5m to 125m. Hence, for instance the people detection
algorithm must handle both extreme cases, a person only the size of a few pixels,
where an arm or a leg maybe thinner than a single pixel and a person much larger
than the camera’s field of view, only partially visible. Furthermore, in an image
from the player’s perspective, other players will extend beyond the green lawn
of the field into the general background. Hence it is not possible to search for
non-green blobs as an easy first processing step. This can also happen for a flying
ball, which is then particularly difficult to detect.
However, the third and most severe problem is, that from our experience,
most of the academic computer vision systems perform on the level of lab demonstrators requiring nicely setup scenes and lighting conditions and usually considerable parameter tweaking. So, to summarize, for the vision part of the RoboCup
2050 challenge, we do not need a new level of functionality as for many other
grand challenges, but we need a new level of robustness.
4
Robustness Through Context
We propose to address the question of robustness by utilizing probabilistically
modeled context information, formulating the overall scene understanding and
prediction problem as a global likelihood optimization task. Using contextual
109
associations is not entirely new [31, 5] and recent progress has shown that exploiting context leads to promisingly robust results. Recent work worth noting
includes modeling of object to object relations as done by [26]. Here, the authors
improved the object categorization accuracy by integrating a post-processing
step that maximizes the object categorization agreement based on contextual
relevance. More generally, [20] proposed a hierarchical framework allowing interaction between objects and regions within the image plane.
Leaving the image plane and therefore 2D models behind, [15] propose a
mechanism for estimating rough 3D scene geometry from a single image which
is then used as an additional input to the object detection methods. Therefore,
detections which fail to fit into the estimated scene geometry at the detected
position are rejected from the global scene interpretation (e. g. pedestrians who
do not stand on the ground plane). This idea is taken one step further by [9]
and [23] where the relationships between objects and scene geometry is modeled,
allowing suitable recognition of objects for a mobile robot application.
These promising results show that exploiting context within the image and
the scene is particularly well suited to the task of creating a vision system for
RoboCup 2050 and also that this task is well suited to study this methodology.
4.1
Data-Driven Bottom-Up Processing
Most current vision systems use a data-driven bottom-up approach [10, 29, 3,
as examples]. Usually, low level features are extracted from the image and then
aggregated through several stages to high level information. Each stage may
incorporate background knowledge at its own level but does not take information
from higher levels into account. It simply takes some input from the previous
lower level and passes the result of the computation to the next higher level.
As an example, a classical Hough transform starts by classifying pixels as
edge or not by thresholding the result, for instance, of a Sobel filter. Similarly,
the system by [3] starts by classifying pixels as lawn or not on the basis of
their color. This is a hard decision taken on the lowest level without any higher
level knowledge, such as the fact that we are looking for a ball or the ball’s
motion model. Such a pixel-wise classification can be very ambiguous. Often we
could, for instance, classify a borderline pixel correctly as belonging to the ball,
although it looks rather greenish, if we considered the context of the ball or its
motion model. However, in conventional vision systems, on the low level this
knowledge does not exist and on the higher level, the fact, that this pixel was
borderline in the classification, is lost due to committing to a hard decision on
the lower level.
To make this idea more concrete, we will now describe a ball tracking system
in the conventional data-driven design (Fig. 5). We will later discuss how it
would look like in the context based design we propose in this paper.
In a data-driven approach, a ball finder algorithm is executed on every image
yielding the 2D image circle (position, radius) corresponding to the ball (or
several hypotheses thereof). These circles are passed to a probabilistic estimator,
such as Gaussian-Maximum-Likelihood that finds the mostly likely states, i. e.
110
Fig. 5. Information flow in a ball tracker based on the conventional data-driven
paradigm (see text). The ball finder is not part of the optimization loop (gray box).
ball positions and velocities. Such an estimator is based on model likelihoods,
i. e. functions with uncertainty, that specify the following:
1. A camera model defines when the ball was at a certain position, where it
would be in the image. This links 3D positions to 2D circles.
2. A ball motion model defines when the ball was at a certain position having
a certain velocity, where it would be and which velocity it would have after
one timestep. This links 3D positions and velocities over time.
The estimator then finds the most likely states by minimizing a sum of squared
differences, which indicates how likely an assumed sequence of states is. It consists of 1. the squared differences between where the ball has been observed by
the ball finder and where the ball model would expect it to be observed if the
assumed state sequence was true and 2. the squared difference between the states
predicted by the ball model and the assumed corresponding states.
The models 1. and 2. are Gaussian likelihoods, i. e. formulas with uncertainty,
so there is a wealth of algorithms for handling this kind of problems. In particular, figure 5 is a conceptual description. Most actual implementations do not
minimize over the whole state sequence in every step but instead use an incremental estimator, such as the Extended or Unscented Kalman Filter [32, 6, 21].
However, this changes only the computation not the result. The key point of this
111
approach is that the ball finder, more general the computer vision, is outside the
optimization loop and there is no information flow from the estimator back to
the ball finder. Hence the ball finder cannot take the high level context of the
ball’s motion into account when deciding on the low level, whether something is
accepted as a ball or not.
4.2
Global Likelihood Optimization
We believe that much of the brittleness of current vision systems originates from
this way of committing to hard decisions on an early level of processing. So our
approach for increased robustness is an overall likelihood optimization.
The problem is to understand an image sequence, i. e. estimating over time.
Indeed, successive images are linked by a motion model and this provides most of
the context we want to build upon. However, we propose not to use incremental
filters, such as EKF, but to look back into the raw images of the last few seconds
at least. This approach has surprising advantages. Imagine the ball is kicked, but
during the first 100ms there is too little contrast to the background so it is not
detected. Now when it is detected, there is new information on where the ball
has been before from the ball’s motion model. The old images are still in memory
and tracking the ball back in time is much less ambiguous than finding the ball
without context and will probably succeed. Paradoxically, once the system has
detected the ball, it has already observed it for 100ms. The first useful prediction
is not delayed at all, because prior to that the ball must have been observed for
some time anyway. In the example above, we would realize this (Fig. 6) by
replacing the ball finder by
3. a ball appearance model that indicates for a given circle in the image plane
how much it looks like a ball. This links the 2D image circle to the actual
image.
In the data-driven approach the ball finder determines where the ball is and the
camera model computes where the ball should be according to the state. Then
both are compared with the difference indicating the likelihood of the state. In
the global likelihood optimization approach, instead the camera model computes
where the ball should be and then the ball appearance model computes how
“ball-like” the image looks there. The difference is, that now the ball appearance
model, i. e. the computer vision, is inside the optimization loop. Now the lower
layer does not have to commit early to a single image circle, but instead it
gradually assesses different circles in the image as requested by the minimization
algorithm. In this approach, an indistinct ball would get a lower likelihood in 3.
but this could be compensated by 1. and 2. if it fits well to the context of a flying
ball. This mechanism allows the now implicit ball finder to utilize the context
provided by the ball motion and implement the behavior described above.
As a final remark, the provided functionality greatly resembles the priming
mechanism known from psychology [33]. This makes us even more believe that
the approach of a global likelihood optimization directly in the images is an
elegant way to greatly increase robustness of computer vision algorithms.
112
Fig. 6. Information flow in a ball tracker based on the global likelihood optimization
paradigm (see text). Here, computer vision, i.e. the ball finder, is part of the optimization loop (gray box).
4.3
Exploiting Semantic Context
Doubtless, if the goal is to track a flying ball, the ball’s motion model provides
the most valuable context. However, there is other, more semantic information
that could be incorporated to increase robustness. Figure 7 illustrates how a
broader context could help distinguishing the real ball from false responses of
the computer vision. We have constructed this example to illustrate what we
believe could happen. It is not based on real computation.
In a) only the first image I1 is considered (p(ball1 |I1 )). The computer vision gives varying responses for different image circles. After applying the necessary threshold needed to give a final result, there are five possible ball locations in the example, namely the true ball and the player’s head, hands,
and both feet together. The same happens in b), where in the example the
9th image I9 is considered alone (p(ball9 |I9 )). Now in c) all images I1...9 are
considered together (p(ball9 |I1...9 , ball model)). The link between the images
is provided by the ball motion model as described in the previous section. In
the example, the player’s head and feet could be ruled out, because their motion does not correspond to the ball motion model. However, in this example, the two hands remain valid hypotheses, because their side-swinging motion would actual correspond to a distant flying ball. In d), the additional con-
113
a) p(ball1 |I1 )
b) p(ball9 |I9 )
c) p(ball9 |I1...9 , ball model)
d) p(ball9 |I1...9 , ball model, field geometry)
e) p(ball9 |I1...9 , ball model, field geometry, f) p(ball1 |I1...9 , ball model, field geometry,
game semantics)
game semantics)
Fig. 7. An illustration, how different types of context information could help to distinguish the real ball from things that look like a ball to the computer vision (see
text).
114
text of the field geometry is added (p(ball9 |I1...9 , ball model, field geometry)).
This context makes the player’s right hand being a ball very unlikely, because from the image position and radius of the circle, the ball would be outside of the field. In e), additional information about game semantics is added
p(ball9 |I1...9 , ball model, field geometry, game semantics). From the game situation, the system expects the player to kick the ball towards the camera. A ball
flying sideward as the falsely detected left hand is implausible with this context and can be ruled out. Of course, as explained above, once the system has
found a unique ball in image 9, it already knows backwards, via the ball motion model, which was the correct ball in image 1. This situation is shown in
f) (p(ball1 |I1...9 , ball model, field geometry, game semantics)). So the system has
already observed the ball for 9 images and can immediately provide a prediction.
How could this example be realized? First of all, using game semantics requires to detect the players, lines and goals to have information about the game
situation. But how could the semantics of soccer be modeled? The classical AI
approach would be to use logic, for instance ontologies to describe a soccer
match. However, typical ontologies, such as OWL [2] express crisp knowledge in
some sort of logic and are difficult to combine with the probabilistic and vague
knowledge discussed in the previous section.
While in the long run modeling a rich and complex semantics, as ontologies
provide, is desirable, probably the first step would be to start with a shallow
semantics that is more easily integrated with probabilistic motion and perception
models. A candidate would be to run a Hidden Markov Model (HMM) [17] for
each player to model the different actions (kicking, running, tackling, etc.) a
player may take, maybe subdivided into different phases. Then one could define
how each phase would affect the players motion and the motion of the ball.
The advantage is that easy algorithms for inference in HMM exist and even
more important that HMM can be well combined with continuous probabilistic
estimation by the Interacting Multiple Model Filter (IMM) [1, §11.6.6].
Overall, it is obvious that modeling the player’s behavior is essential for
understanding a soccer match as a whole and is helpful even for tracking the
ball. However, it is still rather unclear how the required background knowledge
can be modeled in a way that can effectively help a computer vision system.
5
Proposed Experiments
For a vision to become reality, realistic intermediate steps are necessary. It would
not help, if we build a vision system now but then had to wait until a human
level soccer robot is available. So we propose a sequence of experiments that,
without a humanoid robot, ultimately allows to verify that the proposed system
is appropriate for human level soccer (Fig. 8).
115
Fig. 8. Our proposed experiment: Mount a camera and an inertial sensor on the head
of a human soccer player and use them to extract all the information, a humanoid
soccer robot would need to take the human’s place.
5.1
Helmet Camera with Inertial Sensor
The basic idea is to let a human soccer player wear a helmet with a camera and
an inertial sensor and verify, that the information extracted by the vision system
from the sensor data, would allow a humanoid robot to take the human’s place.
As a first experiment, we propose to record data from a soccer match and
run the vision system on that data offline. Since it is hard to obtain groundtruth data, we would use our expert’s judgment to assess, whether the result
would be enough for a humanoid robot to play soccer. It is very advantageous to
work on recorded data allowing to reproduce results for debugging and analysis
and to run the system even if it is still not real-time. Overall, it allows to first
concentrate on functionality and robustness instead of computation time and
integration.
We have already conducted a very first experiment [21, 6], where the ball
and the field lines are manually extracted from the recorded images (available
on request). The ball’s trajectory is predicted by least-square estimation using
the likelihood functions 1. and 2., as well as corresponding equations for how
the inertial sensor observes the free motion of the camera (Fig.9). The results
indicate that if the ball can be detected in the image with about one pixel
precision, the prediction would be precise enough. We believe that these kinds
of studies which deliberately discard essential aspects, such as integration, realtime computation, or autonomy are undervalued by the community who favors
full system approaches. But even from a full system perspective, it is much more
valuable to obtain an extensive result on a subsystem which then can guide the
full system design than to do another increment on a full system.
116
2.0
UKF
Distance (m)
1.5
1.0
0.5
0.0
0
200
400
600
Time (ms)
800
1000
1200
Fig. 9. Predicting the trajectory of a flying ball from a moving camera-inertial system.
As an initial study, the ball, the lines, and the goal corners have been manually extracted from the images. From this data, the trajectory of the ball is predicted (left).
The right plot shows the error of the predicted touchdown point varying over time. It
shows, even though the camera is moving, the prediction is roughly precise enough for
interception.
5.2
Motion Capture Suit
Departing from the experiment above, one might ask whether more sensors are
needed than just camera and inertial. Both human and humanoid robot can
derive their motion from the joint angles. This provides the horizontal motion
(odometry) and the height over ground. The horizontal motion facilitates localization and the height derived from vision is much less precise. Indeed, we
experienced that the uncertain height is a major part of the error in Fig. 9.
An intriguing idea is to equip the human player with a tracker-less motion
capture suit [34] measuring joint angles. Apart from providing the kinematic
information discussed above, it also provides the trajectory of both feet. If the
player hits the ball, one can compare the predicted ball trajectory with the real
foot trajectory and evaluate the precision. This is important since in such an
experiment, ground truth is not available.
5.3
Virtual Reality Display
The experiments above have the drawback that they are evaluated by an expert
looking at the vision system’s output. The most direct proof that this is all you
need for playing soccer would be to give a human just that output via a head
mounted display and see whether s/he can play.
The approach is of course fascinating and direct, but we have some concerns
regarding safety. Anyway, this experiment becomes relevant only after we are
convinced in principle, that the system is feasible. So this is something to worry
about later.
117
6
Conclusion
In this position paper, we have outlined the road to a vision system for a humanrobot soccer match. We claim that, since soccer is a structured environment, the
basic techniques are available and the goal could be reached within a decade. The
main challenge will be robustness, which we propose to address by optimizing a
global likelihood function working on a history of raw images. We have outlined
a sequence of experiments to evaluate such a vision system with data from a
camera-inertial system mounted on the head of a human soccer player.
The reason for being confident such a system can be realized within a decade
is the insight that it does not need general common-sense-reasoning AI. This is
good news for the RoboCup 2050 challenge. But it suggests that, even when we
meet that challenge, it does not imply we have realized the dream of a thinking
machine, the whole challenge had started with.
That would not be the first time.
References
1. Y. Bar-Shalom, X. Li, and T. Kirubarajan. Estimation with Applications to Tracking and Navigation. John Wiley & Sons, Inc., 2001.
2. S. Bechhofer, F. van Harmelen, J. Hendler, I. Horrocks, D. McGuinness, P. PatelSchneider, and L. Stein. OWL web ontology language reference. Technical Report
REC-owl-ref-20040210/, W3C, 2004.
3. M. Beetz, S. Gedikli, J. Bandouch, B. Kirchlechner, N. v. Hoyningen-Huene, and
A. Perzylo. Visually tracking football games based on tv broadcasts. IJCAI 2007,
Proceedings of the 20th International Joint Conference on Artificial Intelligence,
Hyderabad, India, 2007.
4. M. Beetz, N. v. Hoyningen-Huene, J. Bandouch, B. Kirchlechner, S. Gedikli, and
A. Maldonado. Camerabased observation of football games for analyzing multiagent activities. In International Conference on Autonomous Agents, 2006.
5. G. Binnig. Cellenger automated high content analysis of biomedical imagery, 2004.
6. O. Birbach. Accuracy analysis of camera-inertial sensor based ball trajectory prediction. Master’s thesis, Universität Bremen, 2008.
7. H.-D. Burkhard, D. Duhaut, M. Fujita, P. Lima, R. Murphy, and R. Rojas. The
Road to RoboCup 2050. IEEE Robotics and Automation Magazine, 9(2):31–38,
2002.
8. E. R. Davies. Machine Vision. Theory, Algorithms, Practicalities. Morgan Kauffmann, 2004.
9. A. Ess, B. Leibe, and L. V. Gool. Depth and Appearance for Mobile Scene Analysis.
In Proceedings of the IEEE International Conference on Computer Vision, 2007.
10. U. Frese, B. Bäuml, S. Haidacher, G. Schreiber, I. Schaefer, M. Hähnle, and
G. Hirzinger. Off-the-shelf vision for a robotic ball catcher. In Proceedings of
the IEEE/RSJ International Conference on Intelligent Robots and Systems, Maui,
pages 1623 – 1629, 2001.
11. U. Frese and T. Laue. (A) vision for 2050: The road towards image understanding
for a human-robot soccer match. In Proceedings of the 5th International Conference
on Informatics in Control, Automation and Robotics, 2008.
118
12. D. Guru and B. Shekar. A simple and robust line detection algorithm based on
small eigenvalue analysis. Pattern Recognition Letters, 25(1):1–13, 2004.
13. S. Haddadin, T. Laue, U. Frese, and G. Hirzinger. Foul 2050: Thoughts on Physical
Interaction in Human-Robot Soccer. In Proceedings of the IEEE/RSJ International
Conference on Intelligent Robots and Systems, 2007.
14. S. Haddadin, T. Laue, U. Frese, S.Wolf, A. Albu-Schäffer, and G. Hirzinger. Kick
it with elasticity: Safety and performance in human-robot soccer. Robotics and
Autonomous System - Special Issue on Humanoid Soccer Robots, 2009. to appear.
15. D. Hoiem, A. A. Efros, and M. Hebert. Putting Objects in Perspective. In Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern
Recognition, 2006.
16. Honda Worldwide Site. Honda World Wide — Asimo, 2008. http://world.honda.
com/ASIMO/.
17. W. Hu, T. Tan, L. Wang, and S. Maybank. A survey on visual surveillance of object
motion and behaviors. IEEE Transactions on Systems, Man and Cybernetics, Part
C, 34(3):334–352, 2004.
18. H. Kitano and M. Asada. RoboCup Humanoid Challenge: That’s One Small Step
for A Robot, One Giant Leap for Mankind. In International Conference on Intelligent Robots and Systems, Victoria, pages 419–424, 1998.
19. J. J. Kuffner, S. Kagami, K. Nishiwaki, M. Inaba, and H. Inoue. Dynamically-stable
Motion Planning for Humanoid Robots. Auton. Robots, 12(1):105–118, 2002.
20. S. Kumar and M. Hebert. A Hierarchical Field Framework for Unified ContextBased Classification. In Proceedings of the IEEE International Conference on Computer Vision, 2005.
21. J. Kurlbaum. Verfolgung von ballflugbahnen mit einem frei beweglichen kamerainertialsensor. Master’s thesis, Universität Bremen, 2007.
22. T. Laue and T. Röfer. Particle filter-based state estimation in a competitive and
uncertain environment. In Proceedings of the 6th International Workshop on Embedded Systems. VAMK, University of Applied Sciences; Vaasa, Finland, 2007.
23. B. Leibe, K. Schindler, N. Cornelis, and L. V. Gool. Coupled Object Detection
and Tracking from Static Cameras and Moving Vehicles. IEEE Transactions on
Pattern Analysis and Machine Intelligence, 30(10):1683–1698, 2008.
24. D. Lowe. Distinctive image features from scale-invariant keypoints. International
Journal of Computer Vision, 60(2):91 – 110, 2004.
25. K. Price. The annotated computer vision bibliography, 2008. http://www.
visionbib.com/.
26. A. Rabinovich, A. Vedaldi, C. Galleguillos, E. Wiewiora, and S. Belongie. Objects
in Context. In Proceedings of the IEEE International Conference on Computer
Vision, 2007.
27. D. Ramanan and D. Forsyth. Finding and tracking people from the bottom up.
In IEEE Conference on Computer Vision and Pattern Recognition, 2003.
28. RoboCup Federation. RoboCup Official Site, 2008. http://www.robocup.org.
29. T. Röfer et al. GermanTeam RoboCup 2005, 2005. http://www.germanteam.org/
GT2005.pdf.
30. P. Stone, R. S. Sutton, and G. Kuhlmann. Reinforcement Learning for RoboCupSoccer Keepaway. Adaptive Behavior, 13(3):165–188, 2005.
31. S. Ullman. Sequence seeking and counter streams: A computational model for
bidirectional information flow in the visual cortex. Cerebral Cortex, 5(1), 1995.
32. G. Welch and G. Bishop. An introduction to the kalman filter. Technical Report
TR 95-041, University of North Carolina, 1997.
119
33. Wikipedia. Priming (psychology) — wikipedia, the free encyclopedia, 2008. [Online; accessed 10-December-2008].
34. Xsens Technologies B.V. Moven, Inertial Motion Capture, Product Leaflet, 2007.
120
Udo Frese received his Ph.D. from University of
Erlangen-Nürnberg in 2004. From 2004 until 2008
he has been a post-doc researcher in Bernd KriegBrückner’s lab at University of Bremen and in the
German Research Center for Artificial Intelligence
(DFKI). He is now assistant professor for real-time
computer vision at University of Bremen and further
on affiliated with the DFKI. His research topics are
real time computer vision, in particular simultaneous
localization and mapping (SLAM) and computer vision for sport robotics.
Tim Laue received his diploma in computer science
from the University of Bremen, Germany, in 2004.
Until 2008 he has been a researcher in Bernd KriegBrückner’s working group at the University of Bremen
Currently, he is with the German Research Center for
Artificial Intelligence at the research group for Safe
and Secure Cognitive Systems, Bremen, Germany. His
research and development fields are simulation and localization for autonomous robots, i. e. for wheelchairs
as well as for different soccer robots participating in
RoboCup competitions.
Oliver Birbach received the diploma degree in computer science from University of Bremen in 2008. He is
currently a junior researcher at the German Research
Center for Artificial Intelligence in Bremen. His main
research areas are multi-target tracking, context-based
computer vision and computer vision in sports environments.
Jörg Kurlbaum studied computer science at the University of Bremen with personal preference in robotics.
After he received the diploma in 2007, he started to
work in the SFB/TR8 Spatial Cognition group on map
building and representation for real-time operator applications. Jörg is affiliated with Udo Frese in his realtime vision group.
Thomas Röfer received his diploma in computer science and his Dr.ing. degree from Universität Bremen,
Germany, in 1993 and 1998, respectively. He is in the
Executive Committee of the RoboCup Federation, and
he is member of the Transregional Collaborative Research Center SFB/TR 8 “Spatial Cognition” at Bremen. He is currently with the German Research Center
for Artificial Intelligence at the research group for Safe
and Secure Cognitive Systems, Bremen, Germany. His
research interests include rehabilitation robotics, robot
soccer, real-time computer vision, world modeling, and
humanoid robots.
On the whereabouts of CSP-CASL – A Survey
The Processes & Data Gang
∗
Andy Gimblett, Temesghen Kahsai, Liam O’Reilly, Markus Roggenbach
Swansea University, United Kingdom
Abstract. Csp-Casl is but one of the many languages, Bernd KriegBrückner (BKB) was influential to: From the initial idea of working towards an integration of the process algebra Csp with the algebraic specification language Casl, over the concrete design of the language’s syntax,
to the tool’s side where the theorem prover Isabelle should provide the
common platform, BKB provided inspiration and guidance, funding, and
also a helping hand when needed.
This paper provides a survey on the technology developed so far for CspCasl, covering results of theoretical nature, an industrial case study, tool
support as well as a testing approach. In honour of BKB’s 60th birthday,
this survey documents what has become out of one of BKB’s visions.
1
Introduction
“Deep is the fountain of the past. Shouldn’t it be called unfathomable?”1 So,
where to start, when it comes to the formal specification language Csp-Casl
[23, 2]? Back in Munich, in the early 1970s, when Bernd Krieg-Brückner (BKB)
met Mila Majster, who later became Markus Roggenbach’s (MR) PhD supervisor? Or, jumping forward in time, at the NWPT 1996 in Oslo, when MR met
Christoph Lüth from BKB’s group? Or at the CoFI meeting in Paris-Cachan
in 1998, where various ideas on reactive Casl extensions were discussed? Even
geography is far from being stable when it comes to Csp-Casl — the language
originated in Bremen, and has now come into bloom at Berlin and Swansea.
What is clear is that BKB was – very much to MR’s dislike – a fan of Csp, and
so we truly can call BKB the matchmaker. The marriage of Csp and Casl was
finally announced at AMiLP 2003, where else but in the town of Romeo and
Juliet, Verona.
As unfathomable as Csp-Casl’s past might appear, Csp-Casl’s reason is
clear: Distributed computer applications like flight booking systems, web services, and electronic payment systems such as the EP2 standard [1] involve
the parallel processing of data. Consequently, these systems exhibit concurrent
aspects (e.g. deadlock-freedom) as well as data aspects (e.g. functional correctness). Often, these aspects depend on each other. Casl alone can deal with data
∗
1
This work was supported by EPSRC under the grant EP/D037212/1.
T. Mann: Joseph and his Brothers, S. Fischer, Berlin, 1933.
122
aspect, Csp is quite capable of modelling concurrent systems. The mutual dependencies between processes and data, however, require a new type of language:
e.g., Csp-Casl.
Previous approaches to integrate processes and data, such as LOTOS [6],
E-LOTOS [9], or CSPM [24] restrict data to initial or concrete data types. This
has the shortcoming that data needs to be fixed the moment the modelling is
done. A step forward in this respect is the specification language μCRL [4].
Here, data types have loose semantics and are specified in equational logic with
total functions. Casl, and thus Csp-Casl, offers a more expressive logic for the
specification of data, namely many-sorted first order logic with sort-generation
constraints, partiality, and sub-sorting. On the process side, μCRL employs a
branching time semantics based on bisimulation. Csp-Casl offers the choice
between the various trace semantics defined in the context of Csp, i.e. CspCasl provides a linear time view on systems. For Casl, various reactive extensions have been defined. They include CCS-CASL [25] and CASL-Charts [22],
which combine Casl with reactive systems of a particular kind, the semantics
of which are defined in terms of set theory. Casl extensions such as CoCASL
[15] (being much simpler than full set theory) and CASL-Ltl [21] provide more
of what one might call a meta-framework: for example, the process algebra CCS
has been formalized in both these languages.
Based on previous publications, this paper provides a survey on the technology available for Csp-Casl. It is organized as follows: Section 2 presents the
language and the basic ideas underlying its semantics, [23]. We then review a
major industrial case study, namely, how to model the electronic payment system EP2 in Csp-Casl, [3]. Section 4 reports on parsing and statically analyzing
Csp-Casl, [2]. Csp-Casl refinement [11] is the topic of Section 5: Besides its
obvious use for system development, refinement is the fundamental means to analyze systems within the Csp-Casl context, e.g. in order to check a specification
for deadlock-freedom. Section 6 documents the construction of Csp-Casl-Prover
[19, 20], which offers theorem proving support for checking Csp-Casl refinement.
With this technology around, we mechanically verify parts of the EP2 system in
Section 7: we prove that the formal specifications representing the different layers of the EP2 system are related via refinement, and demonstrate how to prove
deadlock freedom of typical EP2 dialogues. Modelling and formally analyzing
a system improves the quality of the design process, however, there still is no
direct link between specification and implementation. Section 8, finally, presents
a framework for black-box testing from Csp-Casl specifications, [12, 13].
2
The design of CSP-CASL
Csp-Casl [23, 2] is a comprehensive language which combines processes written
in Csp [5, 24] with the specification of data types in Casl [17]. The general idea
is to describe reactive systems in the form of processes based on Csp operators,
where the communications of these processes are the values of data types, which
are loosely specified in Casl. All standard Csp operators are available, such as
123
multiple prefix, the various parallel operators, operators for non-deterministic
choice, communication over channels. Concerning Casl features, the full language is avaliable to specify data types, namely many-sorted first order logic
with sort-generation constraints, partiality, and sub-sorting. Furthermore, the
various Casl structuring constructs are included, where the structured free
construct adds the possibility to specify data types with initial semantics.
Syntactically, a Csp-Casl specification with name N consists of a data part
Sp, which is a structured Casl specification, an (optional) channel part Ch to
declare channels, which are typed according to the data part, and a process
part P written in Csp, within which Casl terms are used as communications,
Casl sorts denote sets of communications, relational renaming is described by a
binary Casl predicate, and the Csp conditional construct uses Casl formulae
as conditions:
ccspec N = data Sp channel Ch process P end
See the below for concrete instances of this scheme. In choosing the loose semantics of Casl, semantically, such a Csp-Casl specification is a family of process
denotations for a Csp process, where each model of the data part Sp gives rise
to one process denotation.
The definition of the language Csp-Casl is generic in the choice of a specific
Csp semantics. Whether a Csp-semantics can be used within our construction
depends on the semantics’ requirements concerning what we call here the data
type of communications. This data type takes as values the alphabet of communications, but additionally provides certain test functions. In this respect, our
construction provides as operations
– test on equality for arbitrary Casl terms (can two communications synchronise?),
– test on membership for a Casl term concerning a Casl sort (does a communication belong to a certain subset of the alphabet of communications?),
– test whether a binary predicate holds between two Casl terms (are the terms
in a renaming relation?), and
– satisfaction of a Casl first order formula (is the formula of the conditional
construct true?).
These test functions can solely be formulated in Casl. Fig. 1 illustrates this setting. The above listed, seemingly small set of test operations allows for all denotational semantics described in [24], namely trace-semantics, failure-divergencesemantics and stable-failure-semantics.
The data types specified by algebraic specification consist of many-sorted
algebras. The data type of communications required by the process algebraic
semantics is a one-sorted algebra. Thus, in order to integrate data into processes,
we need to turn a many-sorted algebra into one set of values such that the above
described tests are closely connected with the original data type. We illustrate
our construction here for the example of many-sorted total algebras:
There are two natural ways to define the alphabet of communications in terms
of the carrier sets of a Casl model: union and disjoint union of all carrier sets.
124
Casl specification Sp
?
has as semantics
Subsorted signature (S , TF , PF , P , ≤)
Class of Σ-models M ; (Ms )s∈S family of carrier sets
Interface: Alphabet construction Set of communications A
?
is parameter for
Csp process P
?
denotes
j
Tests over A :
a = b?, a ∈ X ?, aRb?, ϕ?
Set of traces over A
Fig. 1. Translating data into an alphabet of communication
To illustrate the effect of both possibilities, consider the following Csp-Casl
specification:
data
sorts S , T
ops c : S ; d : T
process c → SKIP || d → SKIP
Its data part, written in Casl, provides two constants c and d of type S and
T , respectively. The process part, written in Csp with Casl terms denoting
communications, combines two processes by the synchronous parallel operator,
i.e. they have to agree on all actions.
The question is, may c and d synchronise or not? In all the various Csp
semantics, c and d synchronise iff they are equal. Now consider two isomorphic
Casl models A and B of the data part:
A(S ) = {∗}, A(T ) = {+}, A(c) = ∗, A(d ) = +
B(S ) = B(T ) = {}, B(c) = B(d ) = Choosing the union of all carrier sets as alphabet has the effect, that c and d
do not synchronise for algebra A while they synchronise for algebra B. Thus,
isomorphic algebras give rise to different behaviour. Therefore, we define the
alphabet to be the disjoint union — with the consequence that c and d do not
synchronise.
Similar ‘experiments’ for partiality, subsorting, and the combination of both
– see [23] – finally lead to an alphabet construction
Ms ∪ {⊥})/∼ .
A(M) := (
s∈S
Here M is a Casl model and S is the set of all sorts declared in the data part.
The special element ⊥ encodes partiality, while ∼ is an equivalence relation
which – on the alphabet level – deals with subsorting.
125
3
The first challenge: Modelling EP2
In [3] we describe the formal specification of a banking system using Csp-Casl.
The system in question is called EP2, which stands for EFT/POS 2000 , short for
‘Electronic Fund Transfer/Point Of Service 2000’; it is a joint project established
by a consortium of (mainly Swiss) financial institutes and companies in order to
define EFT/POS infrastructure for credit, debit, and electronic purse terminals
in Switzerland2 .
Formalising
Specifications
Informal
Design Process
Formal Design
Process
..
.
Architectural
Level
Abstract
Component
Level
Concrete
Component
Level
...
...
Formalisation
Csp-Casl Spec
SP0
...
Formalisation
Csp-Casl Spec
SP1
Formalisation
Csp-Casl Spec
SP2
..
.
...
Intuitive Refinement
Fig. 2. Formalising specifications at various levels of abstraction.
At the architectural level, the system consists of seven autonomous entities
centred around the EP2 Terminal (see Figure 2, which shows several of these
entities), a hardware device concerned with processing card details and authorising financial transactions. The other entities are the Cardholder (i.e. customer),
Point of Service/POS (i.e. cash register), Attendant, POS Management System/PMS , Acquirer , Service Center , and Card . These entities communicate with
the Terminal, and to a certain extent one another, mostly via XML messages
over TCP/IP; the messages contain information about authorisation, financial
transactions, and initialisation and status data. Each component is thus a reactive system, and there are both reactive parts and data parts which need to be
modelled, with these parts heavily intertwined.
2
www.eftpos2000.ch
126
The EP2 specification, a typical informal specification as used in industry,
consists of twelve documents, each of which either considers some particular component of the system in detail, or considers some aspect common to many or all
components. The major components have specification documents setting out
general , functional , and supplementary requirements, where the functional requirements carry the most detail, and consist mainly of use cases discussing how
that particular component behaves in various situations. Additionally, there is a
Security Specification, an Interface Specification and a Data Dictionary. The Security Specification discusses the various security mechanisms used within EP2,
many of which are inherited from EMV and other related specifications. The
Interface Specification describes every interface between components in terms of
the messages allowed on the interface, and in what order they may flow. The
Data Dictionary describes in more detail the data elements found in messages
and, in some cases, in the state of some component. It includes brief discussion
of every data element occurring in the system. The documents are written using
a number of different specification notations: plain English; UML-like graphics
(use cases, activity diagrams, message sequence charts, class models, etc.); pictures; tables; lists; file descriptions; encoding rules. In these regards it is entirely
typical of a modern industrial strength system specification.
With such a document structure, the information required to understand a
particular aspect of the system is contained in several different documents, each
of which has its own things to say about the situation in question. For example,
the SI-Init interface between Terminal and Acquirer is documented in the
Terminal Specification, the Acquirer Specification, the Interface Specification,
and the Data Dictionary. In general, as we found, this results in a specification
which is difficult to understand, and which easily leads to inconsistencies and
ambiguities.
As such, this is exactly the kind of system we would like to specify formally,
with all the associated benefits of tool support that this brings. When [3] was
written, such tool support for Csp-Casl was unavailable; nonetheless, we were
able to specify in Csp-Casl a number of aspects of the system at varying levels
of abstraction, as indicated in Figure 2, and corresponding to different views used
in the informal specification documents. In particular, we wrote an architectural
specification providing the system overview, and abstract component specifications of several elements including the SI-Init interface between Terminal and
Acquirer. In doing so, we identified a number of ambiguities and even contradictions in the EP2 specification. For example, the order of communications over
the SI-Init interface was specified in two contradictory ways in two different
documents. Thus, the very act of writing specifications in a formal language
required a level of rigour at which such problems were forced into the light.
With our specifications we were then able to perform (relatively simple)
proofs on paper — for example of the deadlock-freedom of part of our SI-Init
specification. It quickly became clear, however, that in order to deal with the
full complexity of a specification of this sort, tool support would be absolutely
essential.
127
4
From blackboard langugage to tool support
For quite a while, Csp-Casl existed only as a ‘blackboard language’: while the
subject of several papers, its syntax and semantics remained malleable, allowing
experimentation and discussion. In order to achieve its full promise, however, it
was clear that tool support would be required. The first step towards such tool
support was to fix Csp-Casl’s syntax and static semantics, and to implement
them within an appropriate tool. This work is described fully in [2].
Element
Machine readable
Named process
P (t1 ,...,tn )
Skip
SKIP
Stop
STOP
Div
DIV
Run
RUN(es)
Chaos
CHAOS(es)
Term prefix
t->p
Internal prefix choice
|~|x ::s->p
External prefix choice
[]x ::s->p
Channel send
c!t->p
Channel nondeterministic send c!x ::s->p
Channel receive
c?x ::s->p
Sequential composition
p;q
Internal choice
p|~|q
External choice
p[]q
Interleaving
p|||q
Synchronous
p||q
Generalised parallel
p[|es|]q
Alphabetised parallel
p[es1 ||es2 ]q
Hiding
p\es
Renaming
p[[R]]
Conditional
if ϕ then p else q
Pretty printed Precedence
P (t1 , . . . , tn )
0
SKIP
0
STOP
0
DIV
0
RUN (es)
0
CHAOS (es)
0
t →p
1
x :: s → p
1
2 x :: s → p
1
c!t → p
1
c!x :: s → p
1
c?x :: s → p
1
p;q
2
pq
3
p2q
3
p ||| q
4
p q
4
p|[es]|q
4
p|[es1||es2 ]|q
4
p \ es
5
p[R]
5
if ϕ then p else q
6
Fig. 3. Summary of Csp-Casl process terms.
Csp-Casl is presented as a Casl extension language; its syntax and static
semantics extend, and refer to, those of Casl; as such, they are formalised in a
style matching that of Casl found in [17]. The abstract and concrete syntaxes
are written in the normal BNF-like style (see figure 3 for an informal overview
of the concrete syntax of process terms), whereas the static semantics is in the
style of natural semantics [10]. As a representative example, here is the static
semantics rule for a conditional process term:
Σ, Λ, G, L P1 p, αp
Σ, Λ, G, L P2 q, αq
(Σ, G ∪ L) F ϕ
Σ, Λ, G, L condition-proc F P1 P2 if ϕ then p else q, αp ∪ αq ∪ sorts(ϕ)
128
The rule shows how to devolve the static check on an abstract syntax element for a conditional process term (condition-proc F P1 P2 ), into checks on
the condition (a Casl formaula F) and processes (P1 and P2 ) involved. It operates within some context (a Casl signature Σ; a so-called Csp-Casl signature
Λ, containing process and channel declarations; and global and local variable
bindings G and L), elements of which then provide context for the devolved
rules. Note that the FORMULA rule is found in the Casl static semantics. If the
devolved checks succeed, they yield semantic objects representing the formula
(ϕ), process terms (p, q) and terms’ constituent alphabets (αp , αq ), which are
combined into the semantic objects yielded as the overall product of the rule: an
object representing the entire conditional term, and an object representing its
constituent alphabet, itself the subject of later checks, at the process equation
level.
The syntax and static checks are implemented within the framework of the
heterogeneous toolset, Hets [16]. Casl sits at the centre of a family of related
specification languages – its extensions and sublanguages – and the language
HetCasl[14] then provides a framework for heterogeneous specification over
this family.3 That is, it enables specification using a mix of languages, rather
than requiring all aspects of a system to be specificied using the same formalism.
Hets then provides tool support for heterogeneous specification in this setting.
It consists of a general framework for working with specifications across multiple
logics, with facilities for parsing, static analysis, error reporting, translation between logics, and integration with proof tools; logics for the various Casl-family
languages (including Csp-Casl) are implemented within this framework.
5
Refinement and property verfication in CSP-CASL
Our notions of refinement for Csp-Casl [11, 23] are based on refinements developed in the context of the single languages Csp and Casl. For a denotational
Csp model with domain D, the semantic domain of Csp-Casl consists of families of process denotations dM ∈ D over some index set I ,
(dM )M ∈I
where I is a class of Casl models over the same signature. Intuitively, a refinement step, which we write here as ‘;’, reduces the number of possible implementations. Concerning data, this means a reduced model class, concerning
processes this mean less non-deterministic choice:
Definition 1 (Model class inclusion). Let (dM )M ∈I and (dM
)M ∈I be families of process denotations.
(dM )M ∈I ;D (dM
)M ∈I ⇔ I ⊆ I ∧ ∀ M ∈ I : dM %D dM .
3
In fact, HetCasl is generic in its language family, and could be used as a heteregenous specification framework in an entirely different setting.
129
Here, I ⊆ I denotes inclusion of model classes over the same signature, and
%D is the refinement notion in the chosen Csp model D. In the traces model
T we have for instance P %T P :⇔ traces(P ) ⊆ traces(P ), where traces(P )
and traces(P ) are prefixed closed sets of traces. Given Csp-Casl specifications
Sp = (D, P ) and Sp = (D , P ), by abuse of notation we also write
(D, P ) ;D (D , P )
if the above refinement notion holds for the denotations of Sp and Sp , respectively. On the syntactic level of specification text, we additionally define the
notions of data refinement and process refinement in order to characterize situations, where one specification part remains constant: In a data refinement, only
the data part changes:
(D, P ) ; (D , P )
data
if
Σ(D) = Σ(D ) and Mod(D ) ⊆ Mod(D)
In a process refinement, the data part is constant:
proc
(D, P ) ;D (D, P )
if
for all M ∈ Mod(D) : den([[P ]]M ) %D den([[P ]]M )
Here, den([[P ]]M ) is the denotation of the process P over the model M in the
chosen Csp denotational model. Clearly, both these refinements are special forms
of Csp-Casl refinement in general, for which the following holds:
Theorem 1. Let Sp = (D, P ) and Sp = (D , P ) be Csp-Casl specifications,
where D and D are data specifications over the same signature. Let (D , P ) be
a new Csp-Casl specification that consists of the process part P of Sp and the
data part D of Sp . For these three specifications holds:
(D, P ) ;D (D , P )
⇐⇒
proc
data
(D, P ) ; (D , P ) and (D , P ) ;D (D , P )
This result forms the basis for the Csp-Casl tool support, see Section 6. In
order to prove that a Csp-Casl refinement (D, P ) ;D (D , P ) holds, first one
uses proof support for Casl alone, see [16], in order to establish Mod(D ) ⊆
Mod(D). Independent of this, one has then to check the process refinement
P %D P .
A similar result can be obtained if the signature of D is embedded into
the signature of D , where σ denotes the signature morphism induced by the
embedding, see [11] for the technical details:
Theorem 2. Let Sp = (D, P ) and Sp = (D , P ) be Csp-Casl specifications, such that the signature of D is embedded into the signature of D . Then
proc
data
(D, P ) ;σ (D , P ) and (D , P ) ;D (D , P ) imply (D, P ) ;σD (D , P ).
For system development one often is interested in liberal notions of refinements, which allow for substantial changes in the design. For system verification,
however, it is important that refinement steps preserve properties. This allows to
130
verify properties on abstract specifications which in general are less complex than
the more concrete ones. Here, we show how to perform deadlock free analysis
using refinement.
In Csp, deadlock is represented by the process STOP . In the Csp context,
the stable failures model F is best suited for deadlock analysis. The model F
records two observations on processes: the first observation is the set of traces a
process can perform, this observation is given by the semantic function traces;
the second observation are the so-called stable failures, given by the semantic
function failures. A failure is a pair (s, X ), where s represents a trace that the
process can perform, after which the process can refuse to engage in all events
of the set X .
A Csp-Casl specification is deadlock-free, if it enforces all its possible implementations to be deadlock-free. On the semantical level, we capture this idea
as follows:
Definition 2. Let (dM )M ∈I be a family of process denotations over the stable
failures model, i.e., dM = (TM , FM ) ∈ F(A(M )) for all M ∈ I , where A(M ) is
the alphabet induced by the model M , see Section 2.
– A denotation dM is deadlock free if (s, X ) ∈ FM implies that X = A(M ) .
– (dM )M ∈I is deadlock-free if for all M ∈ I it holds that dM is deadlock-free.
Deadlock can be analyzed through refinement checking: The most abstract deadlock free Csp-Casl specification over a Casl signature Σ with a set of sort
symbols S = {s1 , . . . , sn } is defined as:
ccspec DFS =
data . . . declaration of Σ . . .
process DF = (!x : s → DF ) Skip
s:S
end
We observe that DFS is deadlock-free. This result on DFS extends to a complete
proof method for Csp-Casl deadlock analysis:
Theorem 3. A Csp-Casl specification (D, P ) is deadlock free iff DFS ;F
(D, P ). Here S is the set of sorts of the signature of D.
Similar ideas allow to employ the notion of refinement to check for divergencefreedom, see [11].
6
CSP-CASL-Prover
Csp-Casl-Prover [19, 20] exploits the fact that Csp-Casl refinement can be
decomposed into first a refinement step on data only and then a refinement
step on processes, see the previous section. The basic idea is to re-use existing
tools for the languages Casl and Csp, namely for Casl the tool Hets [16]
and for Csp the tool Csp-Prover [7, 8], both of which are based on the theorem
prover Isabelle/HOL [18]. The main issue in integrating the tools Hets and CspProver into a Csp-Casl-Prover is to implement – in Isabelle/HOL – Csp-Casl’s
131
construction of an alphabet of communications out of an algebraic specification
of data written in Casl. In Csp-Casl-Prover, we choose to prove the relation ∼ –
see Section 2 – to be an equivalence for each Csp-Casl specification individually.
This adds an additional layer of trust: complementing the algorithmic check of
a static property, we provide a proof in Isabelle/HOL that the construction is
valid. The alphabet construction, the formulation of the justification theorems
(establishing the equivalence relation), and their proofs can all be automatically
generated.
CSP-CASL Prover
Hets +
Process & Data
Refinement
(Sp’,P) <= (Sp’,P’)
Hets
Translator
CSP-Prover
Translated
Process & Data
Refinement
([Sp’],[P]) <= ([Sp’],[P’])
Theory Files
+
Isabelle
Theorem
Prover
Refinement
Holds /
Doesn’t Hold
Interactive
Theorem
Proving
Fig. 4. Diagram of the basic architecture of Csp-Casl-Prover.
The architecture of Csp-Casl-Prover is shown in Figure 4. Csp-Casl-Prover
takes a Csp-Casl process refinement statement as its input. The Csp-Casl
specifications involved are parsed and transformed by Csp-Casl-Prover into a
new file suitable for use in Csp-Prover. This file can then be directly used within
Csp-Prover to interactively prove if the Csp-Casl process refinement holds.
The alphabet construction discussed in Section 2 depends on the signature
of the data part of a Csp-Casl specification, e.g., on the set of sorts S , see
Section 2. Hets, however, produces a shallow encoding of Casl only, i.e., there
is no type available that captures the set of all sorts of the data part of a
Csp-Casl specification. Consequently, it is impossible to give a single alphabet
definition within Isabelle/HOL which is generic in the data part of a Csp-Casl
specification. Instead, we produce an encoding individually crafted for the data
part of any Csp-Casl specification.
7
Towards a verification of EP2
We consider two levels of the EP2 specification, namely: the architectural level
(Arch) and the abstract component level (ACL). We choose a dialogue between
132
Informal Design
Process
Formal Design
Process
..
.
Architectural
Level
...
...
Formalisation
Arch Init
=T
Csp-Casl
Csp-Casl
Abstract
Component
Level
..
.
...
σ T
...
Formalisation
RUN Arch
σ T
F
ACL Init
=F
F
Seq Init
Fig. 5. Refinement in EP2
the Terminal and the Acquirer. In this dialogue, the terminal and the acquirer
are supposed to exchange initialization information. For presentation purposes,
we study here only a nucleus of the full dialogue, which, however, exhibits all
technicalities present in the full version.
Our notion of Csp-Casl refinement mimics the informal refinement step
present in the EP2 documents: There, the first system design sets up the interface between the components (architectural level), then these components are
developed further (abstract component level). Here, we demonstrate how we can
capture such an informal development in a formal way, see Fig. 5.
We first specify the data involved using Casl only. The data specification
of the architectural level (D Arch GetInit) requires only the existence of
sets of values: spec D Arch GetInit = sort D SI Init end. In the EP2
system, these values are communicated over channels; data of sort D SI Init
is interchanged on a channel C SI Init linking the terminal and the acquirer.
On the architectural level, both these processes just ‘run’, i.e., they are always
prepared to communicate an event from D SI Init or to terminate. We formalize
this in Csp-Casl:
ccspec Arch Init =
data D Arch GetInit
channel C SI Init : D SI Init
process
let Acquirer = EP2Run
T erminal = EP2Run
in T erminal |[ C SI Init ]| Acquirer
end
where EP 2Run = (C SI Init ? x : D SI Init → EP 2Run) 2 SKIP . On the
abstract component level (D ACL GetInit), data is refined by introducing a
type system on messages. In Casl, this is realised by introducing subsorts of
D SI Init. For our nucleus, we restrict ourselves to four subsorts, the original
dialogue involves about twelve of them.
spec D ACL GetInit =
sorts SesStart, SesEnd, DataRequest, DataResponse < D SI Init
ops r : DataRequest; e : SesEnd
133
axioms ∀ x : DataRequest; y : SesEnd.¬(x = y)
∀ x : DataRequest; y : SesStart.¬(x = y)
∀ x : DataResponse; y : SesEnd.¬(x = y)
∀ x : DataResponse; y : SesStart.¬(x = y)
end
In the above specification, the axioms prevent confusion between the different
sorts. Using this data, we can specify the EP2 system in Csp-Casl. In the
process part the terminal (TerInit) initiates the dialogue by sending a message
of type SesStart; on the other side the acquirer (AcqInit) receives this message.
In AcqConf , the acquirer takes the internal decision either to end the dialogue by
sending the message e of type SesEnd or to send another type of message. The
terminal (TerConf ), waits for a message from the acquirer, and depending on
the type of this message, the terminal engages in a data exchange. The system
as a whole consists of the parallel composition of terminal and acquirer:
ccspec ACL Init =
data D ACL GetInit
channels C ACL Init : D SI Init
process
let AcqInit = C ACL Init ? session : SesStart → AcqConf
AcqConf =C ACL Init ! e → Skip
C ACL Init ! r
→ C ACL Init ? resp : DataResponse → AcqConf
T erInit = C ACL Init ! session : SesStart → T erConf
T erConf =C ACL Init ? conf M ess →
(if (conf M ess : DataRequest)
then C ACL Init ! resp : DataResponse → T erConf
else if (conf M ess : SesEnd) then SKIP else ST OP )
in T erInit |[ C ACL Init ]| AcqInit
end
Theorem 4. Arch Init ;σT ACL Init
Proof. Using tool support, we establish this refinement by introducing two intermediate specifications RUN Arch and Seq Init:
ccspec RUN Arch =
data D Arch GetInit
channel C SI Init : D SI Init
process EP2Run
end
ccspec Seq Init =
data D ACL GetInit
channels C ACL Init : D SI Init
process
let SeqStart =C ACL Init ! session : SesStart → SeqConf
SeqConf =C ACL Init ! e → Skip
C ACL Init ! r
→ C ACL Init ! resp : DataResponse → SeqConf
in SeqStart
end
134
With Csp-Casl-Prover we proved: Arch Init =T RUN Arch. Now we want
to prove that RUN Arch ;σT Seq Init. Using HETS we showed:
data
D Arch GetInit ;σ D ACL GetInit. Now, we formed the specification (D ACL GetInit, PSeq Init ) and showed in Csp-Casl-Prover that, over
the traces model T , this specification refines to Seq Init. Here, PSeq Init denotes the process part of Seq Init. [20] proves ACL Init =F Seq Init. As
stable failure equivalence implies trace equivalence, we obtain ACL Init =T
Seq Init. Fig. 5 summarizes this proof structure.
As ACL Init involves parallel composition, it is possible for this system
to deadlock. Furthermore, the process TerConf includes the Csp process STOP
within one branch of its conditional. Should this branch of TerConf be reached,
the whole system will be in deadlock. The dialogue between the terminal and
the acquirer for the exchange of initialization messages have been proven to
be deadlock free in [20]. Specifically, it has been proven that the following reproc
finement holds: Seq Init ;F ACL Init, where Seq Init is a sequential
system. Sequential system are regarded to be deadlock-free. With our proof
method from Section 5, we can strengthen this result by actually proving that
Seq Init is deadlock free. To this end, we proved with Csp-Casl-Prover that
proc
DF ;F Seq Init where DF is the least refined deadlock-free process.
8
Specification based testing
The fundamental question when dealing with testing based on formal specifications is the following: which are the intentions or properties formulated in the
specification text? In the end, each test case reflects some intentions described
in the specification.
In order to study the above question for Csp-Casl, in [12] we undertake a
specification exercise in modelling a one-bit calculator. It has two input buttons and can compute the addition function only. Whenever one of the buttons
is pressed on the calculator, the integrated control circuit displays the corresponding digit in the display. After pressing a second button, the corresponding
addition result is displayed and the calculator returns to its initial state. In a
first high-level specification we might want to abstract from the control flow and
just specify the interfaces of the system:
ccspec BCalc0 =
data sort N umber
ops 0, 1 : N umber;
+ : N umber × N umber →? N umber
channel Button, Display : Number
process P0 = (?x : Button → P0 ) (!y : Display → P0 )
end
Relatively to BCalc0, the process
T0 = Button!0 → Button!0 → STOP
135
encodes ‘left open’ behaviour: The specification says that the process P0 internally decides if it will engage in an event of type Button or in an event of type
Display. Thus, if the environement offers to press the zero button twice, a legal implementation of BCalc0 can in one experiment refuse to engage in this
behaviour (by deciding internally that initially it engages in an event of type
Display), while in a second run it engages in it.
A more refined version requires that the first displayed digit is echoing the
input, and the second displays the result of the computation:
ccspec BCalc1 =
data sort N umber
ops 0, 1 : N umber;
+ : N umber × N umber →? N umber
channel Button, Display : Number
process P1 =?x : Button → Display!x → ?y : ButtonDisplay!(x +y) → P1
end
Relatively to BCalc1, now the process T0 encodes ‘forbidden’ behaviour as it
does not alternate between pressing buttons and display of digits. The process
T1 = Button!0 → Display!0 → Button!1 → Display!1 → STOP
however, again represents ‘left open’ behaviour: We haven’t yet specified the
arithmetic properties of addition, i.e. it is undecided yet if 0 + 1 = 1 or 0 + 1 = 0.
This underspecification can be resolved by adding suitable axioms:
ccspec BCalc2 =
data sort
N umber
ops
0, 1 : N umber;
+ : N umber × N umber →? N umber
axioms 0 + 0 = 0; 0 + 1 = 1; 1 + 0 = 1
channel Button, Display : Number
process P2 =?x : Button → Display!x → ?y : ButtonDisplay!(x +y) → P2
end
Relatively to BCalc2, T1 represents ‘required’ behaviour.
We encode the intention of a test case with respect to a specification by
a colouring scheme. Intuitively, green test cases reflect required behaviour of
the specification. Red test cases reflect forbidden behaviour of the specification.
A test is coloured yellow if it depends on an open design decision, i.e., if the
specification does neither require nor disallow the respective behaviour. Internal
non-determinism and underspecification are the sources of yellow tests.
Formally, a test case is just a Csp-Casl process in the same signature as
the specification, which may additionally use first-order variables ranging over
communications. The colour of a test case T with respect to a Csp-Casl specification (Sp, P ) is a value c ∈ {red , yellow , green}. [12] provides, besides a mathematical definition of test colouring, also a syntactic characterization of test
colouring in terms of refinement. This enables the use of Csp-Casl-Prover in
order to check the colour of a test case.
136
We call a Csp-Casl refinement ‘;’ well-behaved, if, given specifications
(D, P ) ; (D , P ) with consistent data parts D and D , the following holds
for any test process T over (D, P ):
1. colour (T ) = green with respect to (D, P ) implies
colour (T ) = green with respect to (D , P ), and
2. colour (T ) = red with respect to (D, P ) implies
colour (T ) = red with respect to (D , P ).
This means: If a test case T reflects a desired behavioural property in (D, P ),
i.e. colour (T ) = green, after a well-behaved development step from (D, P ) to
(D , P ) this property remains a desired one and the colour of T is green. If a test
case reflects a forbidden behavioural property in (D, P ), i.e. colour (T ) = red ,
after a well-behaved development step from (D, P ) to (D , P ) this property remains a forbidden one and the colour of T is red. A well-behaved development
step can change only the colour of a test case T involving an open design decision, i.e. colour (T ) = yellow . For the various Csp-Casl refinements holds:
Csp-Casl data refinement is well-behaved; Csp-Casl process refinement based
on the Csp models T is not well-behaved; Csp-Casl process refinement based
on Csp models N and F is well-behaved, provided the processes involved are
divergence-free. [13] studies preservation of test colourings under enhancement.
In terms of refinement, for the above specifications holds: BCalc0 ;F
BCalc1 ;F BCalc2. The first step is a process refinement based on F, the
last one is a data refinement, i.e. both refinement steps are well behaved.
In [12, 13], we develop these ideas towards a testing framework: Given a test
and its colour, its execution on a system under test (SUT) is defined algorithmically. The test result is obtained by comparing intended and actual behaviour of
a SUT, where the test verdict is either pass, fail or inconclusive. Intuitively, the
verdict pass means that the test execution increases our confidence that the SUT
is correct with respect to the specification. The verdict fail means that the test
case exhibits a fault in the SUT, i.e., a violation of the intentions described in
the specification. The verdict inconclusive means that the test execution neither
increases nor destroyes our confidence in the correctness of the SUT.
In an industrial cooperation with Rolls-Royce, Derby, UK, our testing approach has successfully been applied to the contol software of air plane engines.
9
Summary and future work
Fig. 6 provides an overview of the Csp-Casl technology discussed: Given an
informal system description such as the EP2 documents, we mirror informal
specifications by formal ones written in Csp-Casl. In this process, the static
analysis of Csp-Casl specifications provides valuable feedback and helps to
avoid modelling errors. Then, we analyze if a Csp-Casl specification exhibits
desired properties, e.g. deadlock freedom. On the theoretical side, we showed
that – for some fundamental system properties – such an analysis can be carried out by studying an equivalent refinement problem. Thus, on the practical
137
side, Csp-Casl-Prover offers mechanized proof support for the analysis. Additionally, we mirror the informal development steps present in such a system
description by Csp-Casl refinement relations. Using Csp-Casl-Prover, we can
verify if these refinement relations actually hold. Properties are guaranteed to be
preserved within a development step provided a suitable Csp-Casl refinement
holds, e.g. deadlock-freedom is preserved under Csp-Casl-stable-failures refinement. Finally, using testing [12, 13], we link the formal specifications with actual
implementations. Csp-Casl refinement relations ensure that test cases which
are designed at an early stage can be used without modification for the test of
a later development stage. Hence, test suites can be developed in parallel with
the implementation. Also testing has tool support in terms of Csp-Casl-Prover,
e.g. in order to colour test cases.
Fig. 6. The Csp-Casl verification approach
Future research will include the integration of the Casl institution independent structuring mechanisms into Csp-Casl, the development of compositional
proof methods in Csp-Casl-Prover, as well as a semantically sound connection
to distributed Java programs.
138
Acknowledgment The authors wish to thank B. Krieg-Brückner for inspiring
Csp-Casl in the first place, Y. Isobe for his pioneering work on Csp-Prover,
B.-H. Schlingloff for being the driving force in applying Csp-Casl to real world
examples, and Erwin R. Catesbeiana (Jr) for being the mastermind of our gang.
References
1. eft/pos 2000 Specification, version 1.0.1. EP2 Consortium, 2002.
2. A. Gimblett. Tool Support for CSP-CASL, 2008. Swansea University, MPhil thesis.
3. A. Gimblett, M. Roggenbach, and B.-H. Schlingloff. Towards a formal specification
of an electronic payment system in CSP-CASL. In WADT 2004, LNCS 3423.
Springer, 2005.
4. J. Groote and A. Ponse. The syntax and semantics of μCRL. In Algebra of
Communicating Processes ’94, Workshops in Computing Series. Springer, 1995.
5. C. A. R. Hoare. Communicating Sequential Processes. Prentice Hall, 1985.
6. ISO 8807. LOTOS - a formal description technique based on the temporal ordering
of observational behaviour. Geneva, 1989.
7. Y. Isobe and M. Roggenbach. A generic theorem prover of CSP refinement. In
TACAS 2005, LNCS 3440, pages 108–123. Springer, 2005.
8. Y. Isobe and M. Roggenbach. CSP-Prover - a proof tool for the verification of
scalable concurrent systems. JSSST Computer Software, 25, 2008.
9. ISO/IEC. E-LOTOS, 2001.
10. G. Kahn. Natural Semantics. In STACS, LNCS 247, pages 22–39. Springer, 1987.
11. T. Kahsai and M. Roggenbach. Refinement notions for CSP-CASL. In WADT
2008. Springer, to be published.
12. T. Kahsai, M. Roggenbach, and B.-H. Schlingloff. Specification-based testing for
refinement. In SEFM 2007, pages 237–247. IEEE Computer Society, 2007.
13. T. Kahsai, M. Roggenbach, and B.-H. Schlingloff. Specification-based testing for
software product lines. In SEFM 2008. IEEE Computer Society, 2008.
14. T. Mossakowski. Heterogeneous specification and the heterogeneous tool set. Habilitation thesis, Universitaet Bremen, 2005.
15. T. Mossakowski, L.Schröder, M. Roggenbach, and H. Reichel.
Algebraiccoalgebraic specification in CoCASL. JALP, 2006.
16. T. Mossakowski, C. Maeder, and K. Lüttich. The Heterogeneous Tool Set, HETS.
In TACAS 2007, LNCS 4424, pages 519–522. Springer, 2007.
17. P. D. Mosses, editor. CASL Reference Manual. LNCS 2960. Springer, 2004.
18. T. Nipkow, L. Paulon, and M. Wenzel. Isabelle/HOL. LNCS 2283. Springer, 2002.
19. L. O’Reilly. Integrating Theorem Proving for Processes and Data, 2008. Swansea
University, MPhil thesis.
20. L. O’Reilly, Y. Isobe, and M. Roggenbach. CSP-CASL-Prover – a generic tool for
process and data refinement. In AVOCS’08, ENTCS, to appear.
21. G. Reggio, E. Astesiano, and C. Choppy. CASL-LTL — a Casl extension for
dynamic reactive systems. Technical Report TR-99-34, Università di Genova, 2000.
22. G. Reggio and L. Repetto. CASL-Charts. In AMAST’00, LNCS 1816, pages
243–257. Springer, 2000.
23. M. Roggenbach. CSP-CASL: A new integration of process algebra and algebraic
specification. Theoretical Computer Science, 354(1):42–71, 2006.
24. A. Roscoe. The theory and practice of concurrency. Prentice Hall, 1998.
139
25. G. Salaün, M. Allemand, and C. Attiogbé. A formalism combining CCS and Casl.
Technical Report 00.14, University of Nantes, 2001.
Andy Gimblett studied in Cardiff and received an
MPhil at Swansea University for research on tool support for CSP-CASL undertaken under the supervision
of Markus Roggenbach in the Processes and Data research group. Currently he is working in the Future
Interaction Technology Lab at Swansea University on
the project “Formally-based tools for user interface
analysis and design”.
Temesghen Kahsai joined the ’Processes and Data’
project in October 2006 as a PhD student. He is researching in the area of refinement and testing for
CSP-CASL. Before coming to Swansea, he was a student at Udine University (Italy).
Liam O’Reilly completed his BSc and masters studies at Swansea University in Wales. He is currently
studying a PhD under the supervision of Markus
Roggenbach at Swansea University. He is a member
of the ’Processes and Data Group’ currently specialising in the development and implementation of tool
support for CSP-CASL.
Markus Roggenbach joined BKB’s group in November 1998, worked on all the things around CASL, especially the real ones, and was also playing around
within the MMISS project. He is famous for promoting the ‘CoFI’ time and is a special friend of Erwin
R. Catesbeiana. He left Bremen in May 2003 to become a lecturer in Swansea, Wales. There he founded
his research group ‘Processes & Data Gang’.
Enhanced Formal Verification Flow for Circuits
Integrating Debugging and Coverage Analysis
Daniel Große, Görschwin Fey, and Rolf Drechsler
Institute of Computer Science, University of Bremen,
28359 Bremen, Germany
{grosse,fey,drechsle}@informatik.uni-bremen.de
Abstract. In this paper we present the state-of-the-art in formal hardware verification with the focus on automatic proof techniques. On top
of necessary fundamentals we describe the well accepted and widely used
industrial technique of Bounded Model Checking (BMC). Thereafter, two
major methodological improvements are proposed which significantly
improve the quality of BMC-based verification:debugging and coverage
analysis. The verification engineer can locate the source of a failure with
an automatic debugging approach. Components are identified which explain the discrepancy between the property and the circuit behavior. This
method is complemented by an approach to analyze functional coverage
of the proven BMC properties. The approach automatically determines
whether the property set is complete or not. In the latter case coverage
gaps are returned. Both techniques are integrated in an enhanced verification flow. A running example demonstrates the resulting advantages.
1
Introduction
For economic reasons the number of components in integrated circuits grows at
an exponential pace according to Moore’s law. This growth is expected to continue for another decade. Resulting is the so-called design gap, the productivity
in circuit design does not increase as fast as the technical capabilities. Thus,
more components can be integrated on a physical device than can be assembled
in a meaningful way during circuit design. The verification gap is even wider,
i.e. how to ensure the correctness of a design. This is of particular importance
in safety critical areas like traffic or security-related systems storing confidential
information. Thus, techniques and tools for the verification of circuits received
a lot of attention in the area of Computer Aided Design (CAD).
Simulation has always been in use as a fast method to validate the expected
functionality of circuits. Once the circuit is described in a Hardware Description
Language (HDL) like Verilog or VHDL, a testbench is used to excite the circuit
under stimuli expected during in-field operation. But the full space of potential
assignments and states of a circuit is exponential in the number of inputs and
state elements. Therefore the search space cannot be exhausted by simulation.
Even emulation using prototypical hardware is not sufficient to completely cover
the full state space.
141
Edit/add property
Property checker
fails
Analyze candidates
Autom. debugger
holds
Coverage analyzer
Analyze scenario
<100%
100%
DONE
Fig. 1. Enhanced flow
Thus, more powerful techniques for formal verification have been developed.
In particular, to prove the correctness of a hardware design with respect to a
given textual specification property checking has been developed. Formal properties are derived from the textual specification. The properties are proven fully
automatically to hold on the design. The underlying Boolean reasoning engines
are quite effective. Even for very large designs meaningful properties are handled
effectively. Still the verification gap remains due to low productivity, and intensive training of verification engineers is required to apply property checking in
practice. Therefore besides powerful reasoning engines, tool support is necessary
for several tasks during formal verification.
Here we propose an enhanced verification flow enriched by techniques to
ease the verification task. The flow is based on previous results by the authors.
Figure 1 illustrates this flow. Dark boxes denote automatic tools while light
boxes require manual interaction.
Typically only a property checker is available that returns a counterexample if
a property fails. The subsequent debugging task is only supported by standard
simulators. But techniques automating debugging in the context of property
checking have been presented [FSBD08] to speed up the work flow. The debugger
uses multiple counterexamples to determine candidate sites for the bug location
and thus decreases the amount of manual interaction.
Another major problem in property checking is to decide when a property
set is complete. This is usually done by manual inspection of all properties –
a threat to correctness for any larger design. The work in [GKD08] automates
this step. The coverage analyzer determines whether the properties describe the
behavior of the design under all possible input stimuli. If some input sequences
are not covered, this scenario is returned to the verification engineer for manual
inspection. Additional properties may be required or existing properties have to
be modified.
This work is structured as follows: The next section provides preliminaries
on Boolean reasoning engines. Section 3 explains property checking as consid-
142
ered here. The automatic debugging approach is briefly discussed in Section 4.
Section 5 explains the approach to automatically analyze functional coverage.
An embedded example is used to illustrate the techniques. Section 6 concludes.
2
Boolean Reasoning
Since the introduction of model checking there has been large interest in robust
and scalable approaches for formal verification. Symbolic model checking based
on Binary Decision Diagrams [McM93] has greatly improved scalability in comparison to explicit state enumeration techniques. However, these methods are
impractical for industrial designs.
Due to dramatic advances of the algorithms for solving Boolean Satisfiability
(SAT) many SAT-based verification approaches have emerged. Today, SAT is
the workhorse for Boolean reasoning and is very successful in industrial practice
[GG07]. Hence, in the following a brief introduction to SAT is provided as SAT
is also the basis for BMC.
The SAT problem is defined as follows: Let h be a Boolean function in Conjunctive Normal Form (CNF), i.e. a product-of-sums representation. Then, the
SAT problem is to decide whether an assignment for the variables of h exists
such that h evaluates to 1 or to prove that no such assignment exists.
The CNF consists of a conjunction of clauses. A clause is a disjunction of
literals and each literal is a propositional variable or its negation. SAT is one of
the central N P-complete problems. In fact, it was the first known N P-complete
problem that was proven by Cook in 1971 [Coo71]. Despite this proven complexity today there are SAT algorithms which solve many practical problem
instances, i.e. a SAT instance can consist of hundreds of thousands of variables,
millions of clauses, and tens of millions of literals.
For SAT solving several (backtracking) algorithms have been proposed
[DP60,DLL62,MS99,MMZ+ 01,ES04]. The basic search procedure to find a satisfying assignment is shown in Fig. 2 and follows the structure of the DPLL
algorithm [DP60,DLL62]. Instead of simply traversing the complete space of assignments, intelligent decision heuristics [GN02], conflict based learning [MS99],
and sophisticated engineering of the implication algorithm by Boolean Constraint
Propagation (BCP) [MMZ+ 01] lead to an effective search procedure. The description follows the implementation of the procedure in modern SAT solvers. While
there are free variables left (a), a decision is made (c) to assign a value to one
of these variables. Then, implications are determined due to the last assignment
by BCP (d). This may cause a conflict (e) that is analyzed. If the conflict can be
resolved by undoing assignments from previous decisions, backtracking is done
(f). Otherwise, the instance is unsatisfiable (g). If no further decision can be
done, i.e. a value is assigned to all variables and this assignment did not cause
a conflict, the CNF is satisfied (b).
To apply SAT for solving CAD problems an efficient translation of a circuit
into CNF is necessary. The principle transformation in the context of Boolean
formulas has been proposed by Tseitin [Tse68]. The Tseitin transformation can
143
(a) free var. left?
yes
no
(b) SAT
(c) decision
(d) propagation
no
(e) conflict ?
yes
ok
(f) resolve conflict
failed
(g) UNSAT
Fig. 2. DPLL algorithm in modern SAT solvers
i0
...
i1
ic
...
s0
o0
o1
oc
p
Fig. 3. Unrolled circuit and property
be done in linear time. This is achieved by introducing a new variable for each
sub-formula and constraining that this new variable is equivalent to the subformula. For circuits the respective transformation has been presented in [Lar92].
3
Formal Hardware Verification using Bounded Model
Checking
In the following the basic principles of Bounded Model Checking (BMC) are
provided.
We use BMC [BCCZ99] in the form of interval property checking as described
e.g. in [WTSF04,NTW+ 08]. Thus, a property is only defined over a finite time
interval. For a circuit with its transition relation Tδ , a BMC instance for a
144
property p over the finite interval [0, c] is given by:
c−1
Tδ (si , si+1 ) ∧ ¬ p ,
i=0
where p may depend on the inputs, states and outputs of the circuit in the time
interval [0, c]. This verification problem can be formulated as a SAT problem
by unrolling the circuit for c time frames and generating logic for the property.
The transformation is illustrated in Figure 3. As the property is negated in the
formulation, a satisfying assignment corresponds to a case where the property
fails – a counterexample.
In contrast to the original BMC as proposed in [BCCZ99] interval property
checking does not restrict the state s0 in the first time frame during the proof.
This may lead to false negatives, i.e. counter-examples that start from an unreachable state. In such a case these states are excluded by adding additional
assumptions to the property. But, for BMC as used here, it is not necessary to
determine the diameter of the underlying sequential circuit. Thus, if the SAT
instance is unsatisfiable, the property holds.
In the following we assume that each property is an implication of the form
always(A → C). A is the antecedent and C is the consequent of the property
and both consist of a timed expression. A timed expression is formulated on
top of variables that are evaluated at different points in time within the time
interval [0, c] of the property. The operators in a timed expression are the typical
HDL operators like logic, arithmetic and relational operators. The timing is
expressed using the operators next and prev. An example circuit given in Verilog
and properties specified in the Property Specification Language (PSL) [Acc] are
given in the following.
Example 1. Figure 4 describes the circuit sreg with the reset input and a clock
signal, the data input in and the control input ctrl. The input data from input
in is stored in the internal register s0 at each clock tick (lines 7–9). The control
input ctrl decides whether output out returns the Boolean AND or the Boolean
OR of the current value of in and the previously stored value s0 (lines 11-19).
The simple PSL property pReset in Figure 5 describes the behavior of this
circuit immediately after a reset has been triggered. The property is defined over
the interval [0, 1]. The antecedent consists of a single assumption saying that the
reset is triggered at time step 0. The consequent specifies that the output is 0
in the next time step under this condition.
4
Debugging
As explained above debugging is a manual task in the standard design flow. Tool
automation helps to improve the productivity. An automatic approach for debugging in the context of equivalence checking has been proposed in [SVFAA05]
and extended to property checking in [FSBD08].
145
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
module s r e g ;
input c l k , r s t
input in , c t r l ;
output out ;
reg s0 , out ;
always @( posedge c l k )
i f ( r s t ) s 0 <= 0 ;
else
s 0 <= i n ;
always @( posedge c l k )
i f ( r s t ) out <= 0 ;
else
// i n c o r r e c t l i n e :
// i f ( ! c t r l )
if ( ctrl )
out <= s 0 | | i n ;
else
out <= s 0 && i n ;
endmodule
1
2
3
4
5
property pReset = always (
r s t == 1
) −> (
next [ 1 ] ( out ) == 0
);
Fig. 5. Reset property
1
2
3
4
5
6
7
8
9
property pOperation = always (
next a [ 0 . . 1 ] ( r s t == 0 )
) −> (
next [ 1 ] ( c t r l ==0) ?
next [ 2 ] ( out )==
( i n && next [ 1 ] ( i n ) )
: next [ 2 ] ( out )==
( i n | | next [ 1 ] ( i n ) )
);
Fig. 6. Property for normal operation
Fig. 4. Example circuit
Essentially, the same model is created as for BMC shown in Figure 3. Additionally, for the failing property one or more counterexamples are given. The
primary inputs of the circuit are restricted to this counterexample. While the
property is restricted to hold. This forms a contradictory problem instance: when
a counterexample is applied to the circuit, the property does not hold. Finding
a cause for this contradiction yields a potential bug location, a so called fault
candidate.
A fault candidate is a component in the circuit that can be replaced to fulfill
the property. Here, a component may be a gate, a module in the hierarchical
description, or an expression in the source code.
To determine such fault candidates, each component of the circuit is replaced
by the logic shown in Figure 7. In the circuit component c implements the
function fc . This signal line is modified to calculate fc =!rc → fc , where rc is a
new primary input to the model. This allows to change the output value of c.
When rc is zero, c simply calculates the original output as given by the circuit
description. When rc is one, the circuit can be repaired by injecting a new output
value for c.
A trivial solution at this stage would modify all components of the circuit
at the same time and by this easily fulfill the attached property. Therefore an
additional constraint limits the number of modifications. First a solution with
only one modification is searched, if no such solution exists more modifications
are iteratively allowed until a first solution has been found. For example, if more
146
C
Rdd
R
rc
C
fc
f’c
Fig. 7. Repairing a component
than one bug is present, often multiple modifications are required to fix the
circuit. Then, all modifications that allow to fulfill the property are determined to
retrieve all fault candidates. Finding the ”real” bug among the fault candidates
is left to the verification engineer.
Some improvements have been proposed to improve accuracy or efficiency of
this simple approach [FSBD08]. Improving the accuracy can be achieved by using multiple counterexamples. The same construction as described above is done
for all counterexamples. The same variables rc are reused with respect to all
counterexamples. Thus, the same components are modified to correct all counterexamples at the same time. Alternatively, the specification may be strengthened to improve the accuracy. By using multiple properties to specify correct
behavior, the acceptable behavior is described more accurately. Therefore, false
repairs are excluded. Finally, so called Ackermann constraints force all modifications to be realizable by combinational circuitry. The approach considered so far
allows components to behave non-deterministic for repair, which is not feasible
in practice. Ackermann constraints can be used to remove these infeasible fault
candidates.
Efficiency can be improved, by incrementally using more and more counterexamples or more and more properties. Simulation-based preprocessing can
help to remove some fault candidates in case of single faults.
Further works show how to improve the efficiency [SLM+ 07,SFBD08], exploit hierarchical knowledge [FASV+ 05], apply abstraction [SV07], fully correct
a circuit with respect to a given specification [CMB07] or generate better counterexamples [SFB+ 09].
Example 2. Assume that the control signal ctrl of the Verilog circuit was interpreted wrongly. Instead of the correct line 16 in Figure 4, line 15 was used.
In this case property pOperation in Figure 6 does not hold on the circuit. One
counterexample may set ctrl to zero, so the then-branch of the if-statement is
executed erroneously. The resulting output may be corrected by changing either
line 17, where the operation is carried out, or line 16, the faulty condition. These
two locations are returned as fault candidates by the approach. When adding
another counterexample that sets ctrl to one, the else-branch is erroneously executed. Lines 16 or 19 are fault locations. Thus, only line 16 remains as a common
147
fault candidate when both of the counterexamples are applied during automatic
debugging.
5
Coverage Analysis
After debugging and finally proving a set of properties, the verification engineer
wants to know if the property set describes the complete functional behavior of
the circuit. Thus, in the standard design flow the properties are manually reviewed and the verification engineer checks whether properties have been specified for each output (and important internal signals) which prove the expected
behavior in all possible scenarios. The coverage analysis approach introduced in
[GKD07,GKD08] automatically detects scenarios – assignments to inputs and
states – where none of the properties specify the value of the considered output.
The general idea of the coverage approach is based on the generation of a
coverage property for each output. If this coverage property holds, the value of
the output o is specified by at least one property in any scenario. Essentially, the
method shows that the union of all properties that involve the output o does not
admit behavior else than the one defined by the circuit. For this task a multiplexor construct is inserted into the circuit for the actual output o, the properties
describing the behavior of o are aligned and finally the coverage property for o is
generated. Figure 8 depicts the final result of the multiplexor insertion which is
carried out before unrolling. As can be seen for each bit of the n-bit output o a
multiplexor is inserted. Each multiplexor is driven by the respective output bit
and the inverted output bit. Then, in a renaming step the output of each multiplexor becomes oi and the input o origi , respectively. Moreover, the predicate
sel computes the conjunction of all the select inputs of the multiplexers. Based
on this construction, the coverage check can be performed by proving that the
multiplexor is forced to select the original value of o (i.e. o orig) at the maximum
time point, assuming all properties involved. This is expressed in the generated
coverage property of the considered output:
⎛
⎞
|Po |
tmax
−1
⎝
p̂j ∧
Xt sel = 1⎠ → Xtmax sel = 1,
j=1
t=0
where |Po | is the set of properties which involve o in the consequent, tmax is the
maximum time point of time intervals defined by the properties in |Po |, p̂j are
the adjusted properties over time, and Xk denotes the application of the next
operator for k times. By this, the coverage problem is transformed into a BMC
problem.
Complete coverage in terms of the approach is achieved by considering all
outputs of a circuit. If all outputs are successfully proven to be covered by the
properties, then the functional behavior of the circuit is fully specified.
Further works consider to aid the verification engineer while formulating
properties with the goal to achieve full coverage. In [GWKD08] an approach to
understand the reasons for contradictions in the antecedent of a property has
148
sel0
0
1
0
o0
...
o_orig
sel
sel1
o_orig
1
1
0
o1
...
Fig. 8. Insertion of the multiplexor
1
2
3
4
5
6
p r o p e r t y pNew = always (
( r s t == 1 ) && next [ 1 ] ( r s t == 0 )
) −> (
next [ 1 ] ( c t r l ==0) ? next [ 2 ] ( out ) == 0
: next [ 2 ] ( out ) == next [ 1 ] ( i n )
);
Fig. 9. Property required for full coverage
been proposed. The technique of [KGD09] removes redundant assumptions in
a property and generates valid properties for a given specified behavior. Both
techniques can be combined with the coverage analysis to reduce the number of
iterations to obtain full coverage.
Example 3. Again consider the example circuit and the properties pReset and
pOperation. Property pReset covers the behavior after the reset while property
pOperation describes the normal operation of the circuit. Thus, full coverage
should be reached and both properties are passed to the coverage analyzer.
The algorithm unexpectedly returns the uncovered scenario where rst is one
in the first time step and zero in the next. Indeed none of the properties covers
this case, because pOperation assumes rst to be zero in two consecutive time
steps. Thus, a new property pNew to cover the remaining scenario is formulated
as shown in Figure 9. Applying the coverage analysis to the three properties
yields 100% coverage.
6
Conclusions
In this paper we have presented an enhanced formal verification flow. For formal
verification the flow uses bounded model checking. The two major improvements
in the new flow are the integration of debugging and coverage analysis. Both techniques automate manual tasks and hence the productivity improves significantly
in comparison to a traditional flow.
149
References
[Acc]
Accellera Property Specification Language Reference Manual, version 1.1.
http://www.pslsugar.org, 2005.
[BCCZ99] A. Biere, A. Cimatti, E. Clarke, and Y. Zhu. Symbolic model checking
without BDDs. In Tools and Algorithms for the Construction and Analysis
of Systems, volume 1579 of LNCS, pages 193–207. Springer Verlag, 1999.
[CMB07] K. Chang, I.L. Markov, and V. Bertacco. Fixing design errors with counterexamples and resynthesis. In ASP Design Automation Conf., pages 944–
949, 2007.
[Coo71]
S.A. Cook. The complexity of theorem proving procedures. In 3. ACM
Symposium on Theory of Computing, pages 151–158, 1971.
[DLL62]
M. Davis, G. Logeman, and D. Loveland. A machine program for theorem
proving. Comm. of the ACM, 5:394–397, 1962.
[DP60]
M. Davis and H. Putnam. A computing procedure for quantification theory.
Journal of the ACM, 7:506–521, 1960.
[ES04]
N. Eén and N. Sörensson. An extensible SAT solver. In SAT 2003, volume
2919 of LNCS, pages 502–518, 2004.
[FASV+ 05] M. Fahim Ali, S. Safarpour, A. Veneris, M.S. Abadir, and R. Drechsler.
Post-verification debugging of hierarchical designs. In Int’l Conf. on CAD,
pages 871–876, 2005.
[FSBD08] G. Fey, S. Staber, R. Bloem, and R. Drechsler. Automatic fault localization
for property checking. IEEE Trans. on CAD, 27(6):1138–1149, 2008.
[GG07]
M. Ganai and A. Gupta. SAT-Based Scalable Formal Verification Solutions
(Series on Integrated Circuits and Systems). Springer, 2007.
[GKD07] D. Große, U. Kühne, and R. Drechsler. Estimating functional coverage in
bounded model checking. In Design, Automation and Test in Europe, pages
1176–1181, 2007.
[GKD08] D. Große, U. Kühne, and R. Drechsler. Analyzing functional coverage in
bounded model checking. IEEE Trans. on CAD, 27(7):1305–1314, 2008.
[GN02]
E. Goldberg and Y. Novikov. BerkMin: a fast and robust SAT-solver. In
Design, Automation and Test in Europe, pages 142–149, 2002.
[GWKD08] D. Große, R. Wille, U. Kühne, and R. Drechsler. Using contradiction analysis for antecedent debugging in bounded model checking. In
ITG/GI/GMM-Workshop “Methoden und Beschreibungssprachen zur Modellierung und Verifikation von Schaltungen und Systemen”, pages 169–178,
2008.
[KGD09] U. Kühne, D. Große, and R. Drechsler. Property analysis and design understanding in a quality-driven bounded model checking flow. In Design,
Automation and Test in Europe, 2009.
[Lar92]
T. Larrabee. Test pattern generation using Boolean satisfiability. IEEE
Trans. on CAD, 11:4–15, 1992.
[McM93] K.L. McMillan. Symbolic Model Checking. Kluwer Academic Publisher,
1993.
[MMZ+ 01] M.W. Moskewicz, C.F. Madigan, Y. Zhao, L. Zhang, and S. Malik. Chaff:
Engineering an efficient SAT solver. In Design Automation Conf., pages
530–535, 2001.
[MS99]
J.P. Marques-Silva and K.A. Sakallah. GRASP: A search algorithm for
propositional satisfiability. IEEE Trans. on Comp., 48(5):506–521, 1999.
150
[NTW+ 08] M. D. Nguyen, M. Thalmaier, M. Wedler, J. Bormann, D. Stoffel, and
W. Kunz. Unbounded protocol compliance verification using interval property checking with invariants. IEEE Trans. on CAD, 27(11):2068–2082,
2008.
[SFB+ 09] A. Sülflow, G. Fey, C. Braunstein, Ulrich Kühne, and R. Drechsler. Increasing the accuracy of sat-based debugging. In Design, Automation and
Test in Europe, 2009. to appear.
[SFBD08] A. Sülflow, G. Fey, R. Bloem, and R. Drechsler. Using unsatisfiable cores
to debug multiple design errors. In Great Lakes Symp. VLSI, pages 77–82,
2008.
[SLM+ 07] S. Safarpour, M. Liffton, H. Mangassarian, A. Veneris, and K. A. Sakallah.
Improved design debugging using maximum satisfiability. In Int’l Conf. on
Formal Methods in CAD, pages 13–19, 2007.
[SV07]
S. Safarpour and A. Veneris. Abstraction and refinement techniques in
automated design debugging. In Design, Automation and Test in Europe,
pages 1182–1187, 2007.
[SVFAA05] A. Smith, A. Veneris, M. Fahim Ali, and A.Viglas. Fault diagnosis
and logic debugging using boolean satisfiability. IEEE Trans. on CAD,
24(10):1606–1621, 2005.
[Tse68]
G. Tseitin. On the complexity of derivation in propositional calculus. In
Studies in Constructive Mathematics and Mathematical Logic, Part 2, pages
115–125, 1968. (Reprinted in: J. Siekmann, G. Wrightson (Ed.), Automation of Reasoning, Vol. 2, Springer, Berlin, 1983, pp. 466-483.).
[WTSF04] K. Winkelmann, H.-J. Trylus, D. Stoffel, and G. Fey. Cost-efficient block
verification for a UMTS up-link chip-rate coprocessor. In Design, Automation and Test in Europe, volume 1, pages 162–167, 2004.
Daniel Große received the Diploma degree in computer science from the Albert-Ludwigs-University,
Freiburg im Breisgau, Germany in 2002 and the Dr.
degree from the University of Bremen, Bremen, Germany, in 2008. He is currently with the research group
of computer architecture at the University of Bremen.
His research interests include verification, logic synthesis, and high-level languages like SystemC.
Görschwin Fey received his Diploma degree in computer science from the Martin-Luther Universität,
Halle-Wittenberg, Germany in 2001. Since then he has
been with the research group of computer architecture
at the University of Bremen, where he received the Dr.
degree in 2006. He was a guest associate professor at
the University of Tokyo from Dec. 2007 to June 2008.
Currently, he is a postdoctoral researcher at the University of Bremen. His research interests are in testing
and formal verification of circuits and systems. In particular he focuses on debugging and reliability.
151
Rolf Drechsler received the Diploma and Dr. Phil.
Nat. degrees in computer science from the J.W.
Goethe-University, Frankfurt am Main, Germany, in
1992 and 1995, respectively. He was with the Institute of Computer Science at the Albert-LudwigsUniversity, Freiburg im Breisgau, Germany, from 1995
to 2000. He joined the Corporate Technology Department of Siemens AG, Munich, in 2000, where he
worked as a Senior Engineer in the Formal Verification Group. Since October 2001, he has been with
the University of Bremen, Bremen, Germany, where
he is now a Full Professor for Computer Architecture.
He has authored more than 10 books in the area of
VLSI CAD and he published more than 250 papers in
archival journals and refereed conference proceedings.
He was the recipient of Best Paper Awards at the 2006
Haifa Verification Conference and the 2007 Forum on
Specification and Design Languages. He served on the
program committee of major conferences in Electronic
Design Automation (EDA), including Design Automation Conference (DAC), Asia and South Pacific Design
Automation Conference (ASP-DAC), and Design Automation and Test Europe (DATE). His research interests include verification, testing and synthesis.
Program Graph Transformation
Berthold Hoffmann
Universität Bremen
[email protected]
Abstract. Graph transformation, a branch of theoretical computer science, is about the definition of graph languages by grammars, and the
study of computations on graphs by rewrite rules. In this paper, we sketch
a software engineering problem – the refactoring of object-oriented software – and indicate how graph grammars and graph rewrite rules can
be extended for specifying and implementing refactoring operations on
a high level of abstraction.
1
Introduction
Graph transformation generalizes two areas of theoretical computer science.
Chomsky grammars and Lindenmeyer systems [28, chap. 4, 5] generate sets,
or formal languages, of words. Graph grammars allow more general languages
to be generated: data structures like double-linked lists, the admissible states
of some system, or the growth and habitus of plants [26]. Term rewriting and
narrowing define rule-based computational models on terms, or trees [1]. With
graph rewriting, it is possible to study rule-based computations on more general
structures: term graph rewriting [25], or petri nets [12], to mention just a few.
Graphs are often drawn as diagrams, and diagrams are abstractly represented
as graphs. This makes graph transformation a natural choice for specifying and
implementing systems that generate and modify diagrams. Diagram languages
like Uml are now widely used for the model-driven design of object-oriented software. Graph transformation has a high potential for specifying and implementing
visual software models and their transformation, witnessed by a series of international workshops on graph transformation and visual modeling techniques [16]. In
this paper, we use a particular scenario from this area to illustrate the potential
of graph transformation.
Refactoring of Object-Oriented Code. Refactoring aims at improving the
structure of object-oriented software for maintenence and adaptation, without
changing its behavior. Martin Fowler’s book [15] gives a catalog of basic refactoring operations, which, skillfully combined, shall attain these goals. A typical
operation in this catalog, Pull-up Method, is summarized as follows:
– All subclasses of a class c contain equivalent implementations of a method m.
– Pull one of these implementations up to c, and remove all other implementations of m in the subclasses of c.
Program Graph Transformation
153
An informal description of the “mechanics” then prescribes a series of steps by
which a programmer should interactively change the code, compile, and test it,
in order to make sure that the consistency of the code, and the behavior of
the program is preserved. Nowadays, many case tools do support refactoring.
However, these implementation have no formal basis so that is remains unclear
whether the application of a refactoring will always preserves code consistency,
let alone program behavior.
The extensive case study [20, 21] by D. Janssens, T. Mens et al. has shown
that graph transformation has a high potential for formalizing refactoring operations. Many advantages are equally true for software models and model transformation in general.
Graph grammars allow languages of graphs and diagrams to be described
in a general way. In Section 2 we show how the shape of program graphs, a
language-independent representation of object-oriented code, can be be defined
with adaptive star grammars. Refactoring operations can be defined by graph
rewriting rules. In Section 3, the Pull-up Method refactoring is specified by a
single generic graph rewrite rule. These rules allow to specify expressive patterns,
including sub-patterns of variable number and shape. In Section 4, we study
how it can be checked whether rewriting steps preserve the shape of graphs. In
particular, we discuss under which conditions generic rules do preserve shape
automatically. Only if all these concepts are supported by an efficient tool, they
can be used for developing refactoring tools. In Section 5 we outline how such a
tool could be developed. We conclude with some remarks on related and future
work in Section 6.
2
Models and Graph Grammars
We describe program graphs informally, discuss why their properties cannot be
captured easily by meta-models like class diagrams, and define them precisely,
by adaptive star grammars with application conditions.
Program Graphs. The case study [20] has proposed a language-independent
representation of object-oriented programs as program graphs, which can be used
for refactoring code written in many languages. In a program graph, nodes represent syntactic entities, and edges establish relations between entities. Relations
represent a hierarchical structure representing abstract syntax, and context links
between uses and definitions of variables and methods that result from the analysis of declarations. Thus the names of variables or methods are not relevant in
this representation.
Fig. 1 below shows two program graphs G and H. (We prefer the classical
way to draw graphs. A denotation as Uml object diagrams would be less concise.) In G, the subgraph P (underlaid in light grey) represents a class with a
method signature and two sub-classes (the nodes labeled C and M). Every subclass contains bodies for the method (the subgraphs underlaid in darker grey),
containing body, expression, and variable nodes, which are labeled B, E, and V.
154
Berthold Hoffmann
1
G
2
3
4
C
M
B
V
1
H
C
C
2
C
3
C
4
M
P
5
C
R
5
=⇒
B
V
B
pum
E
6
V
E
E
E
E
M
E
E
6
V
7
E
M
7
Fig. 1. Two program graphs
In H, one of the method bodies in P has been deleted, and the other one pulled
up to the superclass (by altering the source of the red edge). Actually, this is
what a Pull-up Method refactoring should do to any pattern P of a program
graph G.
In model-driven design, the shape of models is usually specified by metamodeling. In Uml, for instance, a class diagram specifies the shape of the object
diagrams that can be generated by a system. A class diagram usually specifies
subclass relations on nodes, incidence constraints for edges (i.e., which type of
node may be the source or target of some type of edge), and cardinality constraints (i.e., how many edges of some type may be the source or target of a
certain type of node). More involved constraints concerning, e.g., the existence
of paths in a graph, can be specified by logical formulas, as in the Uml’s sublanguage Ocl.
Program graphs are trees representing the nesting of the abstract syntax (by
solid arrows), with contextual links from the uses of names to their declaration
(by dashed arrows). A class diagram can capture the syntactic structure by
subclasses. But, the definition-to-use links of program trees can only be specified
by Ocl formulas.
Therefore we propose to use graph grammars for specifying program graphs,
rather than meta-models. A graph grammar generates a language of graphs (the
valid models) by inductive rules.
Star Grammars. We start from a class G of graphs that are labeled and
directed, and may have loops and parallel edges (with the same label, connecting
the same nodes).
Star grammars are based on variable substitution. In a graph, a variable is a
star-like subgraph X with a center node that is labeled with a variable symbol,
and has edges connecting the center node to border nodes that are labeled with
terminal symbols. The graph in Figure 7 is a star.
A star rule is written Y ::= S, where Y is a variable and S a graph so that
Y and S share the border nodes of Y . Figure 2 shows a star rule for the body of
a method. The box annotated e that surrounds the variable named Exp (with
155
Program Graph Transformation
center node x and border node v) indicates a multiple subgraph: this box may
have n > 0 clones, where each of them is connected to the rest of the graph in
the same way (by an edge from the B-node to each clone of the node v in this
case). Multiple subgraphs transfer the notations for repetitions in the EBNF of
context-free grammars to graphs.
We draw star rules by “blowing up” the center node of Y and drawing the
inserted nodes and edges of S inside, as can be seen in Figure 4. (The second
rule in the lower row of is that of Figure 2.).
A variable X in a graph G can be substituted with a star rule Y ::= S
if there is an isomorphism m : Y → X. Then a variable substitution yields the
graph denoted as G[X/m S], which is constructed from the disjoint union of G and
S by identifying the border nodes of X and Y according to m and removing their
edges and center nodes. Let Σ be a finite set of star rules. Then we write G ⇒Σ H
if H = G[X/m S] for some y ::= S ∈ Σ, and denote the reflexive-transitive closure
of this relation by ⇒∗Σ . Let Y be the set of all variables occurring in Σ. We call
them syntactic variables, like in Chomsky grammars, and denote the graphs with
syntactical variables by G(Y). Then Γ = G(Y), Y, Σ, Z is a star grammar with
a start variable Z ∈ Y. The language of Γ is obtained by exhaustive substitution
of its rules to the start variable:
L(Γ ) = {G ∈ G | Z ⇒∗Σ G}
Figure 4 shows the star rules generating the program trees that underlie program
graphs, and Figure 3 shows the generation of a method body tree.
Adaptive Star Grammars. The dashed edges in the rules use, assign, and call
point to new nodes generated by these rules. In program graphs, dashed edges
are contextual links that shall point to a visible declaration of an attribute,
a variable, or a method, respectively. The contextual links of program graphs
cannot be generated with star grammars. Because, in a finite star grammar, the
number of border nodes in stars is bounded. For the insertion of a contextual
link, say to a method in a call rule, all potential targets of the link had to be
border nodes of the left-hand side of the call rule. Since the number of visible
B
Bdy
B
B
B
B
B
Bdy
E
⇓
v
E
Exp
⇒
B
e
Exp
x
E
E
Exp
Exp
E
E
E
Exp
⇒
Exp
E
⇒
E
E
E
E
⇒
E
E
E
E
V
V
Exp
bdy
Fig. 2. A star rule
V
V
V
V
V
V
M
V
Fig. 3. Variable substitutions generating a method body
M
156
Berthold Hoffmann
C
C
Prg
A
C
Hy
Fea
Cls
C
start
M
Sig
B
Hy
hy
Sig
Bdy
Fea
Fea
att
cls
method
E
E
B
E
Exp
Exp
Bdy
B
F
Cls
Hy
M
Exp
E
E
E
A
Exp
sig
Exp
V
A
use
bdy
M
assign
Exp
call
Fig. 4. Star rules generating program trees
methods in a program graph is not bounded, but grows with the size of the
graph, this is not possible.
This observation, however, gives a clue how this limitation can be overcome.
We make stars, star rules, and variable substitutions adaptive wrt. the numbers
of border nodes in stars, as proposed in [7]. In Figure 5, the star rule bdy on the
left is adaptive: The node d is a multiple node (which is drawn with a shade, like
the multiple subgraph in Figure 2), representing a set of clone nodes. The clones
of d are the declarations visible in the body1 A clone of a graph X is denoted
as X vk , where v is a multiple node or subgraph, and k the number of its clones.
On the right, a schematic view of the clones bdy ke nd of the adaptive star rule bdy
is given, for k > 0, n 0. When all multiple nodes and subgraphs in a rule are
cloned, the result is an ordinary star rule, but every adaptive rule may have an
infinite number of clones.
B
B
Bdy
Bdy
e
E
Exp
v
x
bdy
v1
x1
bdy ke
D
d
E
E
···
Exp · · · Exp
··· ···
d
n
d1
···
D
···
···
D
vk
xk
dn
Fig. 5. An adaptive star rule for bodies with clones
1
The type D has subtypes K, V, and M representing declarations of constants, variables, and methods. We silently assume that types in rules may be replaced by
subtypes.
Program Graph Transformation
157
Let Σ̈ denote the sets of all possible clones a set Σ of adaptive star rules.
Then, an adaptive star grammar Γ = G(Y), Y, Σ, Z has the same components
as a star grammar, where the syntactic variables in Y and the rules in Σ are
adaptive, but the start variable Z is not. The adaptive star grammar Γ generates
the language
L̈(Γ ) = {G ∈ G | Z ⇒∗Σ̈ G}
The set of cloned star rules Σ̈ is infinite. It has been shown in [7] that this
gives adaptive star grammars greater generative power than grammars based on
hyperedge [17] or node replacement [13], but are still parseable. The adaptive
C
C
A
g
p
D
A
e
e
Hy
i
C
g
F
e
Hy
Prg
C
D
M
p
h
Cls
M
A
p
d
Cls
Sig
p
e
B
Hy
A
Fea
i
sig
hy
cls
D
D
D
start
B
A
M
A
B
B
Bdy
p
Fea
Fea
Fea
E
i
Sig
Bdy
Bdy
h
att
method
override
D
D
h
M
A
Exp
i
p
A
bdy
D
D
E
E
E
Exp
Exp
Exp
E
E
Exp
use
d
A D
assign
d
V
Exp
call
D
p
A
M
A
D
Fig. 6. The set PG of adaptive star rules generating program graphs
rules in Figure 6 generate all program graphs. Since space does not permit to
explain all rules, we refer the reader to [8] for a thorough explanation. Roughly,
the rules start, hy. and cls generate the class hierarchy of a program, collecting
the declarations contained in the classes, and distributing them according to the
visibility rules: If a declaration is global, it is visible in the entire hierarchy; if
it is heritable, it is visible in the class and all its subclasses, if it is internal, it
158
Berthold Hoffmann
is only visible inside the class itself. The features of a class are either attributes
(variables or constants), or new methods, or bodies overriding some inherited
method. Expressions just represent data flow: the use of some visible attribute, or
the assignment of the value of an expression to a variable, or the call of a method
with expressions as actual parameters. The full grammar for program graphs,
with control flow in method bodies, typing, and more sophisticated visibility
rules, appeared in Niels Van Eetvelde’s PhD thesis [31].
Application Conditions. Adaptive star rules do not suffice to capture all
properties of program graphs. In a call, e.g., the number of actual parameters
must match the formal parameters of the method being called. If a program contained m 0 methods, this would require to collect m different sets F1 , . . . , Fm
of formal parameters in distinct multiple border nodes of a variable like Exp.
However, the number m is not bounded, whereas there are only finitely many
edge types. So adaptive stars do not suffice to distinguish all parameter sets
F1 , . . . , Fm .
Our solution is to extend adaptive star rules with positive and negative application conditions as in [11]. A conditional adaptive star rule has the form
A, N → Y ::= S where the graphs A, N ∈ G may share nodes with Y . For
some clone Ä, N̈ → Ÿ ::= S̈ of this rule, a variable substitution G[X/m S̈] is
only defined if the isomorphism m : Ÿ → X can be extended to the positive
application condition Ä, and cannot be extended to the negative application
condition N̈ .
In PG, application conditions are used as follows: The syntactic variables are
attached, not only to all visible declarations, but also to all formal parameters
of all methods. In rule call, the edge from M to A underlaid in green is a positive
condition, collecting formal parameters of M, and the edge from M to the other Anode underlaid in red is a negative condition, forbidding the existence of further
formal parameters of M. Then the multiple subgraph makes sure that an actual
parameter will be generated for every formal parameter of the method being
called.2
The adaptive star grammar in Figure 6 is harder, both to develop, and to
understand, than a class diagram for program graph.
However, the definition of a graph language by adaptive star grammars is
inductive. For instance, the rule bdy ke nd in Figure 5 expresses the following inductive property: If E1 , . . . , Ek are expression graphs, where every graph Ei has
a root vi , and uses the same set {d1 , . . . , dn } of declarations, adding a B-node
as a root, with edges to all nodes v1 , . . . , vk yields a body graph using the same
declarations. This allows to prove properties of adaptive star languages by induction over the rules, and to define operations on these languages by structural
recursion over the rules.
Adaptive star rules have further properties that will be useful in the next
section.
2
In rule override, a similar application condition makes sure that all formal parameters
of a method are visible as variables within its body.
Program Graph Transformation
159
Lemma 1 (Commutativity and Substitutivity [7]). Consider two consecutive variable substitutions G[X/m S][X /m S ]. Then the following holds:
1. Commutativity: If X is in G, then G[X/m S][X /m S ] = G[X /m S ][X/m S].
2. Substitutivity: If X is in R, then G[X/m S][X /m S ] = G[X/mS[X /m S ]].
Commutativity and substitutivity are typical for context-free grammars on words
or graphs, as discussed by B. Courcelle in [5]. However, adaptive star grammars
are not context-free, because they fail to be universal : this would mean that
every variable X in a graph may be substituted by every rule Y ::= S if the
variable symbols at their center nodes are the same. Obviously, this is not always
the case for adaptive star rules. For instance, consider a variable X with center
labeled Exp and a single E-labeled border node, i.e., no declaration node is
“visible” to X. Then X (like all graphs containing X) is a blind alley. None of
the rules use, assign, call applies to X, because each of these rules requires that
an attribute, a variable, or a method is visible to X.
3
Generic Graph Rewriting
When applying a graph rewrite rule to a host graph G, a match m(P ) of a graph
pattern P in G shall be identified in order to insert a replacement graph R for
it. The connection between G and R is established by letting P and R share an
interface at which the match m(P ) and R shall be glued.
Gluing Rules. A gluing rule consists of a pair of star rules Y ::= P , Y ::= R
using the same variable Y as their interface; it is denoted as t = P R : Y ,
and applies to a graph G if there exists a context graph C containing a star X
with an isomorphism m : Y → X so that G = C[X/m P ]. Then, the application
of t yields the graph H = C[X/m R]; such a transformation step is denoted as
G ⇒t H.3
In Figure 1, the grey parts of graphs G and H form pattern and replacement
of a gluing rule pum = P R : Y ; the interface Y is shown in Figure 7. When
drawing rules, we omit the interface, and indicate common border nodes of Y
in P and R by attaching the same numbers to them, as done in Figure 1. The
figure thus shows a rewrite step G ⇒pum H, for a schematic context as shown
in Figure 8.
Rule pum captures just some of the situations where the Pull-up Method
refactoring shall be performed. A general rule for Pull-up Method should apply
to a class with any number of subclasses (not just two), to methods with any
number of parameters (not just one), and to bodies of arbitrary shape. Therefore
we extend gluing rules with sub-patterns of arbitrary number or shape.
3
Rewriting with gluing rules corresponds to double pushout graph transformation [10]
with discrete interfaces and injective occurrences.
160
Berthold Hoffmann
Y
C
C
C
M
C
V
H1
2
C
4
V
C
M
C
V
H1
V
M
Fig. 7. An interface star
M
Fig. 8. A context graph
Cloning and Expansion. Adaptive star rules do already provide multiple
nodes and subgraphs that may have a variable number of clones. It is easy
to allow them in gluing rules as well. According to the other requirement, rule
pum should apply to method bodies of arbitrary shape, not just to the particular
bodies in the example (which are underlaid in dark grey).
We therefore allow that in a generic rewrite rule t : P R : Y , the graphs
P and R may not only contain multiple nodes and multiple subgraphs, but also
graph variables from a set X . Graph variable are expanded to varying subgraphs
before applying a rule.
1
1
C
2
pum =
C
3
4
C
M
C
B
V
B
5
c
B1
2
c
3
C
4
M
C
V
B
5
B2
6
D
B2
6
D
Fig. 9. The generic Pull-up Method rule
Figure 9 shows a generic rule pum for Pull-up Method. The rule contains
a multiple subgraph c, two multiple nodes 5 and 6, and two graph variables
B1 and B2 . In order to apply a generic rule, its multiple nodes and subgraphs
must be cloned, and its variables must be expanded by graphs. For simplicity, we
assume that cloning is done before expansion. (In practice, cloning and expansion
should be incremental, and interleaved. This is useful for implementing matching
of generic rules.)
Program Graph Transformation
1
1
C
2
¨ =
pum
C
3
4
C
M
C
B
V
B
2
B1
B2
V
M
3
C
5
6
161
4
M
C
V
B
5
B2
7
V
6
M
7
Fig. 10. A clone of the generic Pull-up Method rule
¨ in Figure 10 is the clone pum 1c 15 26 of the generic rule pum in
The rule pum
Figure 9.
A substitution σ is a set of star rules with mutually distinct graph variables
on their left-hand sides. Since σ actually defines a partial function from variables
to graphs, we write Dom(σ) for the set of the left-hand sides in σ, and σ(Y ) = S
for every Y ∈ Dom(σ) if Y ::= S is a star rule in σ. The expansion of a graph G
applies σ to each graph variable in G, and is denoted by Gσ . Due to Lemma 1.1,
variables can be expanded in any order, even in parallel.4
β=
8
>
>
>
>
>
>
>
B1
>
>
>
>
>
>
E
>
<
>
>
E
>
>
>
>
>
>
>
>
> b1
>
>
>
: V
B
9
>
>
>
>
>
>
>
>
>
>
>
>
>
>
=
B
B2
E
E
,
E
E
b2
V
M
V
V
V
M
V
>
>
>
>
>
>
>
>
>
>
>
>
>
>
;
Fig. 11. A shaped substitution for method bodies
¨ β of the clone
The gluing rule pum applied in Figure 1 is the expansion pum
¨ in Figure 10, where β is the substitution defined in Figure 11.
pum
A generic rule t : P R : Y applies to a host graph G = C[X/m P̈ σ ], where
ẗ : P̈ R̈ : Ÿ is a clone of t and σ is a substitution. The application of t yields
the graph H = C[X/m R̈σ ], and we write G =⇒ H.
t,σ
4
In this notation, we omit the morphisms from the variables of σ to those of G by
which the expansion becomes unique.
162
Berthold Hoffmann
Shaped Expansion. Expansions insert arbitrary graphs into a generic rule.
This is not only hard to implement, because the substitution has to be found
while applying the generic rule, but it is also too general. In our example, rule
pum shall be applied to program graphs. Then the graph variables B1 and B2
are placeholders for method bodies. So we require that substitutions are shaped
according to adaptive star rules.
Let Γ = G(Y), Y, Σ, Z be an adaptive star grammar with application conditions. We assume that every graph variable X ∈ X refers to a syntactic variable
X ∈ Y, which is called the shape of X. For some graph G ∈ G(X ), replacing
graph variables by the corresponding syntactic variables yields the shape graph
G ∈ G(Y). A graph G ∈ G(X ) is shaped according to Σ if Y ⇒∗Σ G for some
Y ∈ Y; this is written as Σ G : Y .
A substitution σ is shaped (according to Σ) if Σ σ(Y ) : Y for every graph
variable Y ∈ Dom(σ).
¨ have the shape Bdy. Then
Let the graph variables B1 and B2 in rule pum
the substitution β is shaped.
4
Shape Preservation
For our example, it is relevant to ask whether the refactoring of a program graph
will always yield a valid program graph again. A simple solution to this problem
would be to check this property after every rewrite step. Since the parsing for
adaptive star grammar is decidable, this can be done automatically, but is not
efficient. It would be better if the rules and the way they are applied could be
restricted so that shapes are preserved in any case.
A generic rule t = P R : Y is called shaped if Σ P : Y and Σ R :
Y. A shaped generic rule t = P R : Y applies to a graph G if there exists a
shaped context graph C containing a star X with an isomorphism m : Y → X so
that G = C[X/m P σ ], for some substitution σ. Then, the application of t yields
Σ
the graph H = C[X/m Rσ ]; a shaped rewrite step is denoted as G =⇒
t H.
Theorem 1 (Shape-Correctness). For any graph G and any shaped generic
rule t:
Σ
Σ G : Z ∧ G =⇒
T H =⇒ Σ H : Z
Proof Sketch. Assume w.l.o.g. that t = P R : Y and H = C[X/m R̈σ ],
for some clone ẗ = P̈ R̈ : Ÿ of t, some shaped substitution σ and some
isomorphism m : Ÿ → X.
The following fact is easily shown by induction over the number of star rule
applications, using Lemma 1.2. (*) If Y ⇒∗Σ S and Y ⇒∗Σ S with an isomorphism m : Y → X for some syntactic variable X in S, then Y ⇒∗Σ S[X /m S ].
Since t is shaped, we have Y ⇒∗Σ R. This carries over to the clone, so
Y ⇒∗Σ R̈. Consider those variables X1 , . . . , Xn in R̈ where every Xi has a
Yi ∈ Dom(σ) with an isomorphism mi : Yi → Xi , for 1 i n. Then fact
(*) yields that Y ⇒∗Σ R̈[X1 /m1 σ(Y1 )] · · · [Xn /mn σ(Yn )] and hence Y ⇒∗Σ
R̈σ by Theorem 1.1.
Program Graph Transformation
163
Again by fact (*), Z ⇒∗Σ C and Y ⇒∗Σ R̈σ with the isomorphism
m : Y → X for some variable X in C implies Z ⇒Σ C[X/m R̈σ ]. Hence
Z ⇒∗Σ H, and H is shaped.
Let pumPG be the shaped generic rule obtained from rule pum in Figure 9
by removing nodes 2 and 4 from the interface. This rule is shaped according to
the syntactic variable with center Hy as it appears in the adaptive star rule hy
in Figure 6.
Note however that pumPG applies only to (shaped) graphs where the superclass 1 has no further features, and the subclasses 2 and 4 have no other features
or subclasses either. In other words, the rule is still not general enough if its
applied in a shape-preserving fashion.
We can extend rule pumPG by further graph variables that are placeholders
for further features and subclasses to change this. Unfortunately, the resulting
rule is rather complicated.
5
Towards an Implementation
The concepts for shapes and rules outlined in Sections 2 to 4 shall be incorporated into a diagram programming language, diaplan. The structure for an
implementation of this language is outlined in Figure 12. The basic layer of the
diaplan system shall be obtained by extending two kernel components of the
Graph rewrite generator GrGen developed at Universität Karlsruhe [4]. The
GrGen rule compiler translates sets of rewrite rules into abstract machine code
of a rewrite engine, using highly sophisticated optimization techniques, making
GrGen the fastest graph rewriting tool available.5 Since the system already
5
Admittedly, Fujaba [24] is close in efficiency, and even better on some benchmarks,
due to a simpler specification language.
program
diagram
user
interface
diaplan
editor
diagram
editor
structure
layer
diaplan
compiler
diaplan
interpreter
GrGen
system
rule
compiler
rewrite
engine
Fig. 12. Outline of the diaplan programming system
164
Berthold Hoffmann
supports shaped graph variables discussed in Section 4, in form of recursively
defined sub-patterns and sub-rules [18], it has “just” to be extended to handle multiple nodes, multiple subgraphs, cloning, and shapes defined by adaptive
rather than plain star grammars.
The structure layer of the system shall consist of a compiler that translates classes with shape and predicate definitions into rewrite units. The diaplan
interpreter shall link these units, call rewrite code recursively, and initiate backtracking for nondeterministic predicates if necessary.
The user interface consists of a diagram editors for diaplan programs, and
a diagram editor for constructing input graphs, invoking the interpreter, and
displaying its results. The editors can be generated with the generator DiaGen
for diagram editors developed M. Minas [22, 23].
6
Conclusions
Refactoring shows that graph transformation can be useful for specifying software models and model transformations. Graph grammars, in particular adaptive star grammars with application conditions, can capture the shape of models
precisely. Thanks to multiple nodes, multiple subgraphs, and graph variables,
generic graph rewrite rules allow to specify refactorings like Pull-up Method
refactoring by a single rule. It is even possible to define criteria under which
applications of a generic rule will always stay within the language of the host
graph, even if these rules may become quite complicated.
Some work related to the concepts shown in this paper shall be mentioned
here. Shape analysis is about specification and verification of invariants for
pointer structures in imperative programming languages, e.g., whether a data
structure is a leaf-connected tree. Often, logical formalisms are used for this
purpose [29]. When specifying models by meta-models, e.g., in Uml, object constraint logic (ocl) is used for properties that are beyond the descriptive power
of class diagrams. Plump et al. have proposed graph reduction grammars for
shape specification [2]. There is a close relationship between graph languages
and (monadic second-order) logic [6], which has not yet been studied for reduction graph grammars or adaptive star grammars, however. The concept of
multiple nodes in graph rewrite rules dates back to Progres [30], and is used
in several other systems. We have extended it to multiple subgraphs. Graph
variables have been implemented, under the name of recursive sub-patterns and
sub-rules, in the extension [18] of the GrGen system [4]. Viatra2 [3] seems to
be the only other system implementing recursive sub-patterns.
Several questions and challenges remain for future research. Fundamental
properties of generic and shaped generic rewrite rules have still to be investigated, such as parallel and sequential independence of rewrite steps, and critical
pairs of rewrite rules, which are prerequisites to study confluence of (shaped)
generic graph rewrite. Another topic is termination.6 For our example, analy6
Generic graph transformation could be defined as an instance of adhesive high-level
replacement, a categorical axiomatization of graph transformation [9]. Then criteria
Program Graph Transformation
165
sis of these properties would allow to indicate which refactorings could become
applicable after a particular refactoring has been performed. And—last but not
least—the system outlined in Section 5 has to be implemented.
Acknowledgment. This paper describes the outcome of my research while I
am working with Bernd Krieg-Brückner at Universität Bremen. This work was
only possible since Bernd allowed my to continue my original line of research –
next to my “daily work”, the management of an esprit project [19], teaching
(programming languages and compiler construction), and administration of his
group at Universität Bremen. I have been lucky – and are still happy – to get
this freedom and generous support. Danke schön, Bernd!
References
1. F. Baader and T. Nipkow. Term Rewriting and All That. Cambridge University
Press, Cambridge, UK, 1998.
2. A. Bakewell, D. Plump, and C. Runciman. Specifying pointer structures by graph
reduction. Mathematical Structures in Computer Science, 2009. Accespted for
publication.
3. A. Balogh and D. Varró. Pattern composition in graph transformation rules. In
European Workshop on Composition of Model Transformations, Bilbao, Spain, July
2006. See also http://viatra.inf.mit.bme.hu/update/R2.
4. J. Blomer and R. Geiß. GrGen.net: A generative system for graph-rewriting,
user manual. www.grgen.net, 2008.
5. B. Courcelle. An axiomatic definition of context-free rewriting and its application
to NLC rewriting. Theoretical Computer Science, 55:141–181, 1987.
6. B. Courcelle. The expression of graph properties and graph transformations in
monadic second order logic. In Rozenberg [27], chapter 5, pages 313–400.
7. F. Drewes, B. Hoffmann, D. Janssens, and M. Minas. Adaptive star grammars and
their languages. Technical Report 2008-01, Departement Wiskunde-Informatica,
Universiteit Antwerpen, 2008.
8. F. Drewes, B. Hoffmann, and M. Minas. Adaptive star grammars for graph models.
In H. Ehrig, R. Heckel, G. Rozenberg, and G. Taentzer, editors, 4th International
Conference on Graph Transformation (ICGT’08), number 5214 in Lecture Notes
in Computer Science, pages 201–216. Springer, 2008.
9. H. Ehrig and K. Ehrig. An overview of formal concepts for model transformations
based on typed attributes graph transformation. Electronic Notes in Theoretical
Computer Science, 152(4), 2006. Proc. Graph and Model Transformation Workshop (GraMoT’05).
10. H. Ehrig, K. Ehrig, U. Prange, and G. Taentzer. Fundamentals of Algebraic Graph
Transformation. EATCS Monographs on Theoretical Computer Science. Springer,
2006.
11. H. Ehrig and A. Habel. Graph grammars with application conditions. In G. Rozenberg and A. Salomaa, editors, The Book of L, pages 87–100. Springer, Berlin, 1986.
for these properties come “for free”. However, such an instance would be too “coarse”
to provide good results for confluence.
166
Berthold Hoffmann
12. H. Ehrig and J. Padberg. Graph grammars and petri net transformations. In
J. Desel, W. Reisig, and G. Rozenberg, editors, Lectures on Concurrency and Petri
Nets, volume 3098 of Lecture Notes in Computer Science, pages 496–536. Springer,
2003.
13. J. Engelfriet and G. Rozenberg. Node replacement graph grammars. In Rozenberg
[27], chapter 1, pages 1–94.
14. G. Engels, H. Ehrig, H.-J. Kreowski, and G. Rozenberg, editors. Handbook of
Graph Grammars and Computing by Graph Transformation, Vol. II: Applications,
Languages, and Tools. World Scientific, Singapore, 1999.
15. M. Fowler. Refactoring—Improving the Design of Existing Code. Object Technology Series. Addison-Wesley, Reading, MA, 1999.
16. gt-vmt: International workshops on Graph Transformation and Visual Modelling
Techniques. http://www.cs.le.ac.uk/events/gtvmt09/, 2000–2009.
17. A. Habel. Hyperedge Replacement: Grammars and Languages. Number 643 in
Lecture Notes in Computer Science. Springer, 1992.
18. B. Hoffmann, E. Jakumeit, and R. Geiß. Graph rewrite rules with structural
recursion. In M. Mosbah and A. Habel, editors, 2nd Intl. Workshop on Graph
Computational Models (GCM 2008), pages 5–16, 2008.
19. B. Hoffmann and B. Krieg-Brückner, editors. Program Development by Specification and Transformation, number 680 in Lecture Notes in Computer Science.
Springer, 1993.
20. T. Mens, S. Demeyer, and D. Janssens. Formalising behaviour-preserving transformation. In A. Corradini, H. Ehrig, H.-J. Kreowski, and G. Rozenberg, editors,
First International Conference on Graph Transformation (ICGT’02), number 2505
in Lecture Notes in Computer Science, pages 286–301. Springer, 2002.
21. T. Mens, N. Van Eetvelde, S. Demeyer, and D. Janssens. Formalizing refactorings
with graph transformations. Journal on Software Maintenance and Evolution:
Research and Practice, 17(4):247–276, 2005.
22. M. Minas. Specifying and Generating Diagram Editors. Habilitationsschrift, Universität Erlangen, 2001. [In German].
23. M.
Minas.
DiaGen/DiaMeta:
A
diagram
editor
generator.
www.unibw.de/inf2/DiaGen, 2008.
24. J. Niere and A. Zündorf. Using Fujaba for the development of production control systems. In M. Nagl, A. Schürr, and M. Münch, editors, Int’l Workshop on
Applications of Graph Transformations with Industrial Relevance (Agtive’99), Selected Papers, number 1779 in Lecture Notes in Computer Science, pages 181–191.
Springer, 2000.
25. D. Plump. Term graph rewriting. In Engels et al. [14], chapter 1, pages 3–102.
26. P. Prusinkiewicz, M. Hammel, J. Hanan, and R. Měch. Visual models of plant
development. In G. Rozenberg and A. Salomaa, editors, Handbook of Formal Languages, volume 3: Beyond Words, chapter 9, pages 535–597. Springer, 1999.
27. G. Rozenberg, editor. Handbook of Graph Grammars and Computing by Graph
Transformation, Vol. I: Foundations. World Scientific, Singapore, 1997.
28. G. Rozenberg and A. Salomaa, editors. Handbook of Formal Languages, volume 1:
Word, Language, Grammar. Springer, 1997.
29. M. Sagiv, T. Reps, and R. Wilhelm. Solving shape-analysis problems in languages
with destructive updating. ACM Transactions on Programming Languages and
Systems, 20(1):1–50, 1998.
30. A. Schürr, A. Winter, and A. Zündorf. The Progres approach: Language and
environment. In Engels et al. [14], chapter 13, pages 487–550.
Program Graph Transformation
167
31. N. Van Eetvelde. A Graph Transformation Approach to Refactoring. Doctoral
thesis, Universiteit Antwerpen, May 2007.
Berthold Hoffmann received his Diplom and PhD
(Dr.-Ing.) from TU Berlin. He works in Bernd KriegBrüchner’s group since 1982.
Type-Safe Red-Black Trees with Java Generics
Stefan Kahrs
University of Kent at Canterbury
Department of computer Science
[email protected]
Abstract. Java Generics[1] are a powerful extension of Java’s class system, allowing the static type checker to ensure (as their most typical
application) the type-safe usage of homogeneous collections.
The paper shows that the expressiveness of generics goes a very long
way, by using them to enforce the tree balancing in a red-black tree
implementation. The idea originates from Functional Programming [2, 3],
though Java Generics are too different to allow a one-to-one adaptation.
Based on experiences with this application, a number of suggestions are
made for the improvement of Java Generics.
1
Introduction
Java Generics came about as a response to a long-standing problem in Java:
the need for dynamic type casts used to make many Java programs potentially
unsafe, with the possibility of class cast exceptions at run-time. In the late 1990s
Phil Wadler et al. [4, 5] showed how the Java language could be extended to
provided for a richer static analysis that would allow to eliminate dynamic casts
almost entirely. Since Java 5 these ideas have entered the language standard [6].
Red-black trees provide a well-known way of implementing balanced 2-3-4
trees as binary trees. They were originally introduced (under a different name)
in [7] and are nowadays extensively discussed in the standard literature on algorithms [8, 9]. The basic operations for search-trees (search, insertion, deletion)
are realised in O(log(n)) time, for tree size n.
Red-black trees are binary search trees with an additional “colour” field
which is either red or black. The root of the tree will always be black. In a redblack tree each red node is required to have black subtrees and is regarded as
an intermediate auxiliary node. All binary nodes are required to be of the same
black height, which is the nesting depth of binary black nodes. Therefore, every
binary black node has a black height of n + 1, where n is the black height of its
subtrees. Moreover, viewing red nodes only as auxiliary nodes, a black binary
node of height n + 1 has between 2 and 4 black subtrees of height n, depending
on whether 0, 1 or 2 of its direct subtrees have red roots. This is the reason why
red-black trees can be seen as an implementation of 2-3-4 trees.
Since the mid 1990s it has been known in the Functional Programming community that invariances such as tree-balance can be enforced through a sufficiently rich type system [2], using the feature of “nested” datatypes. Characteristic for these types is that the unfolding of their recursion (see [10]) results in
169
non-regular infinite trees; this causes problems for type inference [11] and indeed
Standard ML’s type-inference based type system fails to cope with nested types
[12]. However, this does not affect the type-checking of type-annotated code, as
in Java Generics.
It would be fairly straightforward to “ape” the FP implementation of redblack trees in an object-oriented setting. However, this would require to move
fields into the implementing concrete classes, and consequently operations on
fields higher up the class hierarchy would have to be expressed in terms of
abstract methods instead. This paper uses a more object-oriented approach that
permits directly operating with fields at abstract classes. However, this requires
to abstract the types of the fields and as a consequence the generic parameter
lists for the classes become more complex.
2
Inheriting Generic Fields
All our red-black trees (even the empty ones) are special cases of binary search
trees. Moreover, the abstract class of binary search trees also contains the only
fields for all of our various tree types. This inheritance of the fields is what makes
the implementation overall rather different from its FP cousin.
public abstract class Bin<T extends Comparable<? super T>,
F extends Find<T>>
implements Find<T>
{
public T co;
public F le,ri;
public T find(T x) {
int res=co.compareTo(x);
if (res==0) return co;
if(res<0) return le.find(x);
return ri.find(x);
}
}
The fields for the left and right branches of a tree node (le and ri) are of
the generic type F, which at this stage just suffices to implement binary search.
Subclasses of Bin will restrict the type for F further, to trees of a particular
height and (possibly) a particular colour.
Consequently, other classes in the hierarchy of tree types will need a matching
type parameter for F — it can only be removed from the class interface once it
no longer changes. In a context in which these fields are both read and written
to this concrete type needs to be made available as a generic parameter.
3
Typing Red-Black Trees
Figure 1 shows the raw type hierarchy used for the red-black tree implementation; it is “raw” in the sense that it does not include the type parameters, most
170
Bin
Black
NeBlack
DeepBlack
Red
EmptyTree
Singleton
DeepRed
ShallowBlack
Fig. 1. Raw Type Hierarchy
notably the ones responsible for the GADT effect, though the node colouring
is already modelled. Also not included are the interfaces which guide how these
types are used, and the types modelling tree underflow.
Clients for this implementation would use the trees through the interface
PureTree which hides the information about black heights: the trees need to
know how high they are, the clients do not. Providing a similar abstraction in
Functional Programming is comparatively awkward, as it would typically involve
the relatively advanced feature (it is not part of the Haskell-98 standard [13],
though all Haskell compilers support it) of existential types, as used in [3]. Java
Generics have a form of existential type abstraction of their own: wildcards.
To provide a tree type for clients the use of wildcards would have been less
appropriate, as PureTree<T> would have been replaced by the rather bewildering
Black<T,?,?,?>.
Generally, the class hierarchy regards all red-black trees as cases of binary
search trees. The tree-manipulation methods for red and black trees share common interfaces, but their implementation is considerably different, so that they
are given different classes. Within colours, classes are distinguished by height:
Singleton is the class of red-topped trees of black height 0, DeepRed the class of
red-topped trees of black height n + 1; black-topped trees have an even finer distinction, with DeepBlack having black height n + 2. Why this particular height
distinction is appropriate will be discussed later.
Notice that such a hierarchy has already considerable implications for the
classes: as trees can shrink through deletion and grow through insertion, trees
need to have some knowledge of the types that would be attached to the neighbouring levels.
There is an immediate price to pay for this design: we can neither change
the colour, nor the height of an existing tree node, because either change requires a change of type. Thus, where the standard imperative implementation of
red-black trees freely rewires (and recolours) tree nodes to realise tree rotations
to compensate for insertion overflow or deletion underflow, a type-safe implementation will have to frequently discard existing nodes and create fresh ones
instead.
171
4
Tree Heights
Red-black trees are a form of balanced trees. Left and right branch of any node
are required to have the same black height. This is enforced in this implementation through the type system, as follows:
– As already seen, left and right branch of any tree node have the same generic
type F; thus, if the branch types of the implementing classes encode height
information the tree is necessarily balanced.
– Empty trees have (by definition) black height 0, as have singleton red trees.
– Non-empty black trees have type NeBlack<T,B,O>, and their black height is
1 plus the black height of O, which is the type of their branches.
– Red trees have type Red<T,B,O> where B is the type of their (black) branches.
Their black height is the same as the black height of type B.
– Red-black trees appear as branches of non-empty black trees. They are instances of the interface RBTree<T,B,O>. The only implementing classes of
this interface, i.e. red trees and black trees, set the B parameter to a black
tree class of black height n which is equal to their own black height.
Therefore the black height of a red-black tree can be read off its type. The reason
this is unambiguous is that the type of the branches occurs as a subexpression
in the whole type expression.
The security this gives to the implementation is relative: tree height is maintained as an invariant. In itself this does not stop end users from writing further
subclasses which could break the invariant — such as ordinary binary search
trees, without any balancing whatsoever.
5
Black Trees
The code that sits at abstract class Black looks mostly rather peculiar and
mysterious. The reason is that most code that can be placed at that abstract
class is code that deals with tree operations relative to some context, such as
a tree rotation w.r.t. a parent and grandparent node. Without knowing where
that context is coming from these operations make very little sense. However,
the type interface for class Black is less obscure:
abstract public class Black<
T extends Comparable<? super T>,
F extends Find<T>,
O extends RBTree<T,B,O>,
B extends Black<T,F,O,B>
> extends Bin<T,F>
implements PureTree<T>, RBTree<T,B,O>,Embed<T,B,O>
The type T is the type of search values; type F is some type which has a method
find for searching T values, and type F is used as the types for the left and right
172
children of the tree. The type O is a red-black tree of the same black height, and
it is the type of insertion results.
Based on this, class Black can already provide the implementation for the
insertion and deletion methods when used as a pure tree:
public PureTree<T> insert(T x) { return ins(x).makePure(); }
public PureTree<T> delete(T x) { return del(x).makePure(); }
abstract public O ins(T x);
abstract public MakePure<T> del(T x);
Subclasses of this will need to constrain the result type of the del method further
— but this is possible in Java since version 5, using covariant overriding. Without
that language feature, class Black would need an additional parameter.
The final parameter of the class B is a standard trick in Java Generics (see
e.g. [6]) to refer to an implementing subclass. The value expressed by this is of
type Black<T,F,O,B>, but this can be too general to be of good use, as it forgets
too much information. Instead, the class Black leaves its concrete subclasses an
abstract method B self() to implement; inside the class Black calls to self()
will therefore lose less information than the constant this does.
The class implements (promises to implement) three interfaces: all black trees
are pure trees, all black trees are red-black trees, and furthermore black trees of
height n can be embedded into the red-black trees of height n. The last of these
three promises (the interface Embed) may appear rather peculiar, and is worth a
further look, as it also reveals a weakness in the expressiveness of Java Generics.
The problem is this: all red-black trees implement the RBTree interface; however, for height 0 (and height > 0) the interface is extended with methods specific
to zero heights and non-zero heights. What we need is that our specialised subclass B needs to implement that associated more specialised interface, e.g. be a
subclass of O. However, Java Generics do not allow subclassing constraints between variables, only between variables and classes. Therefore, within the class
Black we cannot even make the promise that B will be a subclass of O.
public interface Embed<
T extends Comparable<? super T>,
A extends Embed<T,A,B>,
B extends RBTree<T,A,B>
> extends Find<T> {
public B embed();
public B ins(T x);
public B mkPair(T x,A sibling);
public NeBlack<T,A,B> mkBpair(T x,A sibling);
}
The way around this problem is to require instead a method embed which coerces a black tree of whatever height into its associated red-black variety. As all
implementing black classes do maintain the property that they subclass their
respective red-black variety, the method always has a trivial implementation.
173
Thus this works but it is not particularly nice and calls for a slight overhaul
of Java Generics, e.g. supporting something like this as a replacement for the
specification of B:
B extends Black<T,F,O,B> & O
This should express that B not only subclasses Black in this way, but that it
also will maintain a subtype relationship with O. This is legal Java syntax for
class/interface names, but not for type variables.
The other methods in that class are needed for forming larger trees, e.g. the
method mkBpair is applied to a value of type T and another black tree of the
same height n to form a non-empty black tree of height n + 1; this particular
method is used for buffering deletion underflow.
As a final remark, it may appear a little bit peculiar that type O is part of the
generic parameter list at all. After all, height information is already recoverable
from the F parameter. It is indeed necessary, because the results of various
operations of the Embed interface will be stored in fields of a node and thus
have to match their type exactly (at this abstract level), and thus when the
class Black promises to implement Embed this type needs to be mentioned itself.
There is also a pragmatic reason to keep generic parameters at class level: even
if we could remove type O from the class interface we would then have to move
that abstraction to the level of individual methods, turning them into generic
methods — generic methods with rather complex type interfaces.
6
Generalised Algebraic Data Types
Algebraic Data Types in Haskell introduce a fresh type, plus constructors that
can be used to create values of this type. In particular, if that new type is
a parametric type A t then all constructors of the type will be polymorphic
functions with return type A t. Generalised Algebraic Data Types weaken this
condition slightly by allowing that constructors can have return type A u, where
u is a substitution instance of t. This generalisation can be used to define, for
example, length-aware lists (vectors) as follows:
data Vector a n
where
Nil :: Vector a ()
Cons :: a -> Vector a k -> Vector a (S k)
In (GHC) Haskell, this defines a fresh type Vector, with two type parameters
a and n. The first parameter is the component type of the vector, the second
the length of the vector, though this length is encoded as a type. This second
parameter has no bearance on the contents of the vectors, it is a phantom type
whose sole purpose is to make the use of vectors more rigid. With this definition,
vectors of different lengths have different types, e.g. Cons 4(Cons 7 Nil) has
type Vector Int (S(S())). A motivation for defining length-aware lists is that
174
they allows us to express indexing operations that are guaranteed to stay within
bounds.
When translating ADTs to Java, the type as a whole becomes an abstract
class, and each data constructor becomes a new subclass. This works without
change in the generalised setting, because when we are subclassing a generic
class in Java we can specialise its type parameters. For example, the subclass
corresponding to the above Cons would commence like this:
class Cons<A,K> extends Vector<A,S<K>> {
A val;
Vector<A,K> tail;
Here, the parameters of Haskell’s data constructor become fields of the new class.
Going back to our case study, the subclasses for Black and Red also use
this technique, though as explained earlier the subclasses have a rigid structure
anyway and thus there is no need to use phantom types (this would be different
for red-black trees without deletion). For example, EmptyTree extends Black
but it specialises the type parameters, as its class header shows:
final public class EmptyTree<
T extends Comparable<? super T>>
extends Black<T,Find<T>,FlatTree<T>,EmptyTree<T>>
Explanation: an empty tree has left and right branches of type Find<T> (which
will always be null), insertion into an empty tree gives a flat tree (red-black
tree of black height 0), and the EmptyTree<T> in the final argument position is
the reflection type of empty trees, which is used e.g. as result type of the self()
method.
Although empty trees already make use of GADT-style subclassing, the exploitation of the expressive power of the feature is perhaps better understood in
connection with the class NeBlack of non-empty black trees:
abstract public class NeBlack<
T extends Comparable<? super T>,
B extends Embed<T,B,O>,
O extends RBTree<T,B,O>>
extends Black<T,O,DeepTree<T,B,O>,NeBlack<T,B,O>>
Explanation: non-empty-black trees (of height n+1) store red-black trees of some
description (type O) as their left and right children. Type B is really the type of
black trees of height n, but NeBlack only needs to know that it implements the
Embed interface — and using the weaker constraint helps to keep the parameter
list of NeBlack short, because Embed mentions fewer type parameters than Black.
The result type of insertions into non-empty black trees is DeepTree<T,B,O>,
the type of red-black trees of height n + 1 (which has a richer interface than
RBTree). Finally, the reflection type is NeBlack<T,B,O>; this suggests that no
subclasses of NeBlack are needed, as they would reflect into their superclass —
175
thus, the subclasses ShallowBlack and DeepBlack are created more for convenience than necessity, as issue that will be discussed later.
Because the class header ensures that the left and right subtree of an NeBlack
object are themselves red-black trees, it can implement the ins method which
is required for all black trees:
public DeepTree<T,B,O> ins(T x) {
int res=co.compareTo(x);
if (res==0) return this;
if (res<0) return le.insL(x,this);
return ri.insR(x,this);
}
The methods insL and insR are required in the RBTree interface which the
fields le and ri must implement since their type O extends that interface. The
purpose of these methods is to insert into a subtree, but to provide the parent
as a context, in case there is an insertion overflow.
7
Iterative vs. Recursive Insertion
A consequence of the type structure for red-black trees is that insertion and
deletion cannot be implemented iteratively (i.e. through loops), the implementation has to be recursive. The reason is that the loop state would not have a
proper type: as the algorithms traverse through the tree, the visited nodes have
different types. However, a recursive implementation should not need to be less
efficient than the iterative implementation: the standard imperative realisation
of red-black trees [8] represents every edge in the tree through two pointers going
in opposite directions. For the recursive version one pointer per edge suffices, as
all required references to parent nodes sit on the run-time stack; in other words,
the method-call overhead in the recursive version is quite similar to the doublepointer overhead in iterative implementations.
During insertion and deletion a sufficient amount of parent context is provided to realise the invariant promised in the types of these operations. Moreover,
if the root of a subtree changes (because a tree rotation is required) this is signalled to the calling context simply through its return value — the iterative
version would instead update the parent through its parent pointer.
Consequently, empty trees are easier to deal with. In particular, the ins
operation of class EmptyTree is the following:
public Singleton<T> ins(T x) {
return new Singleton<T>(x,this);
}
Because the ins methods return the new root, insertion into an empty tree
can be realised by creating and returning a new node. When a tree grows to n
internal nodes it will have n + 1 occurrences of empty trees; however, instead of
176
doubling the memory requirements for a tree all these empty trees are the same
object, which is achieved by passing this to the Singleton constructor.
The need for doing this reveals again a slight weakness of Java generics.
Because the class EmptyTree is generic in T, the Singleton class cannot keep
a static copy of an empty tree, and thus it is passed to it at construction time.
This is correct and proper, but why has EmptyTree to be generic in T? The
answer is this: EmptyTree “could” implement Black without being given the type
parameter T. However, this would make its implementing methods themselves
generic, e.g. for the ins method we would get this signature:
public <T extends Comparable<? super T>> FlatTree<T> ins(T x);
The actual problem with this is that Java does not recognise a generic method as
an implementation of a less generic abstract one, and therefore a type-unaware
empty tree cannot be used to implement type-aware tree interfaces.
g
p
p
t
c
t
a
d
b
a
g
b
c
d
Fig. 2. Insertion with required right rotation
If insertions return new roots so should tree rotations. While this is possible
in a recursive and imperative realisation of red-black trees, tree rotations (as
methods) do not fit terribly well in this typed setting, because they are not
universally applicable with universally meaningful types. Instead they occur in
certain contexts, and these contexts need to be provided so that the same effect
can be achieved.
An example is shown in figure 2; here, the circular nodes indicate red colour,
the square ones black colour. In the example, an insertion into a tree has proceeded in the left child of the red node p and returned another red tree rooted
at t. Because of the colour clash this node cannot be placed as the new left child
of p — this is indicated by the dashed line. Instead, the attempt to place t as
the left child of p is informed by the context, nodes p and g. This is sufficient
to be able to form the desired (legal) tree fragment on the right. This tree fragment cannot entirely be created through pointer rewiring either: although node
t only changes colour this is now a change of class which consequently requires
the creation of a new object; the node p would be re-used as indicated in the
177
picture, but the height of p on the left and right differs, which again requires
a fresh object. Only node g retains colour and height and can be updated as
indicated.
Tree rotations of this kind sit fairly high up in the class hierarchy at classes
Black and Red themselves. The object responsible for performing this tree rotation is t; as t is red, the corresponding method sits in the class Red:
public DeepTree<T,B,O>
goLL(Red<T,B,O> pa, NeBlack<T,B,O> gpa) {
gpa.le=pa.ri.embed();
return blacken().mkPair(pa.co,gpa);
}
The method goLL generally attempts to place this object (if possible) as the left
child of (red) parent pa who is itself left child of (black) grandparent gpa; there
are four such methods for the four possible left/right combinations. The type
parameters to Red and NeBlack indicate that these have indeed types that fit
this scenario. The result type of the rotation is a red-black tree of height n + 1
and will thus implement the corresponding DeepTree interface.
Class Black has also four goXY methods which in its case (when t is black)
will not require tree rotations.
The reason for the call of embed is the earlier mentioned problem that Java
does not allow us to specify that B is a subtype of O; to avoid this call the method
definition could have been duplicated to the more specific classes Singleton and
DeepRed where this subtype relationship is established.
In class Red this method creates two fresh nodes of height n+1 (one black, one
red), and abandons two red nodes of height n. An implementation of red-black
trees in which these nodes have the same types could reclaim the abandoned
ones instead of allocating new memory.
These four goXY methods deal with (potential) insertion overflow, i.e. when
an insertion into a black tree might not result in another black tree, in a context
where a black tree is required. The classes also have methods that deal with
deletion underflow, i.e. when the deletion from a black tree results in a black
tree of lower height. These are more complex, because the trees involved in the
required rotation are of different heights.
8
Deletion
When the root node of a subtree is earmarked for deletion, the deletion process
attempts to replenish that node with the nearest node to it from one of its
subtrees (in the actual code it is always the left child, but this is not important),
in order to keep the effect of deletion local. This attempt can go wrong in one
of two different ways:
– the subtree in question could be empty
– otherwise, the removal of that nearest element might make the subtree underflow, i.e. reduce its height
178
The first of these problems goes away if the strategy is only applied to red trees
of height n + 1, and black trees of height n + 2; the second problem goes away
for trees lower in height. This case distinction is the reason for the chosen class
hierarchy.
In its most elementary form a tree underflow arises when the only node of a
black tree of height 1 is removed. Deletion underflow is propagated up the tree,
until either a red node is encountered which can absorb the underflow, or the
root is reached — in which case the overall tree shrinks in height.
Generally, the methods for underflow propagation/absorption are more complex than the methods that deal with insertion overflow, because at absorption
point tree nodes from three different height levels are involved. An example for
g
r
g
u
z
r
b
x
b
u
x
y
z
y
Fig. 3. Absorption of deletion underflow
the absorption process is in figure 3: a deletion has taken place in the left branch
of black tree g (of height n + 2). The returned result is b, a black tree of height
n — the deletion in this branch has underflown. An attempt to absorb the underflow is made by calling a method absorbL on u, the right child of g. As it is
black, it cannot itself absorb the underflow, but it makes a 2nd-level absorption
attempt, by calling absorbL2 on its left child r.
In this case r is red, and can therefore absorb the deletion underflow by
creating the tree fragment on the right, which is a tree of the same height as the
original tree. The corresponding code for r sits in class Red:
public Grey<T,NeBlack<T,B,I>,DeepTree<T,B,I>>
absorbL2(B un,DeepBlack<T,B,I> gpa,NeBlack<T,B,I> pa){
pa.le=ri.embed();
T aux=gpa.co; gpa.co=co;
gpa.le=un.mkBpair(aux,le);
return gpa;
}
179
One could slightly improve the efficiency of this, as it is not really necessary to
create a fresh black node: an abandoned black node of the right height still sits
at gpa.le, though the types do not tell us that it does.
Notice that the result type of absorbL2 is an instance of Grey, an interface
that allows for potential deletion underflow. In this case, there was no further
underflow, but all non-empty black trees of height n + 1 implement the Grey
interface at height n, so returning gpa is well-typed.
However, the type indicates that a further deletion underflow would have
been possible. Had r been black then the absorption attempt would have failed,
with the situation as in figure 4: In this case, the node r is paired with b, creating
g
u
u
b
r
z
g
z
b
r
Fig. 4. Propagation of deletion underflow
a new red node as the left child of u. The overall result would have been u, a
black tree of height n + 1 — and therefore the overall result in this section of the
tree is another underflow. However, even underflow has a height attached to it,
as is apparent in the absorbL2 method of class Black which produces this tree
fragment:
public Grey<T,NeBlack<T,B,O>,DeepTree<T,B,O>>
absorbL2(B un,DeepBlack<T,B,O> gpa,NeBlack<T,B,O> pa) {
pa.le=un.mkPair(gpa.co,self());
return new Underflow<T,NeBlack<T,B,O>,
DeepTree<T,B,O>>(pa);
}
As pa (the u in figure 4) is a non-empty black tree, it does implement the
Grey interface; however, it implements it at height n, and the signature of the
absorbL2 method requires a grey of height n + 1. Therefore a new underflow
object is created. The class Underflow is just a wrapper for black trees that are
“too low” in height. It implements the Grey interface, which solely consists of
methods for attempting to place a deletion result back into a parent context.
The method that triggered the absorption attempts in figures 3 and 4 is the
method putL, implemented by Underflow as follows:
180
public Grey<T,NeBlack<T,B,O>,DeepTree<T,B,O>>
putL(DeepBlack<T,B,O> pa) {
return pa.ri.absorbL(body,pa);
}
The tree body is a field of Underflow; it is that mentioned “too low” black tree.
The reason this wrapper class is needed is that method selection is not informed
by the type parameters of generic classes. Thus we could not make a black tree
class implement two different type instances of the Grey interface, i.e. one where
the heights match and the parent node can be updated, and one where they are
one level out of step and the absorption methods are called.
9
Java Language Issues
In the process of developing this application a number of problematic issues
about various details in Java Generics were encountered. Only some of them
remain nuisances for the final version of the red-black tree implementation, but
the other issues remain worth discussing anyway.
Beside language issues, there were also problems with the compiler, e.g. certain type dependencies made it crash, and some complex class/interface relations
were unjustly rejected, but this section focuses on language design issues only.
Already mentioned earlier were the problems that it is impossible to specify
subtype relationships between type variables, and that generic methods cannot
implement non-generic ones. It is not easy to see how the second of these ideas
can be sensibly incorporated into Java, but the first one should be possible and
worthwhile to do.
A number of further issues are dedicated a subsection each:
9.1
Type Abbreviations
Java has no feature to abbreviate type expressions. Before generics there was
very little need for that, although import clauses also have a name-abbreviating
effect and purpose.
Within generics, type expressions can grow to rather unmanageable levels.
Earlier we have seen the return type of absorbL2 barely fitting onto a single line,
and this is a return type quite a few methods share. It is possible to introduce
a shorter name by simply subclassing a type expression. This is indeed one of
the reasons for providing the (not strictly necessary) class DeepBlack; without
this subclass, all methods mentioning DeepBlack<T,B,O> in their parameter list
would have to replace each occurrence of this type by the rather unwieldy:
NeBlack<T,NeBlack<T,B,O>,DeepTree<T,B,O>>
However, abbreviating a type expression through subclassing still introduces a
proper subclass instead of an alias, and this can create problems — which is
why the same technique has not been used for example for the return type of
methods such as absorbL2.
181
Thus, the author proposes the following Java extension: permit inner class
declarations like the following:
private class DG =
Grey<T,NeBlack<T,B,O>,DeepTree<T,B,O>>;
This should introduce DG as an alias for the type expression on the right-hand
side, i.e. it should be freely exchangeable with that type expression. Of course,
this would not be necessary if Java had a macro language.
9.2
GADTs with closed world assumptions
The issue of not always being able to use subclassing for type abbreviations also
points to a subtle distinction between GADTs Haskell-style and their realisation
in Java: GADTs come with a closed-world assumption, in that the values formed
with their constructors are the only values of their defined type.
Going back to the vector example from section 6, values of type Vector a
(S k) can only be formed using the constructor Cons. In the Java realisation of
this type, nothing would stop us from defining a third subclass of vectors which
also extends Vector<A,S<K>>. This distinction can be rather irksome. We may
have code that produces a Vector<A,S<K>>, and we have written Cons as the
only instance of that class, and yet they are only assignment-compatible in one
direction (unless we resort to casts).
Looking at it differently: in Java we can only place code at a generic class
when all its generic parameters are uninstantiated. In order to place code at a
class like Vector<A,S<K>> we have to subclass Vector with its second parameter instantiated to S<K> and place the code at that subclass. This leaves the
possibility of other subclasses of Vector<A,S<K>> that may not provide such
code. Potential discrepancies of this ilk have complicated the red-black tree implementation tremendously, as precise type information needs to be threaded
through various class interfaces to make types fit.
There is not an obvious Java language extension that fits into the Java system
and permits such behaviour, which is unfortunate as in practice this is probably
the biggest obstacle in the usability of this approach.
10
Conclusions
On the example of red-black trees this paper has shown how properties such
as tree-balancing can be expressed within Java Generics, to the extent that
the type-checker will guarantee the data invariants. The ideas are known from
functional programming but have been adapted for a better fit to Java’s features.
In contrast to functional programming, the inability to place code at an instantiated class (without introducing a fresh subclass) is extremely inconvenient
in this setting: the resulting potential type incompatibilities have to be counteracted by extending generic parameter lists throughout the implementation.
182
Some of the implementation comes with a performance penalty: each object
has a fixed invariant in this setting. Therefore, object reuse is restricted to situations with the same invariant; instead of changing the invariant of an object
a fresh object with different invariants would have to be created that takes its
place. In particular, this applies to some of the objects involved in tree rotations.
One can use the height-verified code as the basis for a more efficient (but
still recursive) implementation of red-black trees that eliminates this overhead.
This involves mostly fine-tuning tree-rotation methods by reclaiming memory
we know will become free. While this can be seen as a program transformation,
it is one informed by global assumptions about the memory such as single use
of tree pointers, and thus goes beyond mere refactoring.
Expressing red-black tree balancing tests Java Generics a fairly long way,
and so this journey discovered a few deficiencies of Java Generics, which leads
the paper to make several suggestions for improvements.
References
1. Bracha, G.: Add Generic Types to the Java Programming Language. Technical
Report JSR-000014, Sun Microsystems (2004)
2. Hinze, R.: Constructing red-black trees. In: Workshop on Algorithmic Aspects of
Advanced Programming Languages, Paris (1999) 89–99
3. Kahrs, S.: Red-black trees with types. Journal of Functional Programming 11(4)
(2001) 425–432
4. Bracha, G., Odersky, M., Stoutamire, D., Wadler, P.: Making the future safe for
the past: Adding genericity to the Java programming language. In Chambers, C.,
ed.: ACM Symposium on Object Oriented Programming: Systems, Languages, and
Applications (OOPSLA), Vancouver, BC (1998) 183–200
5. Bruce, K.B., Odersky, M., Wadler, P.: A statically safe alternative to virtual types.
In: ECOOP ’98: Proceedings of the 12th European Conference on Object-Oriented
Programming, London, UK, Springer-Verlag (1998) 523–549
6. Naftalin, M., Wadler, P.: Java Generics. O’Reilly (2007)
7. Bayer, R.: Symmetric binary b-trees: Data structure and maintenance algorithms.
Acta Informatica 1 (1972) 290–306
8. Cormen, T.H., Leiserson, C.E., Rivett, R.L.: Introduction to Algorithms. MIT
Press (1990)
9. Sedgewick, R.: Algorithms. Addison-Wesley (1988)
10. Cardone, F., Coppo, M.: Two extensions of Curry’s type inference system. In
Odifreddi, P., ed.: Logic and Computer Science. Academic Press (1990) 19–76
11. Kfoury, A.J., Tiuryn, J., Urzyczyn, P.: Type reconstruction in the presence of polymorphic recursion. ACM Transactions on Programming Languages and Systems
15(2) (1993) 290–311
12. Kahrs, S.: Limits of ML-definability. In: Proceedings of PLILP’96. (1996) 17–31
LNCS 1140.
13. Jones, S.P., ed.: Haskell 98, Languages and Libraries, The Revised Report. Cambridge University Press (2003)
183
Stefan Kahrs I was in my third semester of Computer Science when BKB joined Uni Bremen, which if my memory is right - was in 1983. I first encountered him when he taught our programming module.
After the Vordiplom I picked a project that was supervised by him, and after graduation in 1986 I became
a Wissenschaftlicher Mitarbeiter of his group - which
amongst other things attached me to the Prospectra
project. Finally I finished my PhD early in 1991, and
in May 1991 I left Bremen. I am now lecturer at the
University of Kent.
Compensating the Computational Bias of
Spreadsheets
Andrea Kohlhase and Michael Kohlhase
German Center for Artificial Intelligence (DFKI)
{Andrea,Michael}[email protected]
Abstract. Spreadsheets are heavily employed in administration, financial forecasting, education, and science because of their intuitive, flexible,
and direct approach to computation. They represent examples for “active documents” which are considered very worthwhile in the field of
Knowledge Management. But these activity traits also lead to usability
and maintenance problems, as spreadsheet-based applications grow evermore complex and long-lived. We argue that these difficulties come from
a fundamental bias of spreadsheets towards computational aspects of the
conceptual models underlying the applications, i.e., a computational bias.
Concretely, we present the software architecture of the SACHS project
geared towards a semantic enhancement of MS Excel spreadsheets that
aims at compensating this computational bias.
This paper contains our view on the foundations of the SACHS project which aims
to extend semantic document modeling to spreadsheets. Bernd Krieg Brückner has
been proposing for years to do this. He always maintained that spreadsheets would
ideal targets for semantic preloading. Indeed our research shows that spreadsheets
are semantically much more interesting than anticipated. Therefore we dedicate this
paper to him as a tribute to his intuitions.
1
Introduction
In 1979 Dan Bricklin and Bob Frankston co-authored the famous first
spreadsheet program: “VisiCalc” [BF79]. Terry Winograd considers the key
innovation of spreadsheet programs to be that they virtualize ledger sheets.
In particular, “spreadsheet virtuality combines the regular structure of the familiar ledger sheet with an underlying structure of interlinked formula” [Win06,
p. 230]. Ledger sheets have been well used in financial accounting scenarios and
as spreadsheets they obtain the additional quality of being computational: the
spreadsheet user can manipulate data directly or via the abstract level of formulas in order to appropriate the data setting, to forecast developments, and to
test financial scenarios.
The management qualities led to wide spread use and reuse of spreadsheets
(see [AE06] and its references for a discussion). But unfortunately, the intensity
with which they were used yielded more and more complex documents intended
185
to be shared among people who needed to understand the same (or similar) data
sets. In turn, this very complexity now often results in errors on the data level
and misinterpretation or misapprehension of the underlying model.
The work reported here originates in the SACHS project (Semantic Annotation
for a Controlling Help System) at the German Center for Artificial Intelligence
(DFKI), Bremen. It addresses the usability problem that even though MS Excel
spreadsheets have the potential to serve well as an interface for a financial controlling system, DFKI employees experience them as too complex in practice —
even longtime users cannot interpret all data and are not certain about their origins. A controlling system is a means for the organization to control finances, i.e.,
to understand profits and losses and draw conclusions, thus a lack of overview
hampers the process: if users are not sufficiently informed they cannot optimize
the next data to be gathered.
A key observation in our approach to solving these usability problems is that
spreadsheets are active documents, we take this to mean that they are documents
whose surface structure can adapt to the environment and user input. “Activeness” in documents can only be realized by making some of the relations between
the underlying objects explicit to the software layer presenting the document. In
the case of spreadsheets these relations are given as the spreadsheet formulae. In
this paper we take a foundational stance and analyze spreadsheets as “semantic
documents”, where the formula representation is only the computational part of
the semantic relations about how values were obtained. In this sense the focus on
the computational aspects constitutes a bias towards computational aspects in
spreadsheets. In particular, spreadsheets disregard other semantic relations that
are important for understanding and relegate them to informal treatment, which
we claim accounts for a good part of the usability problems lamented above.
To compensate the diagnosed computational bias we propose to augment the
two existing semantic layers of a spreadsheet — the surface structure and the
formulae by one that makes the intention of the spreadsheet author explicit. In
the SACHS project we encode the intention as an accompanying OMDoc [Koh06]
document and can thereby provide multi-layered, semantic help services.
2
Semantic Layers in Spreadsheets
Instead of developing the general theory, we will expose the salient parts of our
approach using Winograd’s example spreadsheet from [Win06] (re-represented
in MS Excel in Figure 1) as a running example. We will differentiate the three
semantic layers in turn and draw conclusions viewing this spreadsheet as both
an active and a semantic document.
2.1
Active and Semantic Documents
We call a document semantic, iff it contains an infrastructure that distinguishes
between content and form. Note that such an infrastructure can range from superficial styling information in PowerPoint slide masters or LATEX document
186
Fig. 1. A Simple Spreadsheet after [Win06]
classes, over RDFa [W3C08] annotations in web pages to formal logic specifications of program behaviors. The idea is that this infrastructure makes relations
between the objects described in the document explicit, so that they can be
acted upon by machines. In particular, semantic documents can be interpreted
by “presentation engines” that operationalize the semantic relations by allowing
the reader to interact with various aspects of the semantic properties. We call the
combination of a semantic document with a presentation engine that can adapt
the surface structure of the document to the environment and user input an active document. Our definition is between the concept of embedding semantic
networks into hyper-documents employed by Gaines and Shaw in [GS99] and
the rather visionary notion of documents that can answer arbitrary questions
about their content proposed by Heinrich and Maurer [HM00]. Crucially both
presuppose some kind of content representation in or near the document and a
suitable “presentation engine”.
For the purposes of this paper, we will neglect the fact that most presentation engines also incorporate editing facilities and concentrate on the interaction with active documents for reading and exploration. This view is similar to
what [UCI+ 06] call “intelligent documents”.
A paradigmatic example of an active document is a Mathematica notebook [Wol02], where equations and mathematical objects can be inspected, visualized, and manipulated by the user. Here, the semantic document is written in
the Mathematica markup language which includes a content markup scheme
for mathematical formulae and a high-level programming language. The presentation engine is the Mathematica front-end which presents interactive documents to the user and calls the Mathematica kernel for evaluation of program
fragments and computation of mathematical properties.
Spreadsheets are another paradigmatic class of active documents. Here the
semantic document contains representations of the cell values or formulae together with display information such as cell color, font information, and current
viewport. The presentation engine is a spreadsheet program like MS Excel, which
187
presents the semantic document to the user by giving it a grid layout and recalculates values from formulae after each update. But what is the underlying
semantic model, i.e. what is the “activeness” of spreadsheets based on?
2.2
The Surface/Data Layer
If we look at the example in Figure 1, we see that the grid of cells can be
roughly divided into three areas. The darker, ochre area in the center contains
values of actual and past expenses and revenues; the lighter, yellow box on the
right contains values projected from these. The white region that surrounds
both boxes supplies explanatory text or header information that helps users to
interpret these numbers. Generally, non-empty cells that do not contain input or
computed values usually contain text strings that give auxiliary information on
the cells that do; we call these cells collectively the legend of the spreadsheet,
since they serve the same purpose as the legend of a map.
Observe that row 17 displays the central values of the spreadsheet: the
profit/loss situation over time (i.e., in the years 1984-1988 as indicated by the
values in row 4). Moreover note that the meaning of the values in row 17 is that
they represent profits and losses as a function π of time: recall that a function is
a right-unique relation — i.e., a set of pairs of input values and output values.
In our example the pair 1984, 1.662 of values of the cells [B4] and [B17] is one
of the pairs of π. We will call such a grid region a functional block, and the
function it corresponds to its intended function. Empirically, all non-legend ,
semantically relevant cells of a spreadsheet can be assigned to a functional block,
so we will speak of the functional block and the intended function of a cell.
In the general case a functional block consists of multiple rows and columns:
if it represents a binary (i.e., two-argument) function, then its values depend on
two (main) parameters which are usually in the cells of the column on the left
and the row on top of the block. We will call the cell ranges of arguments of the
function of a block its input row/column. In our example the block with cells
[B9:D13] ranges over expense categories (given in [A9:A13]) and years (given in
[B4:D4]) and therefore represents a binary function.
Of course the actual values can also depend on more parameters, but these are
given lower salience by the restriction to a two-dimensional grid. Alternatively, an
n-ary function is split up into a sequence of binary functions and the spreadsheet
grid contains blocks for all of these. If a functional block has only one row or
column, then it represents a unary (i.e. one-argument) function, and in the case
where it only consists of a single cell, it represents a constant (i.e., a function
without arguments).
Our notion of a functional block is related to, but different from Abraham
and Erwig’s of a “table” [AE04,AE06], which also contains row and column
header and footer blocks. Our functional blocks roughly correspond to their
“table core”, whereas we would consider their header blocks as input blocks or
legend cells (but there may be non-header input blocks in our model) and their
footer blocks which contain aggregation cells as separate functional blocks.
188
2.3
The Formula Layer
A spreadsheet cell c may not only be associated with a simple data item (or
value) [[c]], it may also be connected with a formula [[[c]]], i.e., an expression built
up from constants, an extended set of numeric and logic operators, and references
to other cells, which evaluates to the cell value. Without loss of generality we
assume that cells with a constant value given by the spreadsheet author carry
the pseudo-formula input.
In our example, the value of π can be computed from the yearly revenues in
R: = [B6:F 6] and the total expenses in E: = [B15:F 15] by a simple subtraction,
the total expenses can in turn be computed by summing up the various particular
expense categories listed in cells [A9:A13].
Note that the formulae of cells in a functional block have to be “cp-similar ”
[BSRR02], i.e., they can be transformed/copied into each other by adjusting the
respective rows and columns. We call a functional block computed if all of its
formulae are cp-similar. In our example, the functional block P: = [B17:F 17] is
computed: let γ range over the columns B to F in P. Note that the formulae
[[[γ17]]] = γ6 − γ15 in cells [γ17] are indeed cp-similar. Together, they make up
the function
F(P): = {[[γ6]], [[γ15]], [[γ17]]|γ ∈ {B, . . . , F }}
We call F(P) the function induced by the (formulae in) block P. But we also
observe that not all functional blocks in a spreadsheet are computed, for instance
the block [B9:D13] only has the pseudo-formula input, since all the values are
provided by the spreadsheet author. We call such blocks data blocks and note
that the property of being a functional block only depends on a functional correspondence (a conceptual aspect of the data) and not on the existence of formulae
(a property of the spreadsheet).
With spreadsheet formulae, users can express data dependencies on a generic
level, so that the spreadsheet program can do much computational work in the
background. By this virtualization of the traditional ledger sheet (see above),
the user’s role is lifted to a layman programmer and offers according potential.
But Abraham and Erwig report an error rate of up to 90% (!) in spreadsheets [AE06], which shows that this potential comes with a substantial risk.
They analyze the source of many of these errors to be in a mismatch between
what the spreadsheet author wants to express and the formulae he writes. They
try to address this situation by static analysis techniques (type checking) of the
formulae and supplying the author with “spreadsheet templates”. To understand
this mismatch better, let us now turn to the model the author intends to convey.
2.4
The Intention Layer
Note that even though F(P) and π compute the same values, they are completely
different functions. π is defined via the actual or projected profits or losses of
an organization, while F(P) is a finite, partial binary arithmetic function. Even
when we compose F(P) with F(R) and F(E) and restrict them to the years
189
1984-86 yielding F: = F(P) ◦ F(R), F(E)[[B4:D4]] , the functions F and π are
only extensionally equal (they are equal as input/output relations) and still differ
intensionally.
Surprisingly, only F is explicitly represented in the spreadsheet of Figure 1,
moreover, this explicit representation is invisible to the user if she doesn’t look
at the formula boxes — thus, leaving the user to figure out the ‘intention’ (the
function π) from the implicit information given in the white part by herself. This
is why we speak of a computational bias of spreadsheet programs, as some
layers of the semantics are explicated but others are not.
Generally, we can assume that spreadsheet program authors use spreadsheets
to express and compute (measurable) properties of situations; if we look a little
closer then we see that these are not properties of the world as such, but of a
high-level, abstract, or mental model of the world, which we subsume under the
term intention of the spreadsheet. In our example, the function π could be
seen as a concept from the intention, whereas the function F can be seen as its
implementation. In our simple example the intention is easy to deduce from the
text in the legend and basic financial accounting knowledge.
But even here, some parts of the spreadsheet’s intention remain unclear: e.g.
for what company or department are the profits and losses computed or what are
the methods of projection for the years 1987/8. Let us now take stock of what
the cells in the spreadsheet mean and what information we need to infer from
this. As we already remarked above, the values of cells [B17:D17] are (the scalar
parts of) the actual profits/losses in the years 1984-1986. We need information
from cell [A3] for the unit of measurement, from cells [B3:D3] that they are
actual, and from [A17] for the interpretation as a ‘profit/loss’. To understand
the full meaning of these cells, we also need to know about profits and losses of
companies — e.g. that high profits of a company I am employed by or that I
own stock in are good for me, the fact that the company is based in the USA
and therefore calculates finances in US$, and that values that are actual are
computed from measured values. Finally, we need to know that the profit/loss
of an organization over a time interval is defined as the difference between its
revenues and expenses over this interval. This knowledge allows to compute the
values of cells in P with the respective formulae from the values of cells in R ∪ E
(i.e., using the function F). The values of the cells in E can be similarly computed
from the values of the cells [B9:D13]. Note that while the definition of profits
and losses above is general accounting knowledge, this definition is particular to
the respective company, as the applicable expenses vary with the organization.
A similar account can be given for the projected profits/losses in cells [E17:F 17],
only that the interpretation of the cells wrt. the intention is even more difficult — even though the situation is simple if taken at face value. Cell [[E17]]
is the projected profit in the year 1987, which is computed from the revenue
and expenses in column E. But in contrast to the values in the actual block
[B6:D6] ∪ [B9:D13], the values in the projected block [E6:F 6] ∪ [E9:F 13] are
not measured, but projected from the actual values by some financial forecasting
method that is reflected in the respective formulae. Note that the correspondence
190
of the formula need not be as direct as in the case of the total expenses above.
It might be that the forecasting method is defined abstractly, and the concrete
formula is derived from it making some simplifying assumptions. Furthermore,
we need to know what assumptions the forecasting method makes, what parameter values are employed and why, how reliable it is, etc. to fully understand the
values. All of these concerns are not addressed at all in the spreadsheet as an
active document. Abraham and Erwig describe this as follows:
There is a high level of ambiguity associated with spreadsheet template
inference since spreadsheets are the result of a mapping of higher-level abstract models in the user’s mind to a simple two-dimensional grid structure. Moreover, spreadsheets do not impose any restrictions on how the
users map their mental models to the two-dimensional grid (flexibility is
one of the main reasons for the popularity of spreadsheets). Therefore the
relationship between the model and the spreadsheet is essentially manyto-many [...].
[AE06,
p. 5]
3
Compensating the Computational Bias
Our analysis of the example above has shown us that large parts of the intention
of a spreadsheet is left implicit, even though it is crucial for a user’s comprehension. In particular, a user needs to know the following for a spreadsheet:
– The ontology1 , i.e., background information about relations between concepts and objects of the intention. The objects in the intention include the
functions represented in the spreadsheets e.g. π, their properties, e.g. the
units of their arguments and values, and thus of the values in the cells.
– The provenance of data in a cell, i.e., how the value of this data point was
obtained, e.g. by direct measurement, by computation from other values via
a spreadsheet formula, or by import from another source; see [MGM+ 08] for
a general discussion of provenance.
– The interpretation, i.e., a correspondence between functional blocks and
concepts or objects of the intention. We distinguish three parts here
• The functional interpretation, that specifies the intended function of
the functional block.
• The value interpretation, i.e. a bijective function that specifies how
to interpret the values of the block cells as ontology objects.
• The formula interpretation that links the formulae of a block to an
object in the ontology. This mapping must be a refinement in the sense
that the interpretations of proper formulae compute the same value as
the formulae itself and the pseudo-formulae input is mapped to a provenance object.
1
In this paper we use the word “ontology” in its general, philosophical meaning as
the study of the existence of objects, their categories and relations amongst each
other, and not in the Semantic Web usage, where it is restricted to formal systems
with tractable inference properties (description logics).
191
In some spreadsheets that are the digital equivalent to “back-of-the-envelope
calculations”, the interpretation, provenance, and ontology information is simple
to infer, so that the un-documented situation is quite tolerable. Indeed this shows
the cognitive strength of the table metaphor, in our example it is no problem for
the human reader to interpret the legend item “(in Millions)” as a specification of
the value interpretation of the cells [B6:F 17] (but not of the years Y: = [B4:D4]).
In many cases spreadsheets have developed into mission-critical tools that
are shared amongst whole departments, because they encapsulate important,
non-trivial, institutional knowledge. The intention of such spreadsheets is much
harder to infer, a fact that is witnessed by the fact that companies spend considerable energy to train employees in the usage (and intention) of such spreadsheets.
In this situation, it would be natural to extend the activity of spreadsheets
to support the user’s comprehension of the spreadsheet intention. Thus, in light
of the discussion of Section 2.1, we suggest to compensate for the computational bias diagnosed above by extending the underlying semantic document of
a spreadsheet. Concretely, we propose to represent the intention (as the provenance and ontology) and to tie the cells in the spreadsheets to concepts in the
intention (via an interpretation function).
3.1
Fixing the Ontology
We have (at least) two possibilities to extend spreadsheets with an ontology
and provenance component: we can either extend spreadsheets by ontology and
provenance facilities or we can extend them by interpretation facilities that reference external representations of the intention. As we have seen above, the
intention contains quite a lot of information, and making it explicit in a software
framework means a large investment. Therefore we contend that an external
representation of the intention is more sensible, since it can leverage pre-existing
tools and profit from interoperability.
We use the OMDoc format (Open Mathematical Documents, see [Koh06]) to
represent the intention model. OMDoc is an XML-based format for representing
semi-formal, semantically structured document collections. It allows to factorize knowledge into “content dictionaries” that serve as constitutive contexts for
knowledge elements. OMDoc provides a mathematically inspired infrastructure
for knowledge representation: document fragments are classified by their epistemic role as e.g. axioms, definitions, assertions, and proofs and mathematical
formulae are represented in the OpenMath [BCC+ 04] or MathML [ABC+ 03]
formats. Furthermore, OMDoc provides a strong, logically sound module system based on structured “theories” (content dictionaries extended by concept
inheritance and views) [RK08]. Finally, the language has been extended to deal
with units and measurable quantities [HKS06] as a prerequisite for interacting
with the physical world. We make use of all of these features for modeling the
intentions of spreadsheets. In contrast to other ontology modeling languages like
OWL [MvH04], the OMDoc format does not commit to a formal logical language, and therefore lacks a native concept of inference but also does not force
192
the author to fully formalize the spreadsheet intention and to work around the
expressivity limitations of the underlying logical system. Instead, OMDoc allows
to locally formalize elements — and thus provide partial inference — with whatever formal system is most suitable; in our application, we mainly use an formal
system for arithmetics as a counterpart for spreadsheet formulae.
For the intention model in our example we divide the background knowledge
into theories that inherit functionalities from each other via the imports relation.
At the very basis we would have a content dictionary Revenues that defines the
concept of the revenue of an organization over a time interval. This theory defines
the concept of a binary revenue function ρ, such that given an organization o and
a natural number n the value ρ(o, n) is the revenue (as a monetary quantity) of
o over the span of the year n (AD) in the Gregorian calendar. Note that we use
this very naive notion of revenues for didactic purposes only. For instance ρ was
chosen as a binary function to highlight that there is no automatic agreement
between functional correspondences in the table and objects of the intention.
We would be perfectly free to analyze the concepts more thoroughly, embarking
into representing monetary systems, theories of time, etc. For the purposes of
this paper, we assume that we can either appeal to the intuition of the user or
inherit these representations from a suitable foundational ontology.
In the same way we proceed with a content dictionary Expenses, which imports from a content dictionary Salaries. Finally, we build a content dictionary
Profits that imports from both. In the OMDoc document pictured in Figure 2
we have summarized some of the relevant content dictionaries and the concepts
they introduce.
3.2
Fixing the Provenance
We enrich our ontology with provenance information: As we have required the
formula interpretation to be a refinement, we need to represent an abstract
notion of spreadsheet computation in the ontology. This can be readily done
making use of e.g. the CASL libraries [CoF04]. For modeling the provenance of
user inputs in spreadsheets, we can be extremely minimalistic, just establishing
a stub content dictionary that lifts the concept of “user input” to the ontology
level. But in our example we can already see what a more elaborate provenance
model could give us: We could specify that the salary values in [B9:F 9] are not
only user inputs, but really manually copied over from another spreadsheet —
the top spreadsheet “Salaries” in Figure 2. To take advantage of this (see details
in the next section) we have to develop content dictionaries for provenance,
adapting and extending first formalizations reported on in [MGM+ 08]. As this
goes beyond the scope of this paper, we leave this to future work.
3.3
Fixing the Interpretation
To interpret the cells in P for example, we need to
193
– fix the functional interpretation: Identify that P and R form functional
blocks with input row Y.
In Section 2.2 we have already discussed the functional interpretation of P:
it is just the intended function π. Similarly, we link the revenues block R
with a unary function ρ(c, ·), where c is the representation of our example
company. Then we have to express the semantic counterpart of the spreadsheet formulae. In our OMDoc format we can simply represent this function
as λy.ρ(c, y) in OpenMath or MathML.
– fix the value interpretation: In our example we observe that the values in
P are actually only the scalar parts of measurable quantities, in this case
measured “in millions” (of US$ presumably) according to the spreadsheet
legend. Similarly, the (string) values Salaries, Utilities, . . . in [A9:A13]
have to be interpreted as objects in the ontology. Thus in our example,
we choose i: = {y "→ y(AD)} as the value interpretation of block Y and
j: = {x "→ 106 xUS$} for block P; obviously both are bijective.
– fix the formula interpretation: our example the formulae γ6 − γ15 in the
functional block P would be linked to the formula π(year) = ρ(year) −
(year) in the FMP element of the Profit/Year definition. Furthermore, we
would link R to a semantic provenance object “imported from Salaries.xsl”.
In Figure 2 we show the functional interpretation mappings as red, dotted arrows and the formula interpretation as purple, dot-dash-arrows. The totality
of cell interpretations induces a set of content dictionaries that is associated
with a spreadsheet, we call these content dictionary the intention model. Note
that we can see the intention model as a representation of the intention of this
spreadsheet.
We can think of the value interpretations as parser/generator pairs that mediate between the scalar function represented by the formulae in the spreadsheet and the intended function in the intention — which is usually a function between measurable quantities. In particular the functions F and π are
related via the commutative diagram, where the function F is induced by the
spreadsheet formulae as discussed in Section 2.4 above. We see that the three
components of the interpretation fully
F
specify the correspondence between
spreadsheet
Y
P
functional blocks in the spreadsheet and
−1
j
i
objects induced by the intention model.
year π
profit
To see the strength of this construction
intention
let us return to our example and look
at the import of salaries from Salaries.xsl. There we have a different value
interpretation for the functional block [F 6:F 6]: this spreadsheet does not calculate in millions, so we chose k: = {x "→ xUS$} and get the import functions
k ◦ j −1 = {xUS$ "→ 10−6 xUS$} in the intention and j −1 ◦ k = {x "→ 10−6 x} in
the spreadsheet.
In conclusion we observe that setting the ontology, provenance, and the interpretation of a cell gives us a full and explicit account of its intention, and we
can relegate all further semantic services to the intention model. For instance we
194
Fig. 2. The SACHS Information Architecture and Control/Search Panel
can verify (using inference in the intention model) this part of the spreadsheet
by establishing the equality of j −1 ◦ F ◦ j and ρ(c).
The SACHS system is a work-in-progress add-in for MS Excel (written in Visual Basic for Applications) that aims at providing semantic help facilities for
spreadsheets. Though it has been developed for the DFKI controlling system, it
works for any spreadsheet whose intention has been encoded as an OMDoc document, e.g. our running example, and mash-up information has been provided2 .
We have designed SACHS as invasive technology that extends well-used (and
therefore well-known) software systems from within to overcome usage hurdles —
see [Koh05b] for a discussion. We designed the semantic services in SACHS to allow the user to explore the intention of a spreadsheet from her micro-perspective,
i.e., from her here-and-now. In particular, all semantic help facilities start from
the cell level. Moreover, we tried to implement a processual interaction design,
i.e., a design where the semantic help evolves in a user-steered process.
For services that visualize the intention, the cells in the spreadsheet must
be interpreted, i.e., linked to elements of the accompanying OMDoc document
as e.g. shown in Figure 2. Generally, all cells in a functional block are linked
to an OMDoc definition — the definition of its intended function, while OMDoc assertions justify their formulae. Concretely, for instance, the Dublin Core
metadata element dc:subject of an OMDoc definition can be used as a SACHS
2
We will spare the reader the technical details of this mash-up for lack of space.
195
label for the cell it interprets. MS Excel’s comment functionality is hijacked to
create SACHS comments that draw on the respective dc:description element,
which contains a concise description of the object in question. In contrast, the
CMP element of the OMDoc definition contains a detailed explanation using semantically marked up representations of objects from the intention model. These
can be mathematical formulae encoded as OpenMath objects like the revenue
function ρ or technical terms like “difference” which we have decorated in angle
brackets in Figure 2.
A formula underlying a cell is mirrored in the formula element FMP of the
respective definition in the semantic document (see Figure 2) in the OpenMath
5
format, this allows us to present it to the user in math notation: i=1 i (1984)
is more readable than “=SUM(B9:B13)”.
Fig. 3. Navigating the Spreadsheet Intention
3.4
Semantic Services in SACHS
In Figure 2 we have already shown the main control panel of the SACHS system
in the right bottom hand corner. This allows the user to enable various semantic
enhancements in the spreadsheet and also search the intention semantically.
From a semantic point of view, contextual views of the spreadsheet intention as
the one in Figure 3 are probably most interesting and helpful. Such views allow
a user to understand the connections between spreadsheet cells and background
information. This view aggregates information about
196
– the intention of a cell in terms of the intended function of its functional
block, and the intended arguments.
– how the cell value is computed from the values of its arguments.
– and the intentions of the arguments.
All information points in the pop-up are hyperlinked to their respective sources
in the OMDoc document or in the spreadsheet so that the graph view can also
serve as a meaning-driven navigation tool. In a future version of SACHS, these
hyperlinks could pre-instantiate the intention model with the argument values
and allow an exploration from the view of the current cell — in our example in
Figure 3 the intention specialized to the year 1984. Note that our example here
only shows the situation for a formula-computed cell. For other provenances,
the pop-up would visualize the provenance object given by the formula interpretation. For instance, for cell [B9] the provenance object is “imported from
Salaries.xls[B5]”, so we can visualize the data from that cell using the existing
SACHS pop-up.
As OMDoc is organized by theories, the provision of multi-level theory graphs
as in the CPoint system [Koh05a] are nice-to-have services one can think of.
4
Conclusion and Outlook
We have analyzed the reasons for users’ difficulties in understanding and appropriating complex spreadsheets, as they are found e.g. in financial controlling
systems. We claim that the ultimate cause is that spreadsheets are weak as active
documents, because their underlying semantic documents are biased to computational aspects and fail to model the provenance, interpretation, and ontological
relations of the objects and concepts operationalized by the system. To remedy
the situation we propose to explicitly model the intention of a spreadsheet as
an intention model in a collection of OMDoc content dictionaries which serve as
an explicit knowledge base for the spreadsheet. Finally, we present the work-inprogress SACHS system that draws on such an intention model to offer various
semantic services that aid the user in understanding and interacting with the
spreadsheets.
In essence, our approach makes double use of the following duality identified
by Fensel in [Fen08]
– Ontologies define formal semantics for information, consequently allowing information processing by a computer.
– Ontologies define real-world semantics, which makes it possible to
link machine processable content with meaning for humans based on
consensual terminologies.
In the analysis we look at the formal semantics of spreadsheets and find a computational bias that hampers understanding since it fails to model consensual terminologies and therefore leads to real-world usability problems. In our proposed
solution we extend the formal semantics of spreadsheets to draw on explicitly
represented consensual terminologies.
197
While a semantic help system for a spreadsheet-based controlling system was
the original motivation for our analysis, we feel that an explicit representation
of the intention model of a spreadsheet has many more applications: it can be
used for verification of the formulae, for change management along the lines
of [MW07], and automated generation of user-adapted spreadsheets.
Our approach seems to be related to “Class Sheets” [EE05] introduced by
Engels and Erwig. Their class descriptions can also be counted as an external,
structured intention model, which is referenced by an interpretation mapping.
We will have to study the relation to our work more closely, but it seems that
their work suffers the same computational bias as the spreadsheets themselves.
But classes with extensive documentation or UML diagrams might go some ways
towards generating a help system like SACHS.
In the future we plan to extend the repertoire of semantic services of the
SACHS system. For instance, we envision a dual service to the one in Figure 3
which could visualize where the value of a cell is used to get an intuition for the
relevance of a cell value.
At the moment, our approach presupposes that the spreadsheet author documents the spreadsheet intention as an OMDoc document and also annotates the
spreadsheet with interpretation information. Both tasks place a heavy burden on
the author and currently restrict our approach to mission-critical or extremely
complex spreadsheets. But we expect the burden of specifying the ontology to
decrease as more and more OMDoc content dictionaries for common models
(e.g. standard accounting techniques) appear. For the interpretations, we plan
to adapt techniques of header, unit, and template inference [AE04,AE06] to
partially automate the annotation process via suggestions.
But future extensions are not limited to spreadsheets. Note that the semantic analysis in Section 2 is largely independent of the computational information
that is at the heart of spreadsheets. The semantic information we are after only
pertains to the use of data grids as a user interface to complex functions. In
the future we plan to generalize the intention model architecture presented in
Section 3 to the case of data grids — e.g. tables of experimental results or raster
data from satellite images. Moreover, we want to develop a native infrastructure
for representing “data grids” as a user interaction feature in OMDoc: In accordance with the format, the provenance and interpretation functionality would
allow to link grid presentations to their intention without leaving the language.
OMDoc-based systems could then pick up the semantics and offer complex interactions based on them. This would lead to much more general active documents.
We could even turn the approach presented in this paper around and generate
SACHS spreadsheets from OMDoc documents as active documents.
As a theoretical extension, we conjecture that the interpretation mappings
can actually be seen as logic morphisms [RK08], which would allow us to re-use
much of the higher-level OMDoc functionalities based on these.
Finally, we admit that we are still at the very beginning of understanding
the repercussions of viewing spreadsheets as active and semantic documents. We
198
were quite surprised at the depth of meaning in even our very simple example
and are looking forward to further spelunking.
References
ABC+ 03.
AE04.
AE06.
ASK+ 04.
BCC+ 04.
BF79.
BSRR02.
Cla92.
CoF04.
EE05.
Fen08.
GS99.
HKS06.
David Carlisle, Patrick Ion, Robert Miner, et al Mathematical Markup
Language (MathML) version 2.0 (second edition). W3C recommendation,
World Wide Web Consortium, 2003. http://www.w3.org/TR/MathML2.
Robin Abraham and Martin Erwig. Header and unit inference for spreadsheets through spatial analysis. In IEEE International Symposium on Visual Languages and Human-Centric Computing, pages 165–172, 2004.
Robin Abraham and Martin Erwig. Inferring templates from spreadsheets.
In ICSE ’06: Proceedings of the 28th international conference on Software
engineering, pages 182–191, New York, NY, USA, 2006. ACM. doi:http:
//doi.acm.org/10.1145/1134285.1134312.
T. Antoniu, P.A. Steckler, S. Krischnamurthi, E. Neuwirth, and
M. Felleisen. Validating the unit correctness of spreasheet programs. In
26th IEEE International Conference on Software Engineering, pages 439–
448, 2004.
Stephen Buswell, Olga Caprotti, David P. Carlisle, Michael C. Dewar, Marc
Gaetano, and Michael Kohlhase. The Open Math standard, version 2.0.
Technical report, The Open Math Society, 2004. Available from: http:
//www.openmath.org/standard/om20.
Dan Bricklin and Bob Frankston. VisiCalc: Information from its creators,
Dan Bricklin and Bob Frankston. http://www.bricklin.com/visicalc.
htm, 1979.
M. M. Burnett, A. Sheretov, B. Ren, and G. Rothermel. Testing homogenous spreadsheet grids with the “what you see is what you test” methodology. IEEE Transactions on Software Engineering, 29(6):576–594, 2002.
J. Claerbout. Electronic documents give reproducible research a new meaning. pages 601–604, 1992.
CoFI (The Common Framework Initiative). Casl Reference Manual. LNCS
2960 (IFIP Series). Springer, 2004.
Gregor Engels and Martin Erwig. ClassSheets: Automatic generation
of spreadsheet applications from object oriented specifications. In 20th
IEEE/ACM International Conference on Automated Seofware Engineering, pages 124–155. IEEE Computer Society, 2005.
Dieter Fensel. Foreword. In Martin Hepp, Pieter De Leenheer, Aldo
de Moor, and York Sure, editors, Ontology Management: Semantic Web,
Semantic Web Services, and Business Applications, Semantic Web and beyond: Computing for Human Experience, pages 9–11. Springer Verlag, 2008.
Brian R. Gaines and Mildred L. G. Shaw. Enbedding formal knowledge
models in active documents; creating problem-solving documents. Communications of the ACM, 42(1):57–63, 1999.
Eberhard Hilf, Michael Kohlhase, and Heinrich Stamerjohanns. Capturing the content of physics: Systems, observables, and experiments. In
Jon Borwein and William M. Farmer, editors, Mathematical Knowledge
Management, MKM’06, number 4108 in LNAI, pages 165–178. Springer
Verlag, 2006. Available from: http://kwarc.info/kohlhase/papers/
mkm06physml.pdf.
199
HM00.
E. Heinrich and H. Maurer. Active documents: Concept, implementation,
and applications. Journal of Universal Computer Science, 6(12):1197–1202,
2000.
Koh05a.
Andrea Kohlhase. Cpoint, 2005. http://kwarc.info/projects/CPoint/.
Koh05b.
Andrea Kohlhase. Overcoming Proprietary Hurdles: CPoint as Invasive
Editor. In Fred de Vries, Graham Attwell, Raymond Elferink, and Alexandra Tödt, editors, Open Source for Education in Europe: Research and
Practise, pages 51–56, Heerlen, The Netherlands, November 2005. Open
Universiteit Nederland, Open Universiteit Nederland. Proceedings at
http://hdl.handle.net/1820/483.
Koh06.
Michael Kohlhase. OMDoc – An open markup format for mathematical
documents [Version 1.2]. Number 4180 in LNAI. Springer Verlag, 2006.
Available from: http://omdoc.org/omdoc1.2.pdf.
MGM+ 08. Luc Moreau, Paul Groth, Simon Miles, Javier Vazquez, John Ibbotson,
Sheng Jiang, Steve Munroe, Omer Rana, Andreas Schreiber, Victor Tan,
and Laszlo Varga. The provenance of electronic data. Communications of
the ACM, 51(4):52–58, 2008. doi:http://doi.acm.org/10.1145/1330311.
1330323.
MvH04.
Deborah L. McGuinness and Frank van Harmelen. OWL web ontology
language overview. W3C recommendation, W3C, February 2004. Available
from: http://www.w3.org/TR/2004/REC-owl-features-20040210/.
MW07.
Normen Müller and Marc Wagner. Towards Improving Interactive Mathematical Authoring by Ontology-driven Management of Change. In Alexander Hinneburg, editor, Wissens- und Erfahrungsmanagement LWA (Lernen, Wissensentdeckung und Adaptivität) conference proceedings, pages
289–295, 2007.
RK08.
Florian Rabe and Michael Kohlhase. A web-scalable module system for
mathematical theories. Manuscript, to be submitted to the Journal of
Symbolic Computation, 2008. Available from: https://svn.kwarc.info/
repos/kwarc/rabe/Papers/omdoc-spec/paper.pdf.
UCI+ 06.
Victoria Uren, Philipp Cimiano, José Iria, Siegfried Handschuh, Maria
Vargas-Vera, Enrico Motta, and Fabio Ciravegna. Semantic annotation for
knowledge management: Requirements and a state of the art. Web Semantics: Science, Services, and Agents on the World Wide Web, 4(1):14–28,
Jan. 2006.
W3C08.
Rdfa primer. http://www.w3.org/TR/xhtml-rdfa-primer/, 2008. Available from: http://www.w3.org/TR/xhtml-rdfa-primer/.
Win06.
Terry Winograd. The spreadsheet. In Terry Winograd, John Bennett,
Laura de Young, and Bradley Hartfield, editors, Bringing Design to Software, pages 228–231. Addison-Wesley, 1996 (2006).
Wol02.
Stephen Wolfram. The Mathematica Book. Cambridge University Press,
2002.
200
Andrea Kohlhase holds a diploma in Mathematics
from the Technical University Berlin. After graduation she spent ten years in the software industry, but
came back to academia via a research programmership at Carnegie Mellon University, Pittsburgh. She
has finished her PhD at University Bremen in 2008
with Bernd Krieg-Brückner as one of her advisors.
Currently she works as a research associate at the
Bremen Lab “Safe and Secure Cognitive Systems” of
the German Center for Artificial Intelligence (DFKI)
in Bremen, Germany. Her research interest lies in the
field of Human-Computer Interaction, especially Interaction Design with a focus on applications using semantic data. She has developed several software packages offering invasive, semantic applications for MS
Office products, e.g., “CPoint” for PowerPoint or the
semantic help system “SACHS” for Excel.
Michael Kohlhase Michael Kohlhase is Professor
for Computer Science at Jacobs University, Bremen
and Vice Director at the German Research Centre
for Artificial Intelligence (DFKI), Bremen. His current research interests include Automated Theorem
Proving and Knowledge Representation for Mathematics, inference-based techniques for Natural Language Processing, and computer-supported Education.
He has pursued these interests during extended visits to Carnegie Mellon University, SRI International,
Rice University, and the Universities of Amsterdam
and Edinburgh. Michael Kohlhase is recipient of the
dissertation award of the Association of German Artificial Intelligence Institutes (1995) and of a Heisenberg stipend of the German Research Council (DFG
2000-2003). He was a member of the Special Research
Action 378 (Resource-Adaptive Cognitive Processes),
leading projects on both Automated Theorem Proving and Computational Linguistics. Michael Kohlhase
is trustee of the Interest Group Mathematical Knowledge Management, a member of the W3C Math working group and the president of the OpenMath Society
(since 2007).
Formal Management of CAD/CAM Processes
Michael Kohlhase, Johannes Lemburg, Lutz Schröder, and Ewaryst Schulz
DFKI Bremen, Germany
<firstname>.<lastname>@dfki.de
Abstract. Systematic engineering design processes have many commonalities with software engineering, with CAx objects taking the place
of program code. They are, however, currently considerably less formal.
We propose to draw on the mentioned similarities and transfer methods
from software engineering to engineering design in order to enhance in
particular the reliability and reusability of engineering processes. Concretely, we formulate a document-oriented design process that integrates
CAx documents with requirement specifications, and show how to interface a CAD system with program verification workflows, thus allowing for completely formalised development strands within a semi-formal
methodology.
1
Introduction
The engineering design process guides the development of products by describing
methods for solving problems on different levels of abstraction. The product
that comes into being will be described by the requirements list, the function
structure, the priciple solution, and its embodiment design. Within these steps
the degree of concreteness of the solution and the amount of detail increases. The
elements of these forms of representation stand for system elements, which are
linked by dependencies within and across their different levels of abstraction. The
present state of the art regarding CAx object representations does not support
this crosslinking of dependencies. Therefore, e.g. non-embodied priciple solution
knowledge is often shared and stored in the form of hand-made sketches and
oral explanations. In other words, large parts of the engineering process are not
completely representable by embodiment focused CAx systems.
Contrastingly, software engineers have long acknowledged the need for a formal representation of the entire software development process. In particular,
formal specification and verification of software and hardware systems is without alternative in safety-critical or security areas where one cannot take the risk
of failure. Formal method success stories include the verification of the Pentium
IV arithmetic, the Traffic Collision Avoidance System TCAS, and various security protocols. In many cases, only the use of logic-based techniques has been
able to reveal serious bugs in software and hardware systems; in other cases,
spectacular and costly failures such as the loss of the Mars Climate Orbiter
could have been avoided by formal techniques. Norms such as IEC 61508 make
the use of formal methods mandatory for software of the highest safety integrity
intro.tex
432
2009-01-05
16:04:14Z
lutz
202
level (SIL 3). Thus, formal methods will form an integral part of any systematic
methodology for safe system design. The main goal of the present work is to outline how formal methods, hitherto used predominantly in areas such as software
development and circuit design that are inherently dominated by logic-oriented
thinking anyway, can be transferred to the domain of computer-aided design
and manufacture of industrial artifacts (CAx), which is more closely tied to the
physical world1 . Potential benefits for the CAx process include
–
–
–
–
formal verification of physical properties of the objects designed
tracing of (formalised) requirements across the development process
improved control over the coherence of designs
semantically founded change management.
Concretely we want to associate the CAD documents with specifications of their
goals, constraints, manufacturing information, etc. The envisioned extended design process is very similar to the semi-formal software engineering process, where
the software artifact under development is accompanied with and cross-linked to
a series of specification documents that describe its intended purpose and meaning with respect to a background theory that is also provided in the collection
of specification documents.
It is crucial to keep the agile spirit of the robotics component design process,
therefore we need to embed the specification processes into the CAD system
using “invasive technology” [5]. On the other hand, we need to create a formal background theory for the geometric CAD objects in the form of CASL
theories [1].
2
A Document-oriented Process for CAx
The best practices for designing technical artifacts are typically standardized
by the professional societies of engineers to ensure quality. In our exposition
here, we will follow the German VDI 2221 [12], which postulates that the design
process proceeds in well-defined phases, refining an initial idea to a fully specified
product documentation in steps.
We observe that the process is similar to the software engineering process and
that the stages in the design process result in specification documents, as they
are e.g. found in the V-model. In contrast to software engineering approaches
like the V-model, the VDI 2221 (and current practices in engineering design) do
not specify processes to ensure consistency between the design stages or methods
to verify that products actually meet requirements from earlier specifications.
1
caddocs.tex
432
2009-01-05
16:04:14Z
We primarily concern ourselves with constructive aspects of the design process — as
opposed to the aesthetic aspects. The English language does not supply an obvious
label for this aspect of the design process as the term “construction” is already taken
and refers to constructing buildings. We will use the term “systematic engineering
design” to be synonymous to the German field of “Konstruktionslehre” which studies
constructive aspects of design processing in mechanical engineering. The details of
the process are described in [12], we build on [8] for intuitions.
lutz
203
In fact, the VDI 2221 process corresponds only to the left shank of the process
depicted in Figure 1, the quality control process (the right shank in Figure 1 and
the main contribution of the V-model) is left unspecified.
Even in software-engineering methods like the V-model, the quality control
processes are only insufficiently supported by machines, leading to high levels of
personnel costs and human errors. Therefore we propose a process where the design documents are semantically annotated, making relations explicit that were
only implicit before and using this information for extended machine support.
Fig. 1. The V-model of Software Engineering
2.1
The Engineering Design Process
Let us now take a more concrete look into the CAx-based design process. To
make the correspondence between VDI 2221 and the V-model explicit we review
the six2 stages and relate them to the V-model before we illustrate them with a
simple example.
S1 Purpose/Problem: a concise formulation of the purpose of the product to
be designed.
S2 Requirements List: a list of explicit named properties of the envisioned
product. It is developed in cooperation between designer and client and
corresponds to the “User Specification” document in the V-model.
S3 Functional Structure: A document that identifies the functional components of the envisioned product and puts them into relation with each other.
S4 Solution in Principle: a specification of the most salient aspects of the
design. It can either be a CAD design like the one in Figure 4 below or a
hand drawing [8].
S5 Embodiment Design/“Gestalt”: a CAD design which specifies the exact
shape of the finished product.
2
caddocs.tex
432
2009-01-05
In fact, [12] specifies separate stages for determining modular structures and developing their embodiments, which we will subsume in steps S3 and S5
16:04:14Z
lutz
204
S6 Documentation: accompanies all steps of the design process.
Note that most of these design steps result in informal text documents, with
step S5 being the notable exception. In the document-oriented engineering design process we will concentrate on the documents, enhance them with semantic
annotations and link them to background specifications to enable machine support: e.g. requirements tracing, management of change, or verification of physical
properties. Before we do this let us set up our example by considering a rational
reconstruction of the design process of a machinists’ hammer according to DIN
1041.
2.2
The Design of a Hammer
We will now exemplify the stages of the design process of a machinists’ hammer,
as this is a tool familiar to most readers. As the standardization process for hammers is at a very advanced stage, so that most product features are completely
determined by industrial standards, we will present a rational reconstruction of
the design process.
Ideally, the list of requirements of a product should be unambiguous, clear
and complete. This ideal goal is seldom fulfilled in a real life design process. This
is due e.g. to implicit customer wishes, which in fact often are more important
to the market-success of a product than the well-known and explicitly named
requirements.
Explicit Hammer Requirements
E1 The hammer has to fulfill the standard DIN 1041 and all related subsequent
standards, namely: DIN 1193, DIN 1195, DIN 5111, DIN 68340 and DIN
ISO 2786-1.
E2 The handle has to be painted with a clear lacquer over all and with colour
RAL 7011 (iron grey) at 10 cm from the lower end.
E3 Two company logos of 20mm length are to be placed on both sides of the
handle.
Implicit Hammer Requirements In our example this could be:
I1
I2
I3
I4
The
The
The
The
hammer
hammer
hammer
hammer
must be usable for right-handed and left-handed persons.
should be ergonomic.
must fit into a standard tool chest.
shall look strong and matter-of-fact.
The Purpose of a Hammer The first and most important step in setting up a
requirements list is the specification of the purpose of the product. The purpose
describes the intended use of the product solution-neutrally. This is the highest
level of abstraction used within the design process. In the case of a hammering
tool the purpose can be in the form of a very simple definition:
caddocs.tex
432
2009-01-05
16:04:14Z
lutz
205
Fig. 2. Engineering standards affecting machinists’ hammers
A hammer is an apparatus for transmitting an impulse to an object.
Within the scope of a hand-tool in contrast to e.g. a hammer mill the purpose
can be narrowed to:
A hammer is an apparatus for the manual generation and transmission
a defined impulse to an object.
Functional Specification of a Hammer Within the design process, the functional specification is done by setting up the function structure that breaks down
the complex problem into managable smaller subfunctions and represents the
logical interrelationships of the given tasks. As in the design step of naming the
purpose, the function structure is still solution neutral. The aim is to open up a
caddocs.tex
432
2009-01-05
16:04:14Z
lutz
206
preferably broad solution field, which will be narrowed by explicit named criteria
within further steps of the design process.
The function structure as a tool to develop new technical solutions is able
to explain interrelationships within the future embodied product; therefore the
connection between function structure and the given product has to be clear.
Every subfunction can be found within the product, or the product is not a
suitable solution. On the other hand, function structures are not appropriate
as a tool for reverse engineering, because the relation between the embodied
product and the underlying functions is ambiguous.
One possible function structure for the hammer as an apparatus for the
manual generation and transmission of a defined impulse to an object can look
as in Figure 3.
Fig. 3. function structure of a hammer
Fig. 4. A principle solution for a Hammer
The Hammer Principle From the functional specification, we develop a principle solution, which can be found in Figure 4. This solution abstracts from the
caddocs.tex
432
2009-01-05
16:04:14Z
lutz
207
physical traits of the eventual product and identifies the functional parts. For
a hammer, one of these is the handle, here a cylindrical part of the hammer
shaft used for gripping. The fact that it is symmetric/cylindrical is a response
to the requirement E1. The handle is connected to an inert mass (depicted by a
solid ball in Figure 4) which is again connected to an active surface that delivers
the impact on the object. The size and form of the active surface will be determined by the requirement E2. In fact, the principle solution reveals that there
is a second possible active area of the hammer, opposite to the primary one;
Figure 4 shows three variants of the principle solution with differing secondary
active surfaces.
The Embodiment of a Hammer Note that the principle solution is not
a finished design yet, since it abstracts from most of the physical traits of a
hammer, e.g. the dimensions of the shaft and the form of the head, which will be
specified in the embodiment design step. Here, the ultimate three-dimensional
shape and the materials of the product are derived, taking into account material
properties, manufacturability constraints, and aesthetic factors. These can lead
to widely differing final designs (see 5 for examples).
Fig. 5. Embodiments of the Hammer Principle
2.3
A Document-Oriented Design Process
We propose an engineering design process which concentrates on the design documents, semantically enhancing them and linking them to semi-formal specifications, so that quality control can be machine supported. It is crucial to note that
the various design documents necessarily have differing levels of rigour, ranging
from informal and hard-to-quantify requirements like E2 to mathematical proofs
of security-relevant properties e.g. in aerospace applications. Additionally, different product parts and aspects underly differing economic and security-related
caddocs.tex
432
2009-01-05
16:04:14Z
lutz
208
constraints, so that design quality control3 must be supported at various levels of formality. As a consequence we need to encode the design documents in
document formats that support flexible degrees of formality. Therefore we chose
OMDoc (Open Mathematical Documents [6]) as the basis of our work. The
format concentrates on structural aspects of the knowledge embedded in documents and provides a markup infrastructure to make it explicit by annotation.
Crucially for our project, the format supports a fine-granular mixture of formal
and informal elements and thus supports the stepwise migration e.g. from informal user requirements to specifications expressed in formal logics that can be
exported to program verification environments like our HeTS system.
We postulate that most of the design documents only need to be semi-formal
for our design method to be effective. Practical semiformalization in OMDoc is
scaffolded by establishing “content dictionaries” (CDs), i.e. semi-formal, but
explicitly structured documents that specify the meaning of the terms used
in design documents. Incidentally, CDs are also convenient for structuring the
background knowledge involved in engineering design. We call a document semiformal if the relevant meaning-giving concepts are annotated by references to
their CDs, and all defined concepts, model assumptions, and postulates are
marked up as such using an OMDoc-compatible vocabulary.
The OMDoc format provides a strong, web-scalable infrastructure for inheritance and modularization of CD-based knowledge [10] and supports practical
knowledge management processes with a rich tool set [4,7,3,9]. In software engineering and mathematics, the additional effort involved in upgrading design
documents to the semi-formal level is significantly lower than the initial effort
of creating these documents once a basic set of CDs has been established for a
domain. In any case, the semiformalization effort is an order of magnitude lower
than fully formalizing them. We anticipate that this relation will not change by
the generalization of the OMDoc to design documents.
Semi-formal design documents already allow machine support for many common quality control tasks like requirements tracing and management of change
which are based on an explicitly given dependency relation (see [2] for details).
3
Invading SolidWorks
In this chapter we illustrate our formal/semi-formal methodology in design processes by means of an integration of formal methods tools into the widely used
CAD system SolidWorks [11]. The latter serves mainly as a demonstration
platform; our overall approach is sufficiently general to apply equally well to any
other CAD system that provides suitable interfaces.
To obtain a formal representation of CAD documents, we provide a SolidWorks plugin that extracts the constructed objects in the shape of the design
3
solidworks.tex
430
2009-01-05
In our deliberations it is important to distinguish the concept of “quality of design”
from the quality assurance in the production process. We are only concerned with
the former in this paper: our methods are inspired by the software design process,
where production quality problems can largely be ignored.
12:25:12Z
lutz
209
tree, which consists of a tree of objects and constraints relating these objects.
These data are obtained using the SolidWorks API, and are output as an
OMDoc file. This file consists of constants, defined by terms, each representing
a design tree, and of axioms stating the given constraints. Formally, a design tree
is a labeled tree where each node represents a SolidWorks object stemming
from an action or base constructor, recorded as the label of this node, applied
to the SolidWorks objects which are represented by the children of the node.
The representation format of design trees is defined in a content dictionary
containing all actions and base constructors of SolidWorks as constants and
function symbols accessing abstract data types (see Listing 1.1 for a code snippet of this content dictionary in OMDoc format). We base the SolidWorks
content dictionary on the theory of real numbers. The main datatype is then the
type SWObject of objects; these may e.g. be points with 3 real coordinates, arcs,
i.e. sectors of circles given by a start and an end point, a center point, and an
orientation (in the shape of a normal vector), or planes, i.e. 2-dimensional affine
subspaces defined by their normal vector and a point. The type of objects is a
recursive data type — e.g. an extrusion (see Listing 1.1) represents the figure
which results as the space covered by an oriented base object when moved along
or against its orientation (as specified by a direction parameter ±1) for a given
distance.
Listing 1.1. The SolidWorks content dictionary
1
6
11
16
21
<theory xml:id=”solidworks”>
<imports from=”Logic/casl.omdoc” type=”global”/>
...
<adt xml:id=”solidworks−object−adt”>
...
<sortdef name=”SWObject” type=”free”>
<constructor name=”Extrusion”>
<argument>
<type><OMOBJ><OMS cd=’solidworks’ name=”SWObject”/></OMOBJ></type>
<selector name=”From” total=”yes”/>
</argument>
<argument>
<type><OMOBJ><OMS cd=’Reals’ name=”Real”/></OMOBJ></type>
<selector name=”Length” total=”yes”/>
</argument>
<argument>
<type><OMOBJ><OMS cd=’Reals’ name=”Real”/></OMOBJ></type>
<selector name=”ExDirection” total=”yes”/>
</argument>
</constructor>
</sortdef>
</adt>
...
</theory>
In the case study of the next section we will show a concrete example which
illustrates the use of the plugin and how this integrates into the whole development process. The case study is mainly concerned with the verification of
solidworks.tex
430
2009-01-05
12:25:12Z
lutz
210
designs against abstract requirements. Further potential uses of the invasive approach include semantic preloading, i.e. automated rapid prototyping of designs
from abstract specifications, as well as requirements tracing and a closer general integration of specifications and designs e.g. by user-accessible links between
specifications and parts in the SolidWorks design.
4
Case Study: Simple Geometric Objects
One would naively imagine that there is really nothing to verify about a geometric object: a cylinder is a cylinder is a cylinder. But as soon as one starts
using a real CAD system, it becomes clear that the situation is actually more
complex. The mathematical concept of a three-dimensional geometric object is
a set of points in three-dimensional euclidean space, typically described by a
general formation principle and a number of parameters. E.g. in the case of a
(solid) cylinder, the parameters are
–
–
–
–
the
the
the
the
coordinates of some anchor point, say the centre of the cylinder,
spatial direction of the axis,
height h and
radius r,
and the formation principle for cylinders prescribes that these parameters describe the set of points p such that
– p has distance at most r from the axis (regarded as an infinite straight line)
and
– the orthogonal projection of p onto the axis has distance at most h/2 from
the centre point.
On the other hand, the description of a geometric object than can be extracted
from a CAD document will in general take a totally different form: instead of
a point set or the above-mentioned parameters, a CAD system will store the
process involved in drawing an object as a set of so-called features, which are
moreover subject to dependencies (these correspond to what a computer scientist
would call constraints, a term that however is used differently in the context
of CAD). Here, a feature is a very basic, typically one- or two-dimensional,
geometric component which may refer in various ways to other components.
E.g. the process of drawing a cylinder would typically look as follows.
– Create a two-dimensional plane.
– Insert a circle into the plane, described as a circular arc with coincident start
and end points and marked as filled.
– Extrude the circle orthogonally to the plane.
The cylinder is thus anchored in one feature, the plane, and itself consists of
two features, the circle and the extrusion that refers to it. Note that a number
of variants are possible, e.g. one might choose to draw an empty circle, extrude
casestudy.tex
432
2009-01-05
16:04:14Z
lutz
211
spec Vectors3DBase =
RealBase
then free type Vector ::= V (C1 : Real ; C2 : Real ; C3 : Real )
ops 0 : Vector = V (0, 0, 0);
..
.
end
spec Vectors3D = . . .
then ops
∗ (x, y : Vector ) : Real =
C1 x ∗ C1 y + C2 x ∗ C2 y + C3 x ∗ C3 y;
%(scalar product)%
# (x, y : Vector ) : Vector = . . . ;
%(vector product)%
| |(x : Vector ) : Real = sqrt (x ∗ x );
..
.
end
Fig. 6. Specification of the basic geometry
it, and then fill the resulting tube. We shall generally refer to a combination of
features as described above as a concrete design, while a definition via mathematical point sets will be called an abstract design. While in the above case it is
easy to see intuitively that the concrete design matches the abstract design, i.e.
that extruding a circle really yields a cylinder, the formalisation of this intuition
is by no means an entirely trivial enterprise, and more complex objects quickly
lead to quite challenging verification tasks – imagine e.g. having to check that
two given circular extrusions of two circles yield two interlocking chain links.
Additional complexity arises from the above-mentioned dependencies – e.g. one
may initially leave the height of the cylinder open, cut part of the cylinder off
using a skewed plane placed at a defined angle to the base plane and touching
the perimeter of the bottom circle, and then impose that the height of the short
side of the arising cut cylinder is half that of the long side, thus completely
determining the height.
It is therefore desirable to have machine support for checking that an abstract
design is actually implemented by a given concrete design. Besides the mere fact
that one can verify geometric shapes, added benefits include
– Easy proofs of physical and geometric properties of the objects involved –
e.g. once one has matched the abstract cylinder to the concrete cylinder, one
can now prove on the abstract side (much more easily than on the concrete
side) that the cylinder adheres to a prescribed surface area, volume, or mass
(if the density is known).
– Better control over the cohesion and consistency of the design – e.g. if it
turns out that the design fails to match the abstract object, this may mean
that the designer has accidentally left extraneous degrees of freedom. Such
casestudy.tex
432
2009-01-05
16:04:14Z
lutz
212
spec BaseObjects =
AffineRealSpace3DWithSets
then ops Cylinder (offset : Point; r : Real ; axis : Vector )
: PointSet =
{ (λ x : Point
• ∃ l : Real
• ∃ y : Vector
• l isIn [ 0 ... 1 ] ∧ orth (y, axis) ∧ | y | ≤ r
∧ x = offset + l ∗ axis + y)
};
end
Fig. 7. The abstract specification of a cylinder
errors may later lead to blocked designs that cannot be completed due to
unsatisfiability of their dependencies, a notorious problem in computer-aided
construction; verification against abstract designs may help detecting such
errors at an early stage of the development process.
– The abstract design may in fact properly abstract from the concrete shape of
the final object, e.g. by leaving less relevant dimensions open (within certain
ranges) or omitting features that do not play a central role in the present
stage of the design process, thus providing for a property-centered approach
to evolutionary design.
The paradigm proposed in FormalSafe for machine-supported verification of designs within the Bremen heterogeneous tool set HeTS is, in the simplest case,
as follows.
– Specify the abstract design in HasCasl, using a library of three dimensional
space and standard geometric objects.
– Export a concrete design from the CAD system and import the resulting
OMDOC document as a HasCasl specification.
– Set up a refinement view in HasCasl that matches the components of the
abstract design to those of the concrete design.
– Discharge the proof obligations arising from the refinement in an automatic
or semi-automatic theorem prover connected to HasCasl via HeTS, typically in Isabelle/HOL.
Further semantic services provided by the connection between abstract and concrete designs within HeTS include semantic annotation and requirements tracing as discussed in Section 2. A more visionary potential application of abstract
designs is the automated derivation of concrete designs, i.e. rapid prototyping
by compilation of abstract designs into preliminary formally verified CAD documents.
casestudy.tex
432
2009-01-05
16:04:14Z
lutz
213
Fig. 8. A cylinder in SolidWorks
The above-mentioned basic libraries are designed as follows4 . The abstract
specification of three dimensional basic geometry is contained in a HasCasl
library which provides the fundamental types and objects such as the datatypes
Point and Vector for points and vectors in R3 , types for point sets and vector
sets and operations on these types (See Figure 6 for some selected specifications
of this library). We will use PointSet as type of geometric objects, and we understand as standard geometric object the set of point sets which is implicitly given
by the image of an operation with result type PointSet, the constructor of the
standard geometric object in question. To give an example, consider Figure 7:
We define here an operation Cylinder , the constructor for cylinders, and thus
obtain the set of cylinders as image of this operation where the arguments vary
over the domain Point × Real × Vector. On the other side we specify the concrete
SolidWorks objects as described in Section 3. As our abstract type for geometric objects is PointSet, we define the semantics of the SolidWorks objects
in terms of points and point sets: objects of type SWPoint are interpreted as
Point or Vector and objects of type SWObject as PointSet. A concrete specification exported from the SolidWorks project shown in Figure 8 can be seen
in Figure 9. As already mentioned we can observe the straight-forward encoding
of the design tree of the cylinder as a HasCasl term. Furthermore we see that
we linked the exported specification to a theory called SolidWorksCylByArcExtrusion. This and similar theories are collected in a common patterns HasCasl
library for SolidWorks and provide the “heart” of the future semi-formal de4
casestudy.tex
432
2009-01-05
All mentioned HasCasl specifications can be obtained over the web from:
https://svn-agbkb.informatik.uni-bremen.de/Hets-lib/trunk/HasCASL/Real3D/
16:04:14Z
lutz
214
spec Project1 = . . .
then ops arc1 : SWArc =
SWArc
(E1, SWPoint (0, 0, 0), SWPoint (1, 0, 0),
SWPoint (1, 0, 0));
cylinder1 : SWObject = Extrusion (Arc arc1, 3, 1);
end
view cylinder1IsCylinder :
SolidWorksCylByArcExtrusion to
{Project1
then ops direction : Direction = 1;
center : SWPoint = SWPoint (0, 0, 0);
boundarypoint : SWPoint = SWPoint (1, 0, 0);
height : Real = 3;
plane : SWPlane = E1 ;
arc : SWArc = arc1 ;
cylinder : SWObject = cylinder1 ;
}
end
Fig. 9. The concrete specification of a cylinder
sign framework presented in this work, as they link the formal output of the
CAD program to the abstract specification of geometry. The main task will be
to develop this library and to provide proof support to close semi-automatically
the asserted links.
5
Conclusion and Further Work
We have shown that systematic engineering design processes (as laid down e.g.
in VDI 2221) have many commonalities with software engineering. To transfer
methods from software engineering to engineering design we have to deal with the
fact that engineering design processes are considerably less formal and are geared
towards producing CAx objects instead of program code. We have formulated a
semi-formal, document-oriented design process that integrates CAx documents
and extended a commonly used CAD system with an interface for exporting
CAD objects to the HeTS system. Moreover, we have illustrated the kind of
proof obligations that may arise in completely formal parts of the development.
The work reported here is just a first step in an ambitious endeavor where we
want to rethink the systematic engineering design process as a whole.
Acknowledgements
The work reported here was supported by the FormalSafe project conducted by
DFKI Bremen and funded by the German Federal Ministry of Education and
concl.tex
432
2009-01-05
16:04:14Z
lutz
215
Research (FKZ 01IW07002). We gratefully acknowledge discussions with Bernd
Krieg-Brückner, Dieter Hutter , Christoph Lüth, and Till Mossakowski. Erwin
R. Catesbeiana has contributed his known opinions on fundamental matters of
logic.
References
1. E. Astesiano, M. Bidoit, B. Krieg-Brückner, H. Kirchner, P. D. Mosses, D. Sannella,
and A. Tarlecki. CASL – the common algebraic specification language. Theoretical
Computer Science, 286:153–196, 2002.
2. S. Autexier, D. Hutter, T. Mossakowski, and A. Schairer. Maya: Maintaining
structured documents. In OMDoc – An open markup format for mathematical
documents [Version 1.2] [6], chapter 26.12.
3. A. Ioniţă. Developing a rest interface to a database for omdoc. http://kwarc.
info/projects/ombase/pubs/report2.pdf, 2007.
4. JOMDoc Project.
5. A. Kohlhase and M. Kohlhase. CPoint: Dissolving the author’s dilemma. In A. Asperti, G. Bancerek, and A. Trybulec, eds., Mathematical Knowledge Management,
MKM’04, number 3119 in LNAI, pp. 175–189. Springer Verlag, 2004.
6. M. Kohlhase. OMDoc – An open markup format for mathematical documents
[Version 1.2]. Number 4180 in LNAI. Springer Verlag, 2006.
7. C. Lange. SWiM – a semantic wiki for mathematical knowledge management. In
S. Bechhofer, M. Hauswirth, J. Hoffmann, and M. Koubarakis, eds., ESWC, vol.
5021 of Lecture Notes in Computer Science, pp. 832–837. Springer, 2008.
8. J. P. Lemburg. Methodik der schrittweisen Gestaltsynthese. PhD thesis, Fakultät
für Maschinenwesen, RWTH Aachen, 2008.
9. Math WebSearch – a semantic search engine. http://betasearch.mathweb.org.
10. F. Rabe and M. Kohlhase. An exchange format for modular knowledge. In P. Rudnicki and G. Sutcliffe, eds., Knowledge Exchange: Automated Provers and Proof
Assistants (KEAPPA), Nov. 2008.
11. Introducing SolidWorks. SolidWorks Corporation, Concord, MA, 2002.
12. VDI-Gesellschaft Entwicklung Konstruktion Vertrieb. Methodik zum Entwickeln
und Konstruieren technischer Systeme und Produkte, 1995. English title: Systematic approach to the development and design of technical systems and products.
concl.tex
432
2009-01-05
16:04:14Z
lutz
216
Michael Kohlhase Michael Kohlhase is Professor
for Computer Science at Jacobs University, Bremen
and Vice Director at the German Research Centre
for Artificial Intelligence (DFKI), Bremen. His current research interests include Automated Theorem
Proving and Knowledge Representation for Mathematics, inference-based techniques for Natural Language Processing, and computer-supported Education.
He has pursued these interests during extended visits to Carnegie Mellon University, SRI International,
Rice University, and the Universities of Amsterdam
and Edinburgh. Michael Kohlhase is recipient of the
dissertation award of the Association of German Artificial Intelligence Institutes (1995) and of a Heisenberg stipend of the German Research Council (DFG
2000-2003). He was a member of the Special Research
Action 378 (Resource-Adaptive Cognitive Processes),
leading projects on both Automated Theorem Proving and Computational Linguistics. Michael Kohlhase
is trustee of the Interest Group Mathematical Knowledge Management, a member of the W3C Math working group and the president of the OpenMath Society
(since 2007).
Johannes Lemburg received his Dipl.-Ing. from the
RWTH Aachen University, Germany, in 2002. Up to
2008 he stayed at this university as a research assistant for the Chair and Institute for Engineering Design
led by Univ.-Prof. Dr.-Ing. J. Feldhusen. He joined the
DFKI Robotics Research Group in 2008 as a senior researcher to strengthen the mechanical design group.
Lutz Schröder received his Dipl.-Math. from the University of Bremen in 1996, his PhD in mathematics
from the University of Bremen in 1999, and his habilitation in computer science from the university of
Bremen in 2005. He held a substitute associate professorship in theoretical computer science from 2003
to 2006. He joined the DFKI safe and secure cognitive systems group in 2006 as a senior researcher.
His research interests include algebraic, higher order,
and reactive specification and verification, in particular coalgebraic modal logic. He received the EATCS
best paper award at ETAPS 2006 for a paper that laid
the ground for the algorithmic theory of coalgebraic
modal logic. He presently holds two DFG grants on
coalgebraic modal logic and, with Christoph Lüth, on
monad-based program logics, respectively.
Ewaryst Schulz received his M.Sc. degree in mathematical logic from University Jussieu (Paris VII),
France, in 2006, after having worked for several years
in internet industry where he developed the content management system ePilot. He joined the Omega
group in Saarbrücken at the end of 2006. Since June
2008 he works as a junior researcher in the FormalSafe
project with the German Research Center for Articial
Intelligence at the research group for Safe and Secure
Cognitive Systems, Bremen, Germany.
concl.tex
432
2009-01-05
16:04:14Z
lutz
Graph Transformation Units
to Model Model Transformation –
First Ideas
Hans-Jörg Kreowski
University of Bremen
[email protected]
No Institute Given
Abstract. In this paper, graph transformation units are temptatively
proposed as a core concept to model model transformations in a compositional framework. The very first ideas are sketched concerning the
sequential and parallel compositions of model transformations, their input and output types, and the possible notions of correctness.
1
Introduction
Computers are devices that can be used to solve all kinds of data-processing problems – at least in principle. The problems to be solved come from economy, production, administration, military, science, education, entertainment, and many
other areas. There is quite a gap between the problems as one has to face them
in reality and the solutions one has to provide so that they run on a computer.
Therefore, computerization is concerned with bridging this gap by transforming a
problem into a solution. Many efforts in computer science contribute to this need
of transformation. First of all, compilers are devices that transform programs in
a high-level language into programs in a low-level language where the latter are
nearer and more adapted to the computer than the former. The possibility and
success of compilers have fed the dream of transforming descriptions of dataprocessing problems automatically or at least systematically into solutions that
are given by smoothly running programs. This includes the request for corectness meaning that a resulting program solves the original problem completely
and precisely (and does not do anything else in addition). In, e.g., Basin and
Krieg-Brückner [BKB99], this key idea is discussed and investigated under the
heading of program development and program transformation. In recent years,
the term model transformation has become more popular.
In this paper, it is sketched how the concepts of graph transformation units
(cf., e.g., [KKS97,KK99,KKR08]) can be employed to model model transformation. Following the ideas of model-driven architecture (MDA)(cf., e.g., Frankel
[Fra03]), the aim of model transformation is to transform platform-independent
models (PIMs), which allow to describe problems adequately, into platformspecific models (PSMs), which run properly and smoothly on a computer. As
218
a typical description of the PIMs, one may use UML diagrams while PSMs are
often just JAVA programs. In addition, there is another source of inspiration:
model transformation in theoretical computer science – in formal language theory as well as in automata theory in particular. These areas provide a wealth
of transformations between various types of grammars and automata like, for
example, the transformation of nondeterministic finite automata into deterministic ones or of pushdown automata into context-free grammars (or the other
way round) or of arbitrary Chomsky grammars into the Pentonen normal form
(to give a less known example). While the models in the MDA context are often
diagrammatic or textual, the examples of theoretical computer science show that
models may also be tuples with components being sets of something.
Accordingly, graphs, tuples of models, and sets of models are introduced as
models in Section 3. In Section 2, graph transformation units are recalled as
devices to transform initial graphs into terminal graphs by means of regulated
rule application. They can be considered as model transformations if the initial
and terminal graphs represent models. Such units are used as building blocks
of the sequential and parallel compositions of model transformations in Section
4 while some variants of the notion of correctness of model transformations are
considered in Section 5.
2
Graph Transformation Units
Graph transformation units are rule-based devices to transform initial graphs
into terminal graphs by iterated rule applications which obey control conditions. The basic components are taken from an underlying graph transformation
approach as introduced in [KK99] . There are always a class of graphs and a class
of rules that can be applied to these graphs. In many cases, rule application is
highly nondeterministic – a property that is not desirable in any case. Hence, a
graph transformation approach can also provide a class of control conditions so
that the degree of nondeterminism of rule application can be reduced. Moreover,
graph class expressions can be used in order to specify sets of initial and terminal
graphs.
Formally, a graph transformation approach is a system A = (G, R, X , C) the
components of which are defined as follows.
– G is a class of graphs.
– R is a class of rules such that every r ∈ R specifies a binary relation on
graphs SEM (r) ⊆ G × G.
– X is a class of graph class expressions such that each x ∈ X specifies a set
of graphs SEM (x) ⊆ G.
– C is a class of control conditions such that each c ∈ C specifies a set of
sequences SEM (c) ⊆ G × G.
219
A well-known instance for the class G is the class of all directed edge-labeled
graphs. Other classes of graphs are trees, Petri nets, undirected graphs, hypergraphs, etc.
One encounters quite a variety of notions of rules and rule applications in the
literature (see, e.g., [Roz97] for the most typical examples). In all cases, a rule
r provides a left-hand side L and a right-hand side R, and its application to a
graph G involves a match of L in G (usually given by a graph morphism often
subject to further conditions). Then certain parts of the match are removed,
and R is added in some way. Let H be the resulting graph, then such a direct
derivation is denoted by G =⇒ H.
r
Every finite subset M ⊆ G is a graph class expression that specifies itself, i.e.
SEM (M ) = M . Moreover, every set L of labels specifies the class of all graphs
in G the labels of which are elements of L. Every set P ⊆ R of rules can also be
used as a graph class expression specifying the set of all graphs that are reduced
w.r.t. P where a graph is said to be reduced w.r.t. P if no rules of P can be
applied to the graph.
The least restrictive control condition is the term free that allows all graph
transformations, i.e. SEM (free) = G × G. Other typical examples are regular
expressions or priorities specified over a given rule set.
More about graph class expressions and control conditions can be found in
[Kus00a,Kus00b].
Now all the ingredients are ready to introduce graph transformation units.
A simple graph transformation unit is a system tu = (I, P, C, T ), where I and
T are graph class expressions to specify the initial and the terminal graphs
respectively, P is a set of rules, and C is a control condition.
Such a transformation unit tu specifies a binary relation SEM (tu) ⊆ SEM (I) ×
SEM (T ) that contains a pair (G, H) of graphs if and only if there is a derivation
∗
G =⇒ H permitted by C, i.e. (G, H) ∈ SEM (C).
P
It is worth noting that in general transformation units have an import component
that allows to use the semantic relations specified by other units as transformation steps [KK99]. But for the purpose of this paper, simple units are enough.
A unit tu = (I, P, C, T ) may be depicted by
I
P, C
T
A graph transformation can be considered as a model transformation if SEM (I)
and SEM (T ) specify classes of graph models. Typical examples of graph transformation units that model model transformations are the transformation of
220
well-structured flow diagrams into while-programs in [KHK06] and of UML sequence diagrams into collaboration diagrams in [CHK04]
Graph transformation units are used as basic building blocks to construct more
complex model transformations by sequential and parallel compositions in Section 4 while, in the following section, the classes of graph models to be transformed are specified accordingly.
3
Graph-based Models and Model Types
The starting point is the observation that many models used in Computer Science are of a graphical, diagrammatic, and visual nature and that they can be
represented as graphs in an adequate way in most cases. But models may not
occur only as singular items, but also as tuples or as some other collections of
models. To cover this, we define the class M of graph-based models in a recursive
way:
1. G ⊆ M,
2. (M1 , . . . , Mk ) ∈ M if M1 , . . . , Mk ∈ M for all k ≥ 2,
3. M ∈ M if M ⊆ M.
Point 1 makes sure that all graphs and – in this way – all diagrams with graph
representations are models. Point 2 allows one to consider a k−tuple of models
as a model and makes k models simultaneously available in this way. Point 3 also
makes many models available at the same time. But now they are not ordered.
The sets may be infinite so that they may serve as semantic models.
Usually, model transformations do not transform all models, but initial models
of a particular type into terminal models of another type. The models of a type
form a set of models. The following type-building operators are considered in
this paper:
1. a graph class expression X ∈ X with SEM (X) ⊆ G yields a type,
2. if L1 , . . . , Lk are types then L1 × . . . × Lk is a type.
As SEM (X) ⊆ G and G ⊆ M, one gets SEM (X) ⊆ M. As L1 ×. . .×Lk contains
tuples as elements, one gets L1 × . . . × Lk ⊆ M. Therefore, the constructions in
Points 1 and 2 are yielding types.
In addition we assume some standard types including BOOL• , N• , Σ • , and
(Σ ∗ )• . The dot means that Boolean values T and F , the natural numbers, the
symbols of the alphabet Σ, and the strings over Σ are represented as special
graphs:
221
T
a
F
,
,
0
+1
n
begin
a1
...
. . . +1
times
ak
end
for n ∈ N,
for a ∈ Σ,
for a1 . . . ak ∈ Σ ∗ , ai ∈ Σ.
It is easy to see, that types can be described by type expressions. The set of type
expressions T is recursively defined:
1. X ⊆ T ,
2. t1 , . . . , tk ∈ T implies t1 × . . . × tk ∈ T
with the obvious semantics
1. SEM (X) for X ∈ X as before,
2. SEM (t1 × . . . × tk ) = SEM (t1 ) × . . . × SEM (tk ).
A typical example of a graph model is the state graph of a finite state automaton. A tuple of such graphs represents the parallel composition of finite state
automata. A pair of a finite state automaton and a string graph represents a
configuration of an acception process. It is an initial configuration if the current
state is the initial one, it is a terminal configuration if the current state is a final
one and the string is empty. A context-free grammar is a model of a quite different form than a state graph. It is a 4-tuple of two finite sets of nonterminal and
terminal symbols respectively, a finite set of production each being a pair of a
nonterminal symbol and a string, and a nonterminal start symbol. Therefore, it
is a model due to the definition above using mainly tupling and set enumeration.
The introduced model types and type expressions are suitable to provide proper
input and output types to the compositional model transformations of the following section.
4
Sequential and Parallel Composition
As pointed out in Section 2, one may consider graph transformation unit as basic
model transformations that transform initial models into terminal ones. To get
a more general notion, one can close the basic transformations under sequential
and parallel compositions. This allows a nice separation between the syntactic
level of expressions and the semantic level of model transformation.
The set MT E of model transformation expressions is defined recursively:
1. graph transformation units are in MT E,
222
2. mte 1 , . . . , mte k ∈ MT E implies mte 1 ; . . . ; mte k ∈ MT E
(sequential composition),
3. mte 1 , . . . , mte k ∈ MT E implies mte 1 . . . mte k ∈ MT E
(parallel composition).
The relational input-output semantics of graph transformation units can be extended to model transformation expressions using the sequential composition of
relations and the product types:
1. SEM (tu) as defined in Section 2,
2. SEM (mte 1 ; . . . ; mte k ) = SEM (mte 1 ) ◦ . . . ◦ SEM (mte k ),
3. ((M1 , . . . , Mk ), (N1 , . . . , Nk )) ∈ SEM (mte 1 . . . mte k ) if and only if
(Mi , Ni ) ∈ SEM (mte i ).
To make this more transparent, one may assign an input type and an output
type to each model transformation expression. Then the relational semantics of
an expression turns out to be a binary relation between input and output types.
The input type in and the output type out are recursively defined on MT E:
1. in(tu) = SEM (I), out(tu) = SEM (T ),
2. in(mte 1 ; . . . ; mte k ) = in(mte 1 ), out(mte 1 ; . . . ; mte k ) = out(mte k ),
3. in(mte 1 . . . mte k ) = in(mte1 ) × . . . × in(mte k ) and out(mte 1 . . . mte k ) = out(mte1 ) × . . . × out(mte k ).
Due to these definitions, it is easy to see that the following observation holds:
SEM (mte) ⊆ in(mte) × out(mte).
The compositions can be quite intuitively depicted:
tr1 ; tr2
I1
T1
tr1
I2
tr2
T2
tr1 tr2
I1
tr1
T1
I1 × I2
T1 × T2
I2
tr2
T2
223
The formal language theory offers many examples of sequential compositions of
model transformations like the transformation of right-linear grammars into finite state automata followed by their transformation into deterministic automata
followed by the transformation of the latter into minimal automata.
A typical example of a parallel composition is given by the acception processes
of two finite state automata that run simultaneously. If they try to accept the
same input strings, this parallel composition simulates the product automaton
that accepts the intersection of the two accepted regular languages.
5
Correctness
The idea of correctness is that a model transformation is assumed to preserve
or reflect the meaning of the transformed model such that the meaning of an
input model and the meaning of a related output model coincide or correlate in
a meaningful way at least. If the meaning of the input models and of the output
models are given by model transformations into suitable semantic models, then
we have the following situation
I
sem I
tr
T
sem T
The ideal result would be that the meaning of the semantics of the input models
equals the meaning of the sequential composition of the model transformation
tr followed by the semantic construction of the output models:
SEM (sem I ) = SEM (tr; sem T ) = SEM (tr) ◦ SEM (sem T ).
But this is not realistic to expect if the semantic constructions are given independently from each other. Usually, the semantic models differ and need some
adaptation before they can coincide. There are principally four cases of adaptations:
1. The semantic models of the input side are adapted to the semantic models
of the output side.
2. The other way round.
3. The semantic models of the input and the output side are both adapted to
certain common reference models.
4. Or such reference models are adapted to the input and output sides.
224
If the adaptations are transformations between semantic models, we get the
following pictorial versions:
1.
adaptIT
2.
adaptT I
3.
adaptI
adaptT
4.
adaptI
adaptT
In all four cases, one can define the correctness of the original transformation
relative to the adaptations. The model transformation tr is a correct model
transformation with respect to sem I and sem T if
1.
2.
3.
4.
SEM (sem I ) ◦ SEM (adaptIT ) = SEM (tr) ◦ SEM (sem T ),
SEM (sem I ) = SEM (tr) ◦ SEM (sem T ) ◦ SEM (adaptT I ),
SEM (sem I ) ◦ SEM (adaptI ) = SEM (tr) ◦ SEM (sem T ) ◦ SEM (adaptT ),
if (M, N ) ∈ SEM (sem I ) and (M, N ) ∈ SEM (tr) ◦ SEM (sem T ), then there
is N̄ with (N̄ , N ) ∈ SEM (adapt I ) and (N̄ , N ) ∈ SEM (adapt T ).
All four notions of correctness can be depicted as follows.
225
tr
1.
sem I
sem T
=
adaptIT
tr
2.
sem I
sem T
=
adaptT I
tr
sem I
=
sem T
3.
adaptI
adaptT
226
tr
sem I
=
sem T
4.
adaptI
adaptT
If all adaptations are the identity, all four notions yield the ideal case of correctness.
6
Conclusion
In this paper, some first ideas for the modeling of model transformations are
sketched. Models are recursively defined as tuples and sets of models starting
with graphs (including some graph-represented standard types). Models of this
kind cover graphical models like UML diagrams as well as set-theoretic models
like grammars and automata. The transformations of such models are recursively
given by the sequential and parallel compositions of model transformations starting with graph transformation units.
To shed more light on the significance of this temptative novel approach, more
work is needed including the following topics.
1. The parallel composition combines model transformations in such a way
that they run independently of each other on their respective components
of the tuples of models. But this is not always desirable as one may like
to interrelate transformation steps. Therefore, one may like to employ control conditions on the level of parallel compositions as they are available
on the basic level of graph transformation units. In [KKK05], the parallel
composition of units is studied as a step in this direction.
2. The variants of the notion of correctness are introduced for a single model
transformation. One interesting question is whether and how these correctness notions are compatible with the sequential and parallel compositions
so that the correctness of the components yields the correctness of the composed model transformation. Another question is how a proof theory for
model transformations may look like.
227
3. Explicit and detailed examples are needed to illustrate all introduced concepts properly.
References
[BKB99] David Basin and Bernd Krieg-Brückner. Formalization of the development process. In Egidio Astesiano, Hans-Jörg Kreowski, and Bernd KriegBrückner, editors, Algebraic Foundations of Systems Specification, chapter 14,
pages 521–562. Springer, 1999.
[CHK04] Björn Cordes, Karsten Hölscher, and Hans-Jörg Kreowski. UML interaction diagrams: Correct translation of sequence diagrams into collaboration
diagrams. In John l. Pfaltz, Manfred Nagl, and Boris Böhlen, editors, Applications of Graph Transformations with Industrial Relevance (AGTIVE 2003),
volume 3062 of Lecture Notes in Computer Science, pages 275–291, 2004.
[Fra03]
David S. Frankel. Model Driven Architecture. Applying MDA to Enterprise
Computing. Wiley, Indianapolis, Indiana, 2003.
[KHK06] Hans-Jörg Kreowski, Karsten Hölscher, and Peter Knirsch. Semantics of
visual models in a rule-based setting. In R. Heckel, editor, Proceedings of
the School of SegraVis Research Training Network on Foundations of Visual
Modelling Techniques (FoVMT 2004), volume 148 of Electronic Notes in Theoretical Computer Science, pages 75–88. Elsevier Science, 2006.
[KK99]
Hans-Jörg Kreowski and Sabine Kuske. Graph transformation units with
interleaving semantics. Formal Aspects of Computing, 11(6):690–723, 1999.
[KKK05] Renate Klempien-Hinrichs, Hans-Jörg Kreowski, and Sabine Kuske. Rulebased transformation of graphs and the product type. In Patrick van Bommel,
editor, Transformation of Knowledge, Information, and Data: Theory and
Applications, pages 29–51. Idea Group Publishing, Hershey, Pennsylvania,
USA, 2005.
[KKR08] Hans-Jörg Kreowski, Sabine Kuske, and Grzegorz Rozenberg. Graph transformation units – an overview. In P. Degano, R. De Nicola, and J. Meseguer,
editors, Concurrency, Graphs and Models, volume 5065 of Lecture Notes in
Computer Science, pages 57–75. Springer, 2008.
[KKS97] Hans-Jörg Kreowski, Sabine Kuske, and Andy Schürr. Nested graph transformation units. International Journal on Software Engineering and Knowledge
Engineering, 7(4):479–502, 1997.
[Kus00a] Sabine Kuske. More about control conditions for transformation units. In
Hartmut Ehrig, Gregor Engels, Hans-Jörg Kreowski, and Grzegorz Rozenberg, editors, Proc. Theory and Application of Graph Transformations, volume 1764 of Lecture Notes in Computer Science, pages 323–337, 2000.
[Kus00b] Sabine Kuske. Transformation Units—A Structuring Principle for Graph
Transformation Systems. PhD thesis, University of Bremen, 2000.
[Roz97] Grzegorz Rozenberg, editor. Handbook of Graph Grammars and Computing
by Graph Transformation, Vol. 1: Foundations. World Scientific, Singapore,
1997.
228
Hans-Jörg Kreowski is professor for theoretical computer science at the University of Bremen
since October 1982 while his colleague Bernd KriegBrückner was appointed even a month earlier. In all
the years until today, they have worked together and
undertaken many efforts at a prosperous development
of the computer science department although their
views of the right track were not always identical.
Concerning research, they share a common interest in
formal methods that led to various close cooperations
in national and international research projects and in
the IFIP Working Group 1.3 (Foundations of System
Specification). Hans-Jörg Kreowski enjoyed particularly the co-edition of the IFIP State-of-the-Art Report Algebraic Foundations of Systems Specification
with Egidio Astesiano and Bernd Krieg-Brückner.
Test Automation Problems and Their Reduction
to Bounded Model Checking
Jan Peleska
Department of Mathematics and Computer Science
University of Bremen
Germany
[email protected]
Abstract. In this paper we analyse a collection of test automation problems and describe the core concepts to their solution. It turns out that
all of them may be regarded as reachability problems of a finite nature.
This means that a finite number of applications of some sort of transition
relation in conjunction with a test-specific predicate completely characterises the solution set of test data to be constructed for the test objective
under consideration. This characterisation coincides with the notion of
a BMC instance which is the bounded model checking equivalent to the
SAT instances of Boolean model checking. All testing problems described
originate from the field of safety-critical systems verification and validation, where rigorous standards specify the amount of testing required in
order to achieve system certification.
1
Introduction
Objectives. The standards for avionic and railway control systems [16, 5] are quite
explicit with respect to dynamic testing of safety-critical systems in these domains: The whole testing process should be driven by requirements-based tests,
that is, tests establishing that a required system feature really has been implemented and works properly – at least for the situations covered by these
tests. Given formal requirements models of untimed or timed automata nature,
requirements-based tests can be made exhaustive in theory; this means that it is
possible to generate a comprehensive test suite from the model which is capable
to uncover every possible discrepancy between implementation and specification model [3, 18, 9]. The amount of test cases to be executed, however, would
exceed any acceptable time limit for the duration of the test suite execution,
even for systems of only moderate size and complexity. As a consequence it is
necessary to find reasonably well-founded heuristics how to decide when a test
suite may be regarded as “sufficiently effective”, though not exhaustive. To this
end, requirements-based testing is refined and complemented by the following
techniques:
1. Structural testing aims at covering all code portions, interfaces and synchronisation aspects of the implementation, while checking that the requirements
230
are still met. These tests are helpful if seemingly atomic requirements had
to be implemented by different components, such that it is impossible to
guess the different test cases required to cover all implementation aspects by
means of requirements analysis alone.
2. Equivalence class testing tries to structure the input data to the system under
test into different classes, such that it can be reasonably expected that the
implementation handles each member of the same class in a uniform manner.
As a consequence, only one or a very small number of test cases from the
same class have to be exercised on the system.
3. Boundary value testing helps to determine how many members of an equivalence class should be exercised on the system under test: At least the test
cases whose input data is located at the boundaries of each class have to be
executed.
4. Mutation testing gives a further answer to the question if a sufficient number
of test cases has been generated: The model is tainted with one or more errors
(this process is called error seeding) and then it is checked whether the test
suite is capable of uncovering the error. If not, additional test cases have to
be constructed.
Our objective in this paper is to provide a uniform formal approach to the
automated model-based construction of requirements-based, equivalence class,
boundary value and mutation test cases which can be practically applied for
the construction of the associated test data. It turns out that every test case
can be formulated as a reachability problem within the given model: We have
to reach a certain pre-state in the model from where the system feature under
investigation can be checked. Then a specific timed sequence of input vectors
has to be constructed which stimulates exactly the system reaction associated
with this feature. The system reaction can be automatically checked against the
model; therefore we only focus on the input data generation problem for the
rest of this paper. Each of the reachability problems associated with the kinds
of tests described above may be brought into a “format” that coincides with the
problem instances of bounded model checking (BMC).
Structural tests are outside the scope of this paper, because the reachability
problems associated with structural test case generation cannot be based on the
model, but have to be derived from the implementation code structure. The
reader is referred to [2] for further details.
Overview. Section 2 introduces the notion of BMC instances and discusses their
application is a timed automata context. In Section 3 a case study is presented
which is used to illustrate the main results about test case and test data generation and the associated reachability problems. These main results are described
in Section 4. Section 5 gives the conclusion and refers to related work and practical applications, as well as future work.
231
2
Bounded Model Checking
Given a model with transition relation Φ, bounded model checking (BMC) tries
to construct solutions of BMC instances, that is, predicates of the form
b=
c−1
j=0
Φ(i(k0 + j), s(k0 + j), s(k0 + j + 1))
(1)
∧ P (i(k0 ), s(k0 ), o(k0 ), . . . , i(k0 + c), s(k0 + c), o(k0 + c))
with
i(k) = (x1 (k), . . . , xm (k)) input vector at execution step t,
s(k) = (s1 (k), . . . , sn (k)) state vector at execution step t,
o(k) = λ(i(k), s(k)) output vector at execution step t,
Φ the transition relation relating pre-states s(k) and inputs i(k) at execution
step k to post states s(k + 1), and
– P a property about inputs, outputs and internal state to be fulfilled during
these c execution steps.
–
–
–
–
In classical BMC the input, output and state vectors consist of boolean components only. While this is appropriate for many problems related to the verification of electronic circuits, the application of BMC for general embedded systems
requires more general data types, such as integral numbers with large data ranges
and floating point data with their associated operations. While, in theory, these
data types and their operations could be re-coded as boolean vectors and operations, resulting again in classical BMC instances, it is often more efficient to
construct the solution space with methods (“theories”) optimised for the nonboolean data types. The tools for solving these mixed data type problems are
called Satisfiability Modulo Theories (SMT) solvers [15].
For discrete time models the execution steps are obviously induced by discrete
time ticks, as, for example, the clock cycles in electronic circuits. For dense-time
models, however, a further abstraction step is required, because the existence
of real-valued clocks induces an uncountable state space. As a consequence, the
meaning of “unfolding the transition relation for c execution steps” is not as
obvious as in untimed or discrete-time models. For models based on the semantics
of timed automata a suitable abstraction is given by the region or zone graphs [1,
p. 70]. Intuitively speaking, the zone graph abstracts timed states conforming to
the same constraints into a single node of the graph. Transitions between nodes
can either be triggered by inputs to the system or by the passage of time, when
boundaries given by clock conditions are reached. Suppose that the system will
reside in a node n of the zone graph in the time interval [t0 , t1 ] if no further
inputs are supplied, but leave n if some input x is provided during that interval.
Then placing x at time t0 +δ, 0 ≤ δ ≤ t1 −t0 will lead to the same system outputs
and state changes, regardless of the value of δ within the specified bounds. These
considerations lead to a specific structure of an abstracted transition relation:
– Timed transitions shift the time forward, but never beyond the next boundary induced by a clock condition in the zone graph. As a consequence, the
232
system never switches nodes during a timed transition. At the end of a timed
transition new inputs to the system may be placed.
– Zero-time transitions perform switches between nodes in the zone graph,
possibly changing state variables and outputs. They occur while inputs and
time remain fixed.
– Timed transitions can only occur if no zero-time transition is enabled. This
obviously requires some well-formedness conditions on the models, in order to avoid divergence (unbounded number of consecutive zero-time steps)
and Zeno sequences (unbounded number of timed transitions with strictly
decreasing time intervals between transitions).
3
Case Study
As an example illustrating the different types of test cases and their associated
representation as BMC instances the model of a fictive automotive controller is
presented in Figures 1 — 3. This controller manages the interplay between turn
indication and emergency flashing: Inputs to the controller are (1) the command
to switch on/off turn indications left or right, TurnIndLeft = 1/TurnIndLeft = 0 or
TurnIndRight = 1/TurnIndRight = 0, (2) the command to turn on/off emergency
flashing, EmerFlash = 1/EmerFlash = 0 and (3) the electrical current status in
percent, Current ∈ {0, . . . , 100}. Controller outputs act on the interfaces to the
left-hand side and right-hand side indication lights: FlashLeft = 1 and FlashRight
= 1 switch these lights on and they are switched off by setting the respective
outputs to zero.
Fig. 1. Components of the flash control unit.
The system requirements to be reflected in the model are given as follows:
R1 Indication lights are only active if the electrical current status is > 80%.
233
R2 If any lights are flashing, this is done synchronously with a 340ms ON –
220ms OFF period.
R3 An input change from (TurnIndLeft,TurnIndRight) ∈ {(0, 0), (0, 1)} to
(TurnIndLeft,TurnIndRight) = (1,0) switches indication lights left into flashing
mode and switches indication lights right off.
R4 An input change from (TurnIndLeft,TurnIndRight) ∈ {(0, 0), (1, 0)} to
(TurnIndLeft,TurnIndRight) = (0,1) switches indication lights right into flashing mode and switches indication lights left off.
R5 An input change from EmerFlash = 0 to EmerFlash = 1 switches indication
lights left and right into flashing mode, regardless of any previously activated
turn indication.
R6 Activation of the turn indication left or right overrides emergency flashing,
if this has been activated before.
R7 If turn indication left or right is switched of and emergency flashing is still
active, emergency flashing is continued or resumed, respectively.
R8 If emergency flashing is turned off and turn indication left or right is still
active, the turn indication is continued or resumed, respectively.
Fig. 2. Control component FLASH CTRL for the decisions “flash left” and “flash right”.
The controller model consists of two sub-components (Fig. 1): The “flash left”
and/or “flash right” decisions are managed by FLASH CTRL whose behaviour is
depicted in the statechart in Fig. 2. Its control decisions are passed to the second
component via internal interfaces Left, Right ∈ {0, 1}. The output interface control together with the timing control for flashing lights is handled by the second
234
component denoted OUTPUT CTRL (Fig. 3). The behavour of OUTPUT CTRL
is modelled by a timed Statechart with clock variable t: Setting t = 0 resets
the clock to 0. Conditions t ≥ c are interpreted as “after c milliseconds have
elapsed”.
A model state vector s(k) evaluated at execution step k can be structured
into s(k) = (t, (loc(k), v(k), tim(k))), where
–
–
–
–
t ≥ 0 is the relative time stamp of the execution,
loc(k) is the sub-vector of Statechart locations,
v(k) is the sub-vector of variable valuations and
tim(k) is the sub-vector of clock valuations.
A state formula ϕ(k) evaluated at execution step k admits atoms C.(k) where
C is a component and a location. Its interpretation is “At execution step k
component C is in location ”. Atoms on clock variables are of the form t(k) ◦ c,
where t(k) is the valuation of a clock variable t at step k, ◦ ∈ {=, ≤, ≥, <, >}
and c is a constant1 . For example, state formula
OUTPUT CTRL.FLASHING.ON(k) ∧ Left(k) ∧ Current(k) = 100 ∧ t(k) < 340
states: “At execution step k, component OUTPUT CTRL is in location FLASHING.ON, variable Left has value 1, variable Current has value 100 and clock t has
not yet reached deadline 340”.
The Transition relation is structured into timed and zero-time transitions as
described in Section 2.
4
Test Types and Their Associated Reachability
Problems
4.1
Requirements-Based Testing
Provided that our model completely reflects the requirements, every property is
represented by one or more paths through the model, together with a suitable
trace of inputs to be passed from the environment to the system under test
at certain points in time. The specific paths to be covered with the associated
specification of admissible inputs is encoded in the P -component of the BMC
c−1
Φ(i(k0 + j), s(k0 + j), s(k0 + j + 1))-conjunct ensures that
instance, while the
j=0
only legal paths through the model are taken.
If a test case is constructed from scratch, unrolling the transition relation
starts at k0 = 0. In many cases, however, test data previously constructed can
be used to drive the system into a suitable pre-state from where the requirement
can be directly checked. This corresponds to the situation where k0 > 0 and
s(k0 ) is this pre-state reached by means of the test data already available.
1
We do not need the full expressiveness of formulae involving clocks which is possible
for timed automata.
235
Fig. 3. Output control component OUTPUT CTRL managing indications lights and
associated flash cycles.
236
Example 1. For testing requirement R1 defined in Section 3 we need two types
of test cases: The first type starts in a system state where indications are off and
the current is ≤ 80%. Then it is tried to switch indications on and checked that
the controller does not give activation signals on outputs FlashLeft, FlashRight.
These test cases are characterised by the condition
c−1
P1 (loc(k0 ), . . . , loc(k0 + c)) ≡ ( k=0 Current(k0 + k) ≤ 80)∧
c−1
k=0 (FlashLeft(k0 + k) = 0 ∧ FlashRight(k0 + k) = 0 ∧
((TurnIndLeft(k0 + k) = 0 ∧ TurnIndLeft(k0 + k) = 1) ∨
(TurnIndRight(k0 + k) = 0 ∧ TurnIndRight(k0 + k + 1) = 1) ∨
(EmerFlash(k0 + k) = 0 ∧ EmerFlash(k0 + k + 1) = 1)))
The second type of test cases starts in a system state where indication lights
are on and then decreases the current to ≤ 80%, expecting the indication lights
to be switched off. These test cases are characterised by the condition
P2 (i(k0 ), loc(k0 ), . . . , i(k0 + c), loc(k0 + c)) ≡
c−1
k=0 (OUTPUT CTRL.FLASHING.ON(k0 + k) ∧
Current(k0 + k) ≤ 80 ∧ OUTPUT CTRL.IDLE(k0 + k + 1))
4.2
Equivalence Class Testing
The test case characterisations P1 , P2 from Example 1 give rise to an infinite
number of test cases, since we are dealing with a non-terminating system and we
have an infinite number of choices with respect to the points in time when inputs
could be placed. In theory, it would be possible to construct a grid automaton
from the model’s zone graph which can be covered by a finite number of test
cases which would uncover every possible deviation of the implementation from
the model [18]. The number of these test cases, however, would be far too large
to be executed in acceptable time. As a consequence it is desirable to apply a
reasonably justified heuristics allowing to reduce the number of test cases without
omitting the examination of important situations. One of the most important
heuristics is called equivalence testing. This approach tries to structure the input
domain into a finite number of so-called equivalence classes, so that each two
members of a class are equally suitable to check a certain aspect of a given
requirement (requirements-based equivalence class construction). Equivalence
classes can be further justified if it can be shown that these members stimulate
the same execution paths in the system under test (structure-based equivalence
class construction). After construction of the classes, only one or a few members
of each class have to be tested.
The construction of equivalence classes for testing reactive real-time systems
is more complicated than equivalence class construction for the test of sequential
terminating modules because the elements of each class are timed sequences of
input vectors. The availability of a model, however, facilitates the justification
237
and construction of requirements-based equivalence classes: Every cycle-free path
through the zone graph satisfying P gives rise to an equivalence class, and two
timed input traces are equivalent if they stimulate the same cycle-free path. This
notion of equivalence is justified as follows:
– We only have to consider cycle-free paths since returning to a previous node
on the path would mean that a previous model state has been reached again.
– Two timed input traces stimulating the same path through the zone graph
are regarded as equivalent, because they result in the same sequence of operations being applied to state variables, clocks and output variables and
to the application of the same deadlines. Of course, these equivalent traces
may use different input values and result in different outputs occurring at
different times, but since they are processed by the same operations and
are subject to the same deadlines, it can be reasonably expected that a few
equivalent traces suffice to uncover every error in this operational sequence.
– Conversely, two different paths through the zone graph induce different
equivalence classes since they lead to different operations exercised on the
state, clock and output variables and to the application of different deadlines.
Example 2. For the test cases characterised in Example 1 by predicate P2 different paths through the zone graph lead to the required pre-state
OUTPUT CTRL.FLASHING.ON(k0 + k) ∧ Current(k0 + k) ≤ 80
Condition Q1 below specifies a path for some suitable 0 ∈ {1, . . . , k0 −1} leading
through locations
(FLASH CTRL.EMER OFF, OUTPUT CTRL.IDLE) →
(FLASH CTRL.EMER OFF, OUTPUT CTRL.ON) →
(FLASH CTRL.EMER OFF, OUTPUT CTRL.IDLE)
where indication lights are switched on via the turn indication lever, while the
emergency flash switch remains in state off (otherwise the model would not
remain in state FLASH CTRL.EMER OFF). The electric current drops below the
acceptable threshold at execution step k0 , and this already happens during the
very first ON-phase of the flash cycle:
Q1 (i(0), loc(0), . . . , i(k0 − 1), loc(k0 − 1)) ≡
k0 −1
( =0
Current() > 80 ∧ FLASH CTRL.EMER OFF()) ∧
0 −1
OUTPUT CTRL.IDLE()) ∧
( =0
k0 −1
( =
OUTPUT CTRL.FLASHING.ON())
0
Now any solution to the BMC instance
b1 =
k0 +c−1
j=0
Φ(i(j), s(j), s(j + 1)) ∧
Q1 (i(0), loc(0), . . . , i(k0 − 1), loc(k0 − 1)) ∧
P2 (i(k0 ), loc(k0 ), . . . , i(k0 + c), loc(k0 + c))
238
with P2 as defined in Example 1 results in a representative for this equivalence
class. For example, b1 can be solved with the timed input trace2
w1 (0, (TurnIndLeft = 0, TurnIndRight = 0, Current = 100),
(100, (1, 0, 90)), (435, (1, 0, 70))
as well as with the equivalent trace
w2 (0, (TurnIndLeft = 0, TurnIndRight = 0, Current = 100),
(200, (1, 0, 98)), (210, (1, 0, 10))
Another path through the zone graph passes through locations
(FLASH CTRL.EMER OFF, OUTPUT CTRL.IDLE) →
(FLASH CTRL.EMER OFF, OUTPUT CTRL.ON) →
(FLASH CTRL.EMER OFF, OUTPUT CTRL.OFF) →
(FLASH CTRL.EMER OFF, OUTPUT CTRL.IDLE)
and can be characterised by
Q2 (i(0), loc(0), . . . , i(k0 − 1), loc(k0 − 1)) ≡
k0 −1
( =0
Current() > 80 ∧ FLASH CTRL.EMER OFF()) ∧
0 −1
( =0
OUTPUT CTRL.IDLE()) ∧
1 −1
( =
OUTPUT CTRL.FLASHING.ON()) ∧
0
k0 −1
( =1 OUTPUT CTRL.FLASHING.OFF())
with suitable 0 < 0 < 1 < k0 . This equivalence class is determined by BMC
instance
k0 +c−1
Φ(i(j), s(j), s(j + 1)) ∧
b2 =
j=0
Q2 (i(0), loc(0), . . . , i(k0 − 1), loc(k0 − 1)) ∧
P2 (i(k0 ), loc(k0 ), . . . , i(k0 + c), loc(k0 + c))
and b2 could be solved, for example, with the input trace
w3 = (0, (TurnIndLeft = 0, TurnIndRight = 0, Current = 100),
(200, (1, 0, 98)), (550, (1, 0, 10))
Since the size of region graphs is known to grow exponentially with the number of clocks (see [1, p. 70]), the zone graphs – though representing further
abstractions of the region graphs – may still become quite large for real-world
problems, so that the number of cycle-free paths may induce too many equivalence classes to be exercised on the system under test during a test campaign
of acceptable duration. In such a situation a uniform distribution of the paths
should be enforced. Techniques for statistic path selection have been suggested
in [4].
2
Trace elements are pairs (τ, x) with time stamp τ and input vector x.
239
4.3
Boundary Value Testing
It is a well-known fact that many critical implementation errors are revealed by
choosing inputs at the boundary between different types of behaviours, such as
– the boundary between legal and illegal data ranges,
– the time boundary when a clock reaches a given threshold, or
– the boundary between equivalence classes.
Normal behaviour boundary value testing selects inputs which are always on the
admissible side of the specified bounds, robustness boundary value testing selects
the inputs from the illegal side of the bounds.
In analogy to equivalence testing for timed reactive systems we first have to
clarify when a timed input trace should be regarded as a boundary input to the
system under test. To this end, to input traces are defined to be close to each
other if
1. they have the same length,
2. each pair of corresponding trace elements (ti , xi ), (ui , xi ) is close in the
canonic metric induced by the data types.
For integral numbers and enumeration types “close” means a numeric distance
of 0 or 1. For time stamps ti , ui closeness depends on the latency of the system
under test, that is, if |ti − ui | is less or equal to the smallest time interval
where consecutive inputs can still be captured and lead to the system responses
required. For floating point data types closeness depends on the resolution of
the data type: If the floating point number precision available is much finer
than the resolution ε > 0 required for the control task under consideration,
x1 , x2 ∈ floatingpoint type are considered close if |x1 −x2 | < ε. Alternatively,
if the control task really exhausts the resolution offered by the data type, close
values should only differ by a number of ulp (units in the last place), that is, the
distance measured in the number of least significant bits in the floating point
format considered [7, p. 289]. For strings, the closeness notion may be based on
the Hamming distance [6].
With this concept of closeness at hand, we can define an input trace w1 to
represent a boundary value test case if there exists an input trace w2 which is
close to a prefix of w1 and is contained in another equivalence class (or stimulates
another path through the zone graph) than w1 .
The automated construction of boundary test cases is best performed in
combination with the elaboration of equivalence classes: Recall from the sections
above that a representative of an equivalence class is constructed by finding a
solution
z = (t(k), (i(k), loc(k), v(k), tim(k), o(k))) | k ∈ {0, . . . , c}
of a suitable BMC instance b as given in Equation 1. The input trace w is
extracted from z by restriction to the points in time where the input vector
changes, say
w = (t(i ), i(i )) | ∈ {0, . . . , m}, 0 ≤ i ≤ c, 0 ≤ t(i0 ) < . . . < t(i )
240
for some suitable m ≤ c (recall that some time has to pass between changing
input vectors). Now we can perform variations on the input trace w by applying
zero, one or more changes of the following kind to each element (t(i ), i(i )) in
w:
1. For 0 < , time point t(i ) is varied in the interval [t(i−1 ), t(i+1 )]; the
variation value is denoted by ti ∈ [t(i−1 ), t(i+1 )].
2. In analogy, time point t(i0 ) of w is varied in the interval [0, t(i1 )], and time
point t(im ) is varied as [t(m), ∞).
3. Each input component of xn (i ) of i(i ) = (x1 (i ), . . . , xm (i )) can be varied
within its admissible bounds (since we are dealing with normal behaviour
boundary tests), leading to variation value xn (i ) and associated vector i (i ).
For a given input variation
w = (t (i ), i (i )) | ∈ {0, . . . , m}, 0 ≤ i ≤ c, 0 ≤ t(i0 ) < . . . < t(i )
it is checked whether it still solves the BMC instance without changing the locations, that is, we can find suitable v (k), tim (k), o(k) such that
z = (t (k), (i (k), loc(k), v (k), tim (k), o (k))) | k ∈ {0, . . . , c}
is still a solution of b. The variation process is performed until a pair of traces
z , z with the given locations loc(k) is found such that
– z solves b,
– z does not solve b,
– A prefix of the input trace w associated with z is close to a prefix of the
input trace w associated with z .
Example 3. Consider input trace w1 defined in Example 2 and its associated
equivalence class. Assuming that the system under test can capture inputs with
a precision of 1ms, the following input trace represents a boundary value of the
same equivalence class.
w1 (0, (TurnIndLeft = 0, TurnIndRight = 0, Current = 100),
(100, (1, 0, 81)), (439, (1, 0, 80))
Obviously w1 is still a solution of P2 which traverses the same locations as w1 .
The traces
w1 (0, (TurnIndLeft = 0, TurnIndRight = 0, Current = 100), (100, (1, 0, 80))
w1 (0, (TurnIndLeft = 0, TurnIndRight = 0, Current = 100),
(100, (1, 0, 81)), (440, (1, 0, 81))
however, are close to prefixes of w1 and are associated with other equivalence
classes: Trace w1 forces the system to remain in location
(FLASH CTRL.EMER OFF, OUTPUT CTRL.IDLE)
241
and w1 induces a path through locations
(FLASH CTRL.EMER OFF, OUTPUT CTRL.IDLE) →
(FLASH CTRL.EMER OFF, OUTPUT CTRL.ON) →
(FLASH CTRL.EMER OFF, OUTPUT CTRL.OFF) →
(FLASH CTRL.EMER OFF, OUTPUT CTRL.IDLE)
The process of identifying boundary input traces is considerably sped up by
application of the following techniques:
1. For a given input trace element (t(i ), i(i )) only those symbols that are
referenced in the constraints of their associated zone graph nodes need to
undergo a variation: In Example 3 above it was not necessary to perform
variations of the time points t(0) = 0, t(1) = 100 because the initial zone
(FLASH CTRL.EMER OFF, OUTPUT CTRL.OFF, Current > 80)
did not depend on any clocks.
2. Time boundaries are always given by the clock whose deadline is closest to
the current time, but which has not yet elapsed.
3. Zones specified by linear constraints about inputs and state variables can
be solved by specialised algorithms such as the Simplex algorithm for linear
constraint systems of real values or by Fourier Motzkin elimination [17] for
linear inequations involving integral numbers.
4. In the general case involving non-linear zone constraints with transcendent
mathematical functions bi-partitioning can be applied as described in [7]:
(a) The interval vector consisting of initial intervals for each relevant input
variable from i(i ) is partitioned into a collection of sub-vectors Vj , 0 <
j ≤ m such that each function occurring in the constraint is monotonic
when restricted to a vector Vj .
(b) A sub-vector Vj is identified such that the BMC instance b can be solved
without changing the sequence of locations, when some vertex v1 of Vj is
used as input vector, and such that b can not be solved without changing
the sequence of locations, when some other vertex v2 of Vj is used as
input.
(c) By bi-partitioning the edges of Vj two vertexes v1 , v2 with the same
properties as v1 , v2 can be found such that v1 , v2 are sufficiently close to
each other.
(d) Using v1 for the relevant components of i (i ) and keeping the other
components as given in i(i ) result in a boundary input trace.
4.4
Error Seeding and Mutation Testing
After having generated test cases and test data guaranteeing full requirements
coverage, adequate consideration of equivalence classes and boundary values, the
242
resulting test suite will still not be exhaustive in general. This means that some
errors of the system under test will not be uncovered by the test suite. Therefore
error seeding and mutation testing are applied as the preferred technique to measure the strength of a test suite with respect to its error detection capabilities.
In model-based testing this technique is used as follows:
1. Error seeding: The model is changed according to a suitable random distribution by modifying expressions and state transitions. This results in a
mutation of the original model.
2. An executable mutant simulation is generated from the corrupted model.
3. Mutation testing: The mutant is tested against the original test suite which
had been generated before using the original model.
4. If the test suite uncovers an error in the mutant this bug is fixed and a
regression test is performed by re-running the complete test suite.
5. If repetitive application of bug-fixing and regression testing uncovers all errors originally implanted in the mutant the test suite is considered as sufficiently effective for error detection.
6. For each uncovered bug a new test case and test data are constructed so
that this error will be detected by the new test which is then added to the
test suite. In order to make this process as efficient as possible it is desirable
to automate this test generation process.
For the generation of test data uncovering a mutation we consider the system
consisting of original model (transition relation Φ) and mutant (transition relation Φ ), both receiving the same inputs but managing their own state variables
and clocks s and s and writing to their own output vectors o and o , respectively. The mutant is tainted with a single error, this is the one for which the
new test has to be constructed. Our objective is now to generate a timed input
trace such that original model and mutant produce different timed output traces
(back-to-back test principle). This results in a BMC instance (“MT” standing for
“Mutation Testing”)
bM T ≡
c−1
(Φ(i(j), s(j), s(j + 1)) ∧ Φ (i(j), s (j), s (j + 1)))
j=0
∧ P (o(0), o (0), . . . , o(c), o (c))
with
s(0) = s (0)
P (o(0), o (0), . . . , o(c), o (c)) ≡
c
(o(j) = o (j))
j=0
While this BMC instance captures the task of mutation test construction, it
may be too time-consuming to be solved in general; in particular, because the
dimension of the state space has been doubled. In practice it is more efficient to
construct the test by re-using information from the already existing test suite:
243
1. Since the existing test suite already provides requirements coverage, there
exists a test driving the system into the pre-state (s(k), s (k)) which has
been tainted with the error seed, that is, s(k) = s (k).
2. Using the techniques described in [18] we can construct an input trace w
which, when exercised on the pre-state (s(k), s (k)) distinguishes s(k) and
s (k) by producing different timed output traces.
5
Conclusion
In this paper a formal basis for automated requirements-based, equivalence,
boundary value and mutation testing has been presented. For each of these
methods the problem of automated test case and test data generation could
be reduced to the solution of reachability problems typically encountered in
bounded model checking.
The solution of the reachability problems induced by test automation require
extensive support of SMT solvers. Solution techniques based on interval analysis
have been described in [2]. For code-based testing, more powerful interpretation
techniques are needed than in model-based testing in order to cope with pointers,
arrays, union data types and bit vectors; solutions to these problems are given
in [13, 12, 11]. These publications also provide detailed references to the work of
other research groups in the field of test automation.
While this paper investigated the theoretical foundations of automated modelbased testing, the results are currently applied in practice in the fields of avionics, railway and automotive control systems. Aspects related to testing in these
application domains have been described in [10, 8, 14].
Most of the test generation features described in this paper have already
been implemented or are currently integrated into the test automation tool RTTester [19] developed by Verified Systems International GmbH together with the
author’s research team at the University of Bremen. This tool provides the core
components for an integrated testing and verification environment, where codebased and model-based formal verification, abstract interpretation and testing
can be performed in an integrated manner and in compliance with the requirements of applicable standards for safety-critical systems development.
Acknowledgements. I would like to express my gratitude to Bernd Krieg-Bückner
for his invaluable support of both my academic and industrial career and for his
competence and resourcefulness as a project partner, co-author and colleague.
References
1. B. Bérard et. al. Systems and Software Verification. Springer, 2001.
2. Bahareh Badban, Martin Fränzle, Jan Peleska, and Tino Teige. Test automation for hybrid systems. In Proceedings of the Third International Workshop on
SOFTWARE QUALITY ASSURANCE (SOQUA 2006), Portland Oregon, USA,
November 2006.
244
3. Tsun S. Chow. Testing software design modeled by finite-state machines. IEEE
Transactions on Software Engineering, SE-4(3):178–186, March 1978.
4. A. Denise, M.-C. Gaudel, and S.-D. Gouraud. A generic method for statistical
testing. In Proceedings of the 15th International Symposium on Software Reliability
Engineering (ISSRE’04), pages 25–34, 2004.
5. European Committee for Electrotechnical Standardization. EN 50128 – Railway
applications – Communications, signalling and processing systems – Software for
railway control and protection systems. CENELEC, Brussels, 2001.
6. Richard W. Hamming. Error detecting and error correcting codes. Bell System
Technical Journal, 26(2):147–160, 1950.
7. Luc Jaulin, Michel Kieffer, Olivier Didrit, and Éric Walter. Applied Interval Analysis. Springer-Verlag, London, 2001.
8. J. Peleska and A. Tsiolakis. Automated Integration Testing for Avionics Systems.
In Proceedings of the 3rd ICSTEST – International Conference on Software Testing, April 2002.
9. Jan Peleska. Formal Methods and the Development of Dependable Systems. Number
9612. Christian-Albrechts-Universität Kiel, Institut fr Informatik und Praktische
Mathematik, December 1996. Habilitationsschrift.
10. Jan Peleska. Formal methods for test automation - hard real-time testing of controllers for the airbus aircraft family. In Proc. of the Sixth Biennial World Conference on Integrated Design & Process Technology (IDPT2002), Pasadena, California, June 23-28, 2002. Society for Design and Process Science, June 2002. ISSN
1090-9389.
11. Jan Peleska. A unified approach to abstract interpretation, formal verification and
testing of c/c++ modules. In John S. Fitzgerald, Anne E. Haxthausen, and Husnu
Yenigun, editors, Theoretical Aspects of Computing - ICTAC 2008, 5th International Colloquium, volume 5160 of Lecture Notes in Computer Science, pages 3–22.
Springer, 2008.
12. Jan Peleska and Helge Löding. Symbolic and abstract interpretation for c/c++
programs. In Proceedings of the 3rd intl Workshop on Systems Software Verification
(SSV08), Electronic Notes in Theoretical Computer Science. Elsevier, February
2008.
13. Jan Peleska, Helge Löding, and Tatiana Kotas. Test automation meets static analysis. In Rainer Koschke, Karl-Heinz Rödiger Otthein Herzog, and Marc Ronthaler,
editors, Proceedings of the INFORMATIK 2007, Band 2, 24. - 27. September, Bremen (Germany), pages 280–286.
14. Jan Peleska, Oliver Möller, and Helge Löding. Model-based testing for modeldriven development with uml/dsl. In Proceedings of the Software & Systems Quality
Conference 2008 (SQC 2008).
15. S. Ranise and C. Tinelli.
Satisfiability modulo theories.
TRENDS and
CONTROVERSIES–IEEE Magazine on Intelligent Systems, 21(6):71–81, 2006.
16. SC-167. Software Considerations in Airborne Systems and Equipment Certification. RTCA, 1992.
17. Alexander Schrijver. Theory of Linear and Integer Programming. John Wiley &
sons, 1998.
18. J.G. Springintveld, F.W. Vaandrager, and P.R. D’Argenio. Testing timed automata. Theoretical Computer Science, 254(1-2):225–257, March 2001.
19. Verified Systems International GmbH, Bremen. RT-Tester 6.2 – User Manual,
2007.
245
Jan Peleska is professor for computer science at the
University of Bremen, where he teaches operating systems and distributed systems. His research focus is
on verification, validation and testing of safety-critical
embedded real-time Systems. He is co-founder of the
company Verified Systems International GmbH, where
the results of his research work are applied to avionic
software, space mission systems and railway and automotive control systems. His first contacts to the University of Bremen were made through Bernd KriegBrückner, and they were partners in the UNIFORM
research project on the integration of different formal methods and associated tools. Bernd considerably
helped paving the way for the foundation of Verified
Systems by introducing Jan to new industrial project
partners in the field of space transportation systems.
Regularity based on categorical algebra
(dedicated to the 60th birthday of Bernd Krieg–Brückner)
Horst Reichel
Institut für Theoretische Informatik
Technische Universität Dresden
D–01062 Dresden, Germany
[email protected]
Abstract. In [6] the author has sketched an approach to regular subsets
for arbitrary inductively defined data types. This development is motivated by the increasing interest in unranked trees as a foundation for
XML (see [5], [1]) and in semi–structured date.
The aim of this paper is a precise formalization of the ideas developed in
the previous paper. The formalization is based on initial algebras such
that the fundamental operations may have rich structured domains. The
most liberal approach with respect to algebraic operations with structured domains is the categorical notion of T –algebras in the sense of
[2]. Therefore we will use categorical universal algebra as basis of the
formalization. Regular T –algebras are introduced as generalizations of
bottom–up automata for unranked trees [8]. It is shown that the category of regular T –algebras is closed with respect to finite colimits which
provides a useful basis to compose regular T –algebras that represent
accepting automata of generalized regular data sets.
1
Basic notions of categorical algebra
The reader is assumed to have some basic knowledge in category theory as
presented in [2].
Let T : Set → Set be any endofunctor of the category of sets. A T –algebra is
a pair A = (A, α : T (A) → A) consisting of the carrier set A and the structure
map α : T (A) → A of A.
A homomorphism f : A → B, where A = (A, α : T (A) → A) and B = (B, β :
T (B) → B) are given T –algebras, is a morphism f : A → B in Set commuting
with the structure maps, i.e. f ; β = α; T (f ). TAlg denotes the category of all
T –algebras and UT : TAlg → Set denotes the forgetful functor which assigns to
each T –algebra the carrier set. A type functor T : Set → Set is called a varietor
[2] if the forgetful functor UT : TAlg → Set has a left adjoint FT : Set → TAlg .
The endofunctor FT ; UT : Set → Set is called the monad freely generated by
T : Set → Set.
In case of classical algebras the type functor is a finite sum of finite products
of the identity functor.
247
An example outside of classical algebra are list algebras. List algebras are
given by the type functor TList : Set → Set with TList (M ) = M ∗ . In case of list
algebras the structure map assigns to each finite sequence (or list) of elements
of the carrier set another element of the carrier set. It is well known that the
category of list algebras (TList )Alg has an initial algebra which represents finitely
branching unranked trees.
Let L be a given set of labels and TL list : Set → Set be the endofunctor with
TL list (M ) = L × M for each set M and TL list (f ) = IdL × f for each morphism
f : A → B in Set. This type functor is again a varietor and the initial algebra
represents labeled unranked trees where the labels on the nodes are elements of
the given set L.
If one takes the varietor Tf ineSet : Set → Set which assigns to each set M
the set Tf ineSet (M ) of all finite subsets of M then the associated initial algebra
represents unranked, repetition free, unordered trees. By the same construction
as above one could generated also the labeled version of this kind of unranked
trees.
This few examples should be enough to demonstrate the greater expressive
power of categorical algebra.
For a better understanding of the following we will recall from [2] the construction of free T –algebras. We define a chain of objects and morphisms in Set
indexed by all ordinals:
f0,1
f1,2
fω,ω+1
fn,n+1
F0 → F1 → . . . → Fω −→ Fω+1 → . . . → Fn −→ Fn+1 → . . .
where
– F0 = ∅, F1 = T (F0 ), Fn+1 = T (Fn ),
f0,1
f1,2
f0,1
– → is the embedding of the empty set, −→= T (−→) and
fn,n+1
fn−1,n
−→ = T ( −→ )
F0,1
F1,2
– Fω is the colimit of the preceding chain, i.e., the colimit of F0 → f1 →
. . . →, which is also the construction for each limit ordinal. The morphism
fω,ω+1
Fω −→ Fω+1 = T (Fω ) is induced by the universal property of the colimit
construction of Fω .
The varietor T : Set → Set is called constructive if the transfinite construction reaches a fixed point and it is called finitary if Fω is a fixed point, i.e.
T (Fω ) = Fω .
This is the construction of the initial T –algebra, i.e. of the T –algebra freely
generated by the empty set. The construction of a T –algebra freely generated
by the set X can be reduced to the construction of the initial (T + X)–algebra.
If T : Set → Set is a finitary varietor then is the monad freely generated
by T : Set → Set again a finitary varietor. This property of finitary varietors is
basic for the following generalization of regularity from strings and ranked trees
to arbitrary inductively defined data types.
248
2
Generalization of Regularity
In the following we will introduce the concept of regular subsets of initial T –
algebras in an inductive way.
The basic case is given by any finitary varietor Tf ineSet : Set → Set which
preserves finite sets. Any functor which is a finite sum of finite products of the
identity functor is a finitary varietor and preserves finite sets. This implies that
the basic case comprises classical algebras.
If we consider finite T –algebras with a distinguished subset of accepting
states as button up automata then in the basic case a subset of an initial T –
algebra is regular if it is the inverse homomorphic image of the set of accepting
states of a finite T –algebra A with respect to the unique homomorphism from
the initial T –algebra to the finite T –algebra.
Definition 2.1.(base case) Let T : Set → Set be a varietor preserving finite
sets. A subset X ⊆ F of the initial T –algebra F is called regular, if there is a finite
−1
(A0 ),
T –algebra A = (A, α) and a distinguished subset A0 ⊆ A with X = fA
where fA : F → A is the unique morphism from the initial T –algebra F to the
finite algebra A.
Let Σ = {a1 , . . . , an } be a finite alphabet. The set Σ ∗ of all finite words
over the alphabet Σ can be seen as the carrier set of an initial Tstring –algebra,
where the varietor Tstring : Set → Set corresponds to a signature consisting of
a constant and a unary operation for each element of Σ. Finite Tstring –algebras
with a distinguished subset correspond to finite deterministic automata and the
previous definition redefines the concept of regular sets of strings over a finite
alphabet directly.
We see that the notion introduced by the basic case extends the well known
notion of regularity for strings and ranked trees following the line of Thatcher
and Wright from 1968 [7].
But, the previous definition yields also the concept of regular sets for finitely
branching, repetition–free , unordered trees, since the varietor Tf inset : Set →
Set, which assigns to each set the set of all finite subsets, this functor preserves
finite sets.
The more interesting generalization is given by the following induction step,
but this needs some preparations.
Let T : Set → Set be a finitary varietor, UT : TAlg → Set the forgetful
functor and LT : Set → TAlg the unique left–adjoint functor to the forgetful
functor.
The composed functor LT ; UT : Set → Set is called the monad freely generated by T : Set → Set [3].
Definition 2.2. (inductive definition of regularity)
Each finitary varietor T : Set → Set which preserves finite sets is called a type
functor with a regularity concept.
Each monad freely generated by a type functor with a regularity concept is again
a type functor with a regularity concept.
Let LT ; UT : Set → Set be a type functor freely generated by a type functor
T : Set → Set with a regularity concept.
249
Let Y be a finite set. A subset X ⊆ UT (LT (Y )) is called regular if there is a
finite LT ; UT –algebra A = (A, α) with a distinguished subset A0 ⊆ A and a
mapping c : Y → A such that
– for each a ∈ A the set α−1 (X) ⊆ UT (LT (A)) is a regular.
−1
– and X = fA,c
(A0 ), where fA,c denotes the unique homomorphism induced
by the the mapping c : Y → A and the LT ; UT –algebra A.
For a type functor T : Set → Set a finite T –algebra A will be called regular
if for each element a ∈ A the inverse image α−1 ({a}) ⊆ T (A) is a regular subset.
Evidently, if T : Set → Set preserves finite sets, then every finite T –algebra is
regular.
The type functor TList : Set → Set is a type functor with a regularity concept
according to the inductive definition above. The resulting notion of regular sets
for the data type of unranked ordered trees coincides with the recent definitions
in literature.
2.1
Example
Let us consider the type functor T1 : Set → Set with T1 (X) = {a, b, A, B}∗ ×
X ∗ . A T1 –algebra E is then given by a carrier set E and an operation α :
{a, b, A, B}∗ × E ∗ → E. The initial T1 –algebra represents finite unranked unordered trees where each node is decorated with an attribute being a string over
the alphabet {a, b, A, B}.
– Let be E = {s0 , s1 , s2 , s3 , sd }
– and for (w, l) ∈ {a, b, A, B}∗ × E ∗ let be
–
⎧
s0 if w ∈ L(a∗ )
l ∈ {nil}
⎪
⎪
⎪
⎪
l ∈ {s0 s0 , s0 s0 s0 }
⎨ s1 if w ∈ L(b∗ )
α(w, l) = s2 if w ∈ L((A + B)a∗ ) l ∈ {s1 s1 , s0 s1 }
⎪
⎪
s3 if w ∈ L((A + B)b∗ ) l ∈ {s2 s2 s2 , s1 s2 s3 }
⎪
⎪
⎩
else
sd else
– and let {s3 } ⊆ E be a distinguished subset (of accepting states).
Then we have four interesting classes of attributed trees, defined by the
inverse homomorphic images of s0 , s1 , s2 , s3 respectively, and the complement of
the union of this classes, represented by sd .
– s0 represents the class of all trees with exactly one node labeled with a string
w ∈ L(a∗ ).
– s1 represents the class of trees with two or three sons, classified by s0 and
the root is labeled with a string w ∈ L(b∗ ).
– s2 represents the class of trees with two sons such that the first is classified
by s1 or s0 and the second son is classified by s1 . Finally the root is labeled
with a string L((A + B)a∗ ).
250
– Finally, the accepting state s3 represents the class of trees with exactly three
sons, where either all of them are classified by s2 or the first son is classified
by s1 , the second by s2 and the third again by s3 . The root is labeled by a
string w ∈ L((A + B)b∗ ).
The example is basically a partial T1 –algebra with a one–point completion
given by the state sd . In terms of automata this completion point is a trap. If
the computation once reaches the trap it can never leave it.
Motivated by this example we will use in the following finite partial T –
algebras as a model for bottom–up automata wich recognize generalized regular
sets.
Definition 2.3. (one–point completion)
Let A = (A, α : T (A) →?A) be a partial T –algebra. The one–point completion A∗ = (A + 1, α∗ : T (1 + A) → 1 + A) is a total T –algebra such that
α∗ ((T (iA ))(x)) = iA (x) if x ∈ dom(α) and α+ (y) = i1 (∅) else, where 1 = {∅}
and i1 : 1 → 1 + A, iA : A → 1 + A are the canonical injections.
The set accepted by a regular partial T –algebra A is then the set accepted
by the one–point completion A∗ as defined above.
3
Regular Expressions based on Limits and Colimits
In the following we will reduce operations on regular subsets in the initial T –
algebra, where T : Set → Set is a type functor with a regularity concept, to
finite limits and colimits in the category of partial regular T –algebras with week
homomorphisms.
Let T : Set → Set be a varietor with a regularity concept. A partial regular
T –algebra A = (A, α : T (A) →?A) is given by a finite set A together with a
partial mapping α : T (A) →?A such that for each a ∈ A the set α−1 ({a}) ⊆
T (A) is regular.
For partial regular T –algebras A = (A, α), B = (B, β) a total mapping
f : A → B is a weak homomorphism if x ∈ dom(α) implies f (x) ∈ dom(β)
and f (α(x)) = β((T (f ))(x)). TprAlg denotes the category of partial regular T –
algebras with week homomorphism.
First we show that the category TprAlg is closed with respect to finite colimits.
Theorem 3.1. TprAlg is closed with respect to finite colimits.
Proof: The empty T –algebra with empty domain of the structure mapping
is the colimit of the empty diagram.
For finite coproducts it is sufficient to show the existence of the coproduct
A + B for any A = (A, α : T (A) →?A), B = (B, β : T (B) →?B) from TprAlg .
On the set A + B with the injections iA : A → A + B, iB : B → A + B we define
the relation ρ ⊆ (A + B) × (A + B) as the smallest equivalence relation which
satisfies the following two conditions:
(1) ρ ⊇ {(iA (α(jA (x))), iB (β(jB (x)))) | x ∈ T (∅) ∧ jA (x) ∈ dom(α) ∧ jB (x) ∈
dom(β)}
(2) If (wA ∈ dom(α) ∧ wB ∈ dom(β) ∧ (T (iA ; natρ )(wA ) = (T (iB ; natρ )(wB ))
then (((iA (α(wA )), iB (β(wB ))) ∈ ρ.
251
This relations allows to define on the quotient set (A + B)/ρ a T –algebra
Q = ((A + B)/ρ, δ : T ((A + B)/ρ) →?(A + B)/ρ)
with dom(δ) = [T (iA ; natρ )](dom(α)) ∪ [T (iB ; natρ )](dom(β)).
If [x]ρ ∈ [T (iA ; natρ )](dom(α)) then δ([x]ρ ) = natρ (iA (α(x))) and δ([x]ρ ) =
natρ (iB (β(x))) if [x]ρ ∈ [T (iB ; natρ )](dom(β)).
If [x]ρ ∈ [T (iA ; natρ )](dom(α))∩[T (iB ; natρ )](dom(β)) then due to condition (2)
it does not matter if we set δ([x]ρ ) = natρ (iB (β(x))) or δ([x]ρ ) = natρ (iA (α(x)))
the results are equal.
By this definition one gets homomorphisms iA ; natρ : A → Q, iB ; natρ : B →
Q and one can show that with these homomorphisms Q is a coproduct of A and
B in the category of partial T –algebras with weak homomorphisms. Evidently
Q is finite and later we will prove that Q is regular.
Let f : A → C, g : B → C be given weak homomorphisms, let [f, g] : A+B →
C be the canonical mapping induced by f and g and let natker[f,g] ; q = f be the
corresponding factorization. Since homomorphisms preserve existing constants
ker[f, g] ⊇ {(iA (α(jA (x))), iB (β(jB (x)))) | x ∈ T (∅) ∧ jA (x) ∈ dom(α) ∧ jB (x) ∈
dom(β)}
To show that ker[f, g] satisfies also condition (2) we assume wA ∈ dom(α),
wB ∈ dom(β) and T (iA ; natker[f,g] )(wA ) = T (iB ; natker[f,g] )(wB ) This implies
T (q)(T (iA ; natker[f,g] )(wA )) = T (iA ; natker[f,g] ; q)(wA ) = T (f )(wA ) = . . . =
T (g)(wB ) so that γ(T (f )(wA )) = γ(T (g)(wB ). Since f and g are homomorphisms and wA ∈ dom(α), wB ∈ dom(β) we obtain f (α(wA )) = g(β(wB )) or in
other words [f, g](ia (wA )) = [f, g](iB (wB )) what we have to prove.
The next step is the existence of coequalizers. The construction is based on
the notion of a T –congruence. Let A = (A, α : T (A) →?A) be a partial T –
algebra. An equivalence relation ρ ⊆ A × A is called a T –congruence, if x ∈
dom(α), y ∈ dom(α) and T (natρ )(x) = T (natρ )(y) implies (α(x), α(y)) ∈ ρ.
The quotient algebra A/ρ = (A/ρ, δ : T (A/ρ) →?A/ρ) is defined by dom(δ) =
T (natρ )(dom(α)) and δ(T (natρ )(x)) = natρ (α(x)).
Let f : A → B, g : A → B be homomorphisms and ρ(f, g) the smallest
T –congruence in B which contains {(f (x), g(x)) | x ∈ A}. The natural mapping
natρ(f,g) : B → B/ρ(f, g) defines a homomorphism natρ(f,g) : B → B/ρ(f, g)
which is a coequalizer of the given homomorphisms.
Finally we prove inductively that the coproduct of two objects in the category
TprAlg is again in TprAlg and that the coequalizer of f : A → B, g : A → B is in
TprAlg if B is in TprAlg . This will complete the proof.
In the base case a type functor T 0 : Set → Set preserves finite sets, so that
any finite partial T0 –algebra is regular. This means, in the the case the theorem
holds.
0
0
Let UT0 : TAlg
→ Set be the forgetful functor and L0T : Set → TAlg
a left
0
0
adjoint. The LT ; UT : Set → Set is a type functor with regularity concept on
the next inductive level.
We show, that the union of finitely many regular sets in L0T (Y ) is regular
again. It is sufficient, to show this for two regular subsets X1 , X2 ⊆ L0T (Y ).
252
By definition there are regular T 0 –algebras A1 , A2 , distinguished subsets Z1 ⊆
A1 , Z2 ⊆ A2 and mappings c1 : Y → A1 , c2 : Y → A2 such that X1 =
−1
−1
(Z1 ), X2 = fA
(Z2 ).
fA
1 ,c1
2 ,c2
Then the pushout for fA1 ,c1 : L0T (Y ) → A1 , fA2 ,c2 : L0T (Y ) → A2 with the
canonic injection h1 : A1 → Q, h2 : A2 → Q exists. Together with A1 , A2 the
T 0 –algebra Q is regular and it holds
X1 ∪ X2 = (fA1 ,c1 ; h1 )−1 (h1 (Z1 ) ∪ h2 (Z2 )),
i.e., X1 ∪ X2 is regular. Since for both the coproduct and the coequalizer, which
are quotients of finite sets, for each element x the inverse image δ −1 (x) is the
union of finitely many regular set and therefor regular as just proved, the coproduct and the coequalizer are also regular. This means that the category of
regular L0T ; UT0 –algebras is closed with respect to finite colimits. The preceding
construction describes also the induction step.
As next we will show that homomorphisms reflect regular sets.
Theorem 3.2. Let h : A → B be a T –homomorphism, B a partial regular
T –algebra and X ⊆ T (B) a regular set. Then (T (h))−1 (X) ⊆ T (A) is again
regular.
Proof: By definition there is a partial regular T –algebra C, a mapping c : B →
C and a distinguished subset C0 ⊆ C such that (fC,c )−1 (C= ) = X. This implies
the existence of a mapping h; c : A → C so that (fC,h;c )−1 (C0 ) = (T (h))−1 (X)
is regular.
This result helps us to see that TprAlg is also closed with respect to finite
limits.
Theorem 3.3. TprAlg is closed with respect to finite limits.
Proof: Let f, g : A → B be homomorphisms in TprAlg . Then we define the
finite partial T –algebra E = (E, δ : T (E) →?E) with E = {x ∈ A | f (x) =
g(x), x ∈ A}, δ(y) = α(T (iE )(y)) for y ∈ dom(δ) = (T (iE ))−1 (dom(α)), where
iE : E → A is the embedding of the subset E. Since δ −1 (e) = (T (iE ))−1 (α−1 (e))
this subset is regular due to Theorem 3.2. This means, E is regular and evidently
the equalizer of the two homomorphisms.
For any two objects A, B of TprAlg we construct the finite partial T –algebra
A × B = (A × B, (T (pA ); α, T (pB ); β) : T (A × B) →?A × B). Then pA : A × B →
A, pB : A × B → B are homomorphisms and dom((T (pA ); α, T (pB ); β)) =
T (pA )−1 (dom(α)) ∩ T (pB )−1 (dom(β)) is the intersection of two regular subsets.
In the same way as we proved above that a finite union of regular subsets is
is regular we can show that a finite intersection of regular subsets ist regular.
This means TprAlg is closed with respect to finite products. Again by induction
on type functors with a regularity concept one can see that the one–element
T –algebra 1tT with total domain is a final object in TprAlg .
As a direct consequence of the preceding theorems we obtain
Theorem 3.4. For any type functor T with a regularity concept and any
finite set X the regular subsets in T (X) are closed with respect to boolean
operations.
253
In [6] the composition of regular algebras corresponding to sequential composition and the ∗–operation in regular expressions is discussed in a very special
case. This compositions are based on the coproduct and on fusing two elements
of a regular algebra. The fusing of elements is a special case of a pushout construction.
may
Let T : Set → Set be any type functor with regularity concept and 1par
T
denote the one–element regular T –algebra with empty domain of the structure
→ A, b : 1par
→B
mapping. For T –algebras A, B the homomorphisms a : 1par
T
T
select elements and the pushout of these selecting homomorphisms fuses the
selected elements.
For the minimization of automate accepting regular sets of unranked trees
in [8] two different kinds of concatenations for unranked trees are used. This
illustrates, that the syntax of regular expressions can not be simply extended
from strings to more general data types. In [1] the section 10 Hedge and tree
regular expression is empty since 2001. This observation suggests the assumption
that the generalizations of the sequential composition and the ∗–operation on
generalized regular sets depends on the used data type an can not be done in a
uniform way.
We hope that finite colimits of regular T –algebras, understood as accepting
bottom–up automata for regular sets of complex structured data, will be a useful
basis for a future study of the different kinds of regular expression.
In [4] another generalization of regular and context free sets of strings to finite
objects like trees or graphs is presented. This approach is based on universal
algebra of many–sorted algebras and defines the sets in terms of systems of
equations. In the introduction the author claims that automata cannot be used
ar general level since automata have no meaning for abstract algebraic objects.
We hope that we could show that at least to some extend bottom–up automate
can usefully be applied to to abstract algebraic objects.
References
1. M. Murata A. Brüggemann-Klein and D. Wood. Regular tree and regular hedge
languages over unranked alphabets: Version 1. Technical Report HKUST-TCSC2001-0, The Hongkong University of Science and Technology, April 2001.
2. J. Adamek and V. Trnkova. Automata and Algebras in Categories. Kluwer Academic
Publisheres, 1990.
3. M. Barr. Coequalizers and freee triples. Math. Z., (166):307–322, 1970.
4. B. Courcelle. Basic notions of universal algebra for language theory and graph
grammars. Theoretical Computer Science, 163:1–54, 1996.
5. F. Neven. Automata, logic, and XML. In Proceedings CSL 2002, volume 2471 of
LNCS, pages 2–26. Springer, 2002.
6. Horst Reichel. An algebraic approach to regular sets. In Algebra, Meaning, and
Computation, volume 4060 of LNCS, pages 449–458. Springer, 2006.
7. J.W. Thatcher and J.B. Wright. Generlized finite automata theory with an application to a decision problem of second–order logic. Mathematical Systems Theory,
2(1):57–81, 1968.
254
8. Wolfgang Thomas, Julien Christau, and Christof Löding. Deterministic automata
on unranked trees. In Proceedings of the 15th Symposium on Fundamentals of Computation Theory, FCT 2005, number 3623 in Lecture Notes in Computer Science.
Springer, 2005.
Horst Reichel is a professor emeritus at the Technische Universität Dresden. He is obliged to deepest
thanks since in the 80th he has been invited by Bernd
to become a member of the COMPASS Working Group
and to participate in the WADT workshop series. This
was the perfect community to discuss the ideas on algebraic specifications, a topic of low interest and support
in east Germany. Later the author got the possibility
to cooperate with the research group of Bernd in Bremen. One result of this cooperation is the specification
language COCASL. The author is deeply impressed of
the broad field of scientific activities and of Bernd’s
abilities to bring together people.
15 Years of Rolland
Thomas Röfer1 , Christian Mandel2 , Axel Lankenau3 ,
Bernd Gersdorf1 , and Udo Frese2
3
1
Deutsches Forschungszentrum für Künstliche Intelligenz GmbH,
Sichere Kognitive Systeme, Enrique-Schmidt-Str. 5, 28359 Bremen, Germany
E-Mail: {Thomas.Roefer,Bernd.Gersdorf}@dfki.de
2
Fachbereich 3 – Mathematik und Informatik, Universität Bremen,
Postfach 330 440, 28334 Bremen, Germany
E-Mail: {cman,ufrese}@informatik.uni-bremen.de
Daimler AG, Research and Advanced Engineering, 71059 Sindelfingen, Germany
E-Mail: [email protected]
Abstract. The paper presents 15 years of research conducted on the
Bremen Autonomous Wheelchair “Rolland”. Rolland is a mobility assistant for the handicapped and the elderly. The paper presents the different
versions of Rolland that were built over the years, as well as the special
input devices that are used with Rolland. Contributions were made to
the state of the art in the areas of self-localization, mapping, navigation, safety, and shared control. A number of assistants are the practical
results of this research conducted.
1
Introduction
The name “Rolland” was coined by Bernd Krieg-Brückner. It is a combination
of the German word “Rollstuhl” for “wheelchair” and the name of the symbol
of the independence and liberty of the Freie Hansestadt Bremen, the “Roland”
monument. The term is now well-established in the area of research on intelligent
mobility assistants.
This paper is structured as follows: first, the different versions of the Rolland
wheelchairs will be presented, followed by the special input devices that were
used in combination with Rolland. Afterwards the research topics that have
driven the development of Rolland are discussed. This work has resulted in
a number of applications that are presented before the paper closes with the
conclusions.
2
The Different Rolland Models
Over the years, different versions of Rolland were set-up on different commercial wheelchairs. The different versions also reflect the development of sensor
technology and computing technology. [1] describes the state of the art in smart
wheelchairs in the year 2000, while [2] presents the state of the art seven years
later. A clear trend is that the design of Rolland deviates less and less from the
256
(a) Rolland I
(b) Rolland II
(c) Rolland III
(d) Rolland IV
Fig. 1. The different Rolland models constructed from 1993 until 2008.
wheelchair used as base, i. e., the additional equipment has become smaller, and
it has been hidden better. Thereby, the additional hardware does impede the
normal usage of a power wheelchair lesser and lesser.
In addition, our 3-D simulator SimRobot [3] has been employed during the
development, i. e., each version of Rolland also has a simulated counterpart in
SimRobot. Thereby, the software development can be significantly simplified,
because algorithms can be tested in simulation before they are applied to the
real system.
2.1
Rolland I
“Rolland I” is based on an old wheelchair of the company Meyra. It was originally
acquired as a mobile platform that was powerful enough to carry a set of analog
artificial neurons that, in the mid-nineties, were quite big and heavy. The front
axle drives the wheelchair while the back axle is used for steering. Therefore, the
wheelchair moves like a car driving backwards. By the student project SAUS
[4] it was equipped with a Pentium 100 computer, twelve bumpers, six infrared
sensors, 16 ultrasonic sensors, and a camera. The infrared sensors can only detect
whether there is an obstacle within a radius of approximately 15 cm; however,
they cannot measure the distance. Two different kinds of ultrasonic sensors are
fitted to the wheelchair: half of the sensors have an opening-angle of 80◦ while
the other half only measure in a range of 7◦ . In addition, the wheelchair can
measure the rotations of its front wheels.
2.2
Rolland II – Meyra Genius
“Rolland II” is based on the commercial power wheelchair Genius 1.522 manufactured by the German company Meyra. As Rolland I, the wheelchair is a
non-holonomic vehicle driven by its front axle and steered by its rear axle. The
human operator controls the system with a joystick. The wheelchair has been
extended by a standard PC (Pentium III 600MHz, 128 MB RAM) for control
257
and user-wheelchair interaction tasks, 27 sonar sensors, and a laser range sensor behind the seat. The sonars are arranged around the wheelchair such that
they cover the whole surroundings. The electronics is able to simultaneously fire
two sensors, one on the left side and one on the right side of the wheelchair.
An intelligent adaptive firing strategy has been introduced by Röfer and Lankenau [5–7]. The sonars are mainly used for obstacle detection. The laser range
finder has an opening angle of 180◦ towards the backside of the wheelchair and
is able to provide 361 distance measurements every 30 ms. It is used for mapping and localization purposes. The original Meyra wheelchair already provides
two serial ports to set target values for the speed and the steering angle as well
as determining their actual values. Data acquired via this interface is used for
dead reckoning. The odometry system based on these measurements is not very
precise, i. e., it performs relatively well in reckoning distances but it is weak in
tracking angular changes. The overall software architecture is described in [8].
2.3
Rolland III – Meyra Champ
Based on the battery-powered wheelchair Meyra Champ 1.594, the third version
of Rolland for the first time featured two laser range finder mounted at ground
level, which allowed for scanning beneath the feet of a human operator. As an
additional sensor device, the system provides two incremental encoders which
measure the rotational velocity of the two independently actuated rear-wheels.
The differential drive kinematics comes along with two passive front castorwheels, giving it the freedom to turn on the spot. The castor-wheels require a
significant amount of force to let them turn, which requires a complex controlling
behavior when changing the driving direction. Three wheelchairs of type Rolland
III have been constructed:
1. The first instance is used as a development platform within the Transregional Collaborative Research Center SFB/TR 8 “Spatial Cognition”. Since
it has no dedicated control PC, a laptop has to be connected (via USB), and
sensors, wheelchair and software must be started in the correct order and
time.
2. The second instance was built up for a demonstration of the Safety Assistant
(cf. section 5.1) in the Heinz Nixdorf Museumsforum as a diploma thesis (see
[9]). The software runs on a PC104 embedded computer installed inside the
battery box. The system can be started using a single switch in about 80
seconds. The wheelchair was used by thousands of visitors of the exhibition
Computer.Medizin in a small demonstration area.
3. The last instance of Rolland III was built up as mobility platform for the
SHARE-it project (http://www.ist-shareit.eu) that is engaged in the field
of Ambient Assisted Living. It uses a more powerful embedded PC of type
EPIA-NX12000EG with a variety of software assistance. An additional microcontroller board observes PC and wheelchair and stops the wheelchair if
the communication is disturbed.
258
(a) Joystick
(b) Head-Joystick
(c) Speech Control
(d) Brain-Computer
Interface
Fig. 2. Different input devices developed for a variety of scenarios and target groups.
2.4
Rolland IV1 – Otto Bock Xeno
In September 2008, the Rolland software has been adapted to a prototype
wheelchair of type Xeno from Otto Bock Healthcare. Its sensor equipment is
very similar to Rolland III (cf. section 2.3), but instead of passive castor-wheels,
it uses a Single Servo Steering (S 3 ) for active steering of the two wheels. In
contrast to a classical Ackermann steering, this steering method allows turning
on the spot (as with the Champ), but without additional force from the differential drive, which results in a very smooth and precise path following. Another
advantage is the extended wheelbase that allows the front laser scanner to scan
between front and rear axles on the left and right side of the wheelchair. The
control PC is a netbook installed under the seat and connected to the CAN bus
of the wheelchair. An emergency stop can be initiated by the CAN based power
control unit if communication problems with the control PC appear.
3
Special Input Devices
Usually electrical wheelchairs are operated by joystick (cf. Fig. 2(a)). From that
perspective they handle the handicapped’s inability to walk by a control loop
that embeds the operator’s remaining sensory-motor capabilities into the vehicle’s actuating body. In the 15 years of Rolland’s development, several interface
techniques have been investigated that not only serve the walking impaired but
also the quadriplegic.
3.1
Head-Joystick
First proposed in [10] and later evaluated in [2, 11, 12], the basic idea of the headjoystick (cf. Fig. 2(b)) is to let the user of an automated wheelchair control the
translational and rotational velocity by continuous pitch and roll movements of
1
As the system is developed in cooperation with a new partner, it will have a different
name. However, since the name has not been decided yet, it is still referred to as
“Rolland IV”.
259
his/her head. Still able to observe the environment by turning the head around
the free yaw-axis without causing any control commands, the user’s head movements around the remaining two axes must exceed a so-called dead-zone in order
to evoke a desired movement. The necessary hardware configuration consists out
of a small-size 3-DOF orientation tracker (IMU) that is mounted at the back
of the operator’s head by means of an easy to wear headband. Here, the IMU
continually monitors the user’s head posture and acceleration w.r.t. the pitch-,
yaw-, and roll-axis.
3.2
Speech Control
With the interpretation of natural language route descriptions [13–15], speech
control has been tackled as a verbal interface that allows for the execution of utterances such as “Go down the corridor and take the second door to the left” (cf.
Fig. 2(c)). One of the basic concepts involved is the decomposition of instructions
given by humans into sequences of imprecise route segment descriptions. By applying fuzzy rules for the involved spatial relations and actions, a search tree is
constructed that can be searched in a depth-first branch-and-bound manner for
the most probable goal configuration w.r.t. the global workspace knowledge of
the wheelchair.
3.3
Brain-Computer Interface
Brain-Computer interfaces, or BCIs for short, are tools that facilitate communication with artificial artifacts via direct measures of the human brain’s activity.
In cooperation with the Bremen Institute of Automation 4 , classified steady-state
visual evoked potentials (SSVEPs) in brain activity have been used to derive
qualitative directional navigation commands. The overall system that has been
first described in [16] projects the given commands onto a frequently updated
route graph representation of the environment. The metrical target locations
deduced are subsequently navigated to by the application of the well-established
Nearness Diagram Navigation method (cf. section 4.3).
4
Research Topics
During development of Rolland, we have done research in many areas, some of
which are traditional topics in mobile robotics such as mapping, self-localization,
or navigation. Others are more specific to a wheelchair, i. e., a device that carries
a person who is also partially controlling it (shared control ).
4.1
Mapping and Self-Localization
Self-Localization was investigated along routes, in route graphs, and in metric maps. The latter was done in combination with mapping, i. e. simultaneous
localization and mapping (SLAM).
4
Contact: A. Gräser, and T. Lüth. Institute of Automation, 28359 Bremen, Germany.
E-Mail:{ag,thorsten.lueth}@iat.uni-bremen.de.
260
Localization Along Routes. In [23–26] a simple approach for the acquisition and representation of spatial knowledge needed for controlling a semiautonomous wheelchair is presented. Simplicity is required in the domain of
rehabilitation robotics because typical users of assistive technology are persons
with severe impairments who are not technical experts. The approach proposed
is a combination of carrying out so-called basic behaviors and the analysis of
the wheelchair’s track of motion when performing these behaviors. As a result,
autonomous navigation in the user’s apartment or place of work can be learned
by the wheelchair by teaching single routes between potential target locations.
The work focuses on the analysis of the motion tracks recorded by the vehicle’s
dead reckoning system. As a means for unveiling the structure of the environment while the system is moving, an incremental generalization is applied to the
motion tracks. In addition, it is discussed how two of such generalized motion
tracks are matched to perform a one-dimensional self-localization along the route
that is followed.
RouteLoc – Self-Localization in Route Graphs. RouteLoc is a selflocalization approach that needs only minimal input (i. e. no expensive proximity
sensors are required) to absolutely localize a robot even in large-scale environments [27–30]. RouteLoc works robustly in structured large-scale environments.
It can be seen as an extension of the approach discussed above to localize in maps
instead of a single route. As input data, RouteLoc requires a topological-metric
map and a so-called incremental route generalization. It calculates a position
estimate in a human-compatible form, e. g. “The robot is in the corridor leading
from A to B, about x meters past A”. In [27], RouteLoc solves, among others,
the so-called “kidnapped robot problem” in real world experiments with Rolland
on the campus of the Universität Bremen.
Scan-Matching-based SLAM. In [32–34, 31] an approach to generate consistent maps in real-time using a laser range sensor is presented. It improves an
existing scan-matching algorithm in several aspects. It also introduces a mapping method that is based on the scan matching method. Using a laser scanner
mounted on the wheelchair, it generates consistent maps, i. e. maps that do not
have discontinuities resulting from the accumulation of errors in the mapping
process. As such accumulations cannot be avoided, the resulting errors are corrected whenever the wheelchair returns to areas already mapped. An important
property of the approach is its real-time capability. The mapping is performed
on the fly and thus, the resulting maps can immediately be employed for selflocalization. An example of the mapping process is given in Fig. 3.
4.2
Behavior-based Navigation
Behavior-based Navigation, in contrast to metric navigation and planning, is
a reactive way to navigate. Usually, the representation of the environment is
not metric. Instead, features of the environment are associated with necessary
261
Fig. 3. Four snapshots while mapping an office floor.
changes in the behavior. Two approaches have been investigated: image-based
homing and navigation using basic behaviors.
Image-based Homing. In [17–19, 35–38] an image processing method is presented that enables a robot to orientate in unchanged, existing environments.
The approach is based on the use of one-dimensional 360◦ (panoramic) color
images that are taken by a special optical sensor. From two of these images,
the Panama algorithm determines the spatial relationship between the positions
where the images have been taken. The approach can be compared to methods
that determine the optical motion flow but it addresses a different application: it
is used to determine the spatial relations between positions that may be located
several meters apart. The image processing method is embedded in a navigation technique in which the environment is represented as a network of routes
along which the navigating agent can move. These routes are represented as sequences of panoramic images. This route knowledge is acquired by teaching. A
teacher only presets the routes that should be learned but not the images that
are required for their description. These images are independently selected by
the autonomous system during the training, depending on the particular environment and the system’s kinematic restrictions.
Basic Behaviors. In [17, 20–22] an approach is presented that describes routes
as sequences of basic behaviors, such as wall-centering, wall-following, entering
a door, etc. The recognition of routemark constellations triggers the switching
between these behaviors. The information when to use which basic behavior and
when to switch to the next one is acquired during a teaching phase (cf. Fig. 4).
The method detects possible navigational errors and tries to recover from them
by backtracking the previous way.
4.3
Metric Navigation
Local navigation approaches that provide the ability to deal with dynamic and
unforeseen obstacles typically employ a sensor-based map of the robot’s local
262
Behaviors
wall-centering
wall-following left
wall-following right
stop
Routemarks
X routemark X
routemark
constellation
7
5
1
9
2
8
6
3
2
Fig. 4. Teaching of routes as an association of routemark constellations and basic
behaviors.
neighborhood for planning purposes. By means of this data-structure, geometric path planning algorithms and reactive steering methods either calculate a
complete trajectory that leads the vehicle to the designated target, or just determine a set of velocities to be applied in the next time frame. In the following
we present local navigation approaches implemented on Rolland during the last
15 years.
The Nearness Diagram Navigation approach (NDN), initially developed by
Minguez et al., is a representative of the class of reactive steering methods. NDN
describes the sensor measurements of the visible part of the environment along
with the desired target position as a unique element of an exclusive and complete
set of situations. Each of the five situations originally defined is associated with
a specific control law that determines the translational and rotational speed to
be applied, as well as the desired direction of movement. Since Rolland is an
example for a non-circularly shaped mobile robot that cannot independently
control its kinematic degrees of freedom, we have augmented the basic NDN
control laws in such a way that they look ahead for upcoming turn maneuvers
and initiate a preparatory sheer out movement when necessary (cf. [39] for an
in-depth discussion).
Behavior-Based Local Navigation splits the local navigation problem into
two parts. The first one is to approach the goal and the second one is to avoid
obstacles on the way. The first part, i. e. the local navigation module, simply
determines the actual driving command from the current pose of the wheelchair
and a local target pose in each control cycle. A pose consists of the three components (x, y, rotation) in global coordinates. The speeds determined are used to
control the Driving Assistant that is described in section 5.2. The local navigation generates motion commands as if the wheelchair would drive in free space,
263
(a) Illustration of the Nearness Diagram Navigation approach within a simulated environment. Sod and Srd denote the borders of the free walking area, while sgoal and
θ describe the direction to the local target pg and the current orientation of the
wheelchair respectively (cf. [39]).
(b) The Bezier-Curve Path (c) The A* based Path Planning approach
Planner selects an optimal
searches the discretized configuration space
path (red) from a set of
(x,y,θ) for an optimal path. Atomic transpossible paths (green) by
lational moves are illustrated as green lines,
reducing the time of travel
while atomic rotational moves are painted
and maximizing the disas red lines.
tance to nearby obstacles.
Fig. 5. Illustration of three local navigation approaches implemented on Rolland. While
the Nearness Diagram Navigation belongs to the class of reactive navigation strategies,
the Bezier-Curve Path Planner, and the A* based Path Planning approach are geometric and complete planning algorithms respectively.
i. e., it completely ignores the current obstacle situation. However, since the
Driving Assistant avoids obstacles, the wheelchair is robustly able to navigate
even in narrow environments, as long as the target poses are selected appropriately. In general, target poses should be located at narrow passages, e.g. between door posts, with the orientation perpendicular to the narrowness. In case
264
the wheelchair gets stuck or, for any reason, misses the target pose, a shunting
behavior allows retrying and thereby ensures robustness.
Geometric Path Planning. The basic principle of geometric path planning
approaches is to continuously check sets of mathematical curves for a solution
path in order to move the robot from its current pose to a given goal-pose.
Therefore the algorithms evaluate a cost function that minimizes the length of
the path to execute, while maximizing the distance to the obstacles along the
selected path. For Rolland several such algorithms have been developed, applying
circular arc segments, clothoids, and cubic Bezier splines [40, 41, 14].
A* Path Planning. Recent developments have led to a prototypical implementation of a real-time-capable planner based on A*-search in Rolland’s configuration space (x, y, θ). Conceptional drawbacks of NDN and the geometric path
planning approaches such as the non-consideration of target orientations, or the
sparse covering of the configuration space by means of mathematical curves
are avoided. For this reason a typical A*-based planning scenario involves the
discretisation of the configuration space into 6 ∗ 106 cells, yielding a spatial resolution of 3 ∗ 3cm2 , and an orientational resolution of 3.75◦ . The algorithmic
search itself is guided by two heuristics taking the current obstacle situation
and the non-holonomic nature of Rolland into account. Hereby it is possible to
maintain planning-cycles of approximately 20ms.
4.4
Shared Control
The shared-control aspect of manned service robots is to be considered in all
applications of Rolland that share control between the human driver and an
intelligent assistance module. So-called “mode confusion” situations have to be
avoided [42, 43]. In the aviation psychology community, the problem of mode confusion has already been discussed for about a good decade. However, the notion
as such has never been rigorously defined. In addition, the pertinent publications
so far cover almost exclusively the pilot-autopilot interaction. In [27, 44], a rigorous view of mode confusion is presented. A framework based on existing formal
methods is established for separately modeling the technical system, the user’s
mental representation of it, and their safety-relevant abstractions. As a result,
an automated model checking approach can be applied to detect mode confusion
potential already in the design phase. In a case study, the obstacle avoidance
skill of Rolland is checked for mode confusion potential with tool support.
4.5
Qualitative Control
In order to solve high level tasks such as the interpretation of qualitative driving
commands or coarse qualitative route descriptions, one needs an appropriate
global data structure that represents the navigable space in an adequate way. For
Rolland we apply a graph structured representation, i. e. the Route Graph [45],
265
because of its extensibility by annotations, given either by humans or automatic
feature detectors, and because of its interface that supplies basic spatial queries
as a foundation for the evaluation of more elaborated spatial tasks.
The underlying idea of the interpretation of Coarse Verbal Route Descriptions
(CRDs) is that they can be split into route segment descriptions that describe
the start, the progression along the route, intermediate reorientation actions,
and the goal. This observation has been validated by empirical studies in which
subjects had to give in-advance route descriptions to a designated goal. For the
evaluation of CRDs, i. e. the calculation of the most likely target pose, we utilize
a search tree the nodes of which represent fuzzy rated places that result from
the evaluation of single CRD-elements. The corresponding algorithm inputs a
formalized CRD representing the current task description, a global route graph
which holds the available world knowledge, and the global pose of the system.
By evaluating fuzzy functions that model the consecutively appearing spatial
relations in the CRD, our algorithm builds up the mentioned search tree and
searches it in a depth-first branch-and-bound manner to determine the most
probable goal pose. For a more detailed elaboration of this work see [14, 15].
5
Applications
Using the techniques discussed above, several assistants have been developed for
Rolland [46]: the Safety Assistant, the Driving Assistant, the Route Assistant,
the (Autonomous) Navigation Assistant, and the Multi Modal Driving Assistant.
5.1
Safety Assistant
The Safety Assistant mainly consists of a safety layer that ensures that the
vehicle will stop in time before a collision can occur [47–50]. 30 or 50 times per
second (depending on the wheelchair model), the safety layer makes a binary
decision. Either the current driving command is safe, and it can be sent to the
wheelchair, or it is not, and the wheelchair has to stop instead. “Safe” means that
if a stop command would be initiated in the next processing cycle, the wheelchair
would still be able to stop without a collision. Otherwise, it has to be stopped in
this cycle, because in the next cycle it would be too late. Whether the wheelchair
can stop in time depends on the actual speeds of the two drive wheels and the
current drive command, because it will influence the current speeds in the future,
the shape of the wheelchair, and its current surroundings. The surroundings are
measured using the distance sensors (sonars or laser scanners) and a model of
the environment is maintained in a local obstacle map (cf. Fig. 6). Based on the
current speeds and the commanded speeds, a safety area is searched for obstacles
in the map. If the safety area is free of obstacles, the current driving command
is safe. Since the shape of such a safety area is rather complex, a large number
of safety areas were pre-computed and stored in a lookup table.
266
Fig. 6. Intervention of the Driving Assistant. The driver presses the joystick straight
ahead (indicated by the thick line); the Driving Assistant detects an obstacle on the
right side (cf. safety area in the local obstacle map above), and avoids it to the left
(thin line).
5.2
Driving Assistant
The Driving Assistant is a software component that provides obstacle avoidance
for Rolland [51, 5, 47, 48, 52]. The Driving Assistant avoids obstacles while the
wheelchair is still controlled by the user. The user remains in complete control
of the system as long as there are no obstacles on the path driven. Whenever
obstacles block the way and would result in the wheelchair being stopped by the
Safety Layer (the main part of Safety Assistant), the Driving Assistant takes
control and avoids the obstacle with as small deviations from the commands
given by the user as possible (cf. Fig. 6). The Driving Assistant on Rolland II
also provides a basic behavior for turning on the spot, because this can be quite
difficult with an Ackermann steering in a narrow environment. Since Rolland
III and IV are equipped with a differential drive, there is no need for such a
behavior.
5.3
Route Assistant
The Route Assistant of the Bremen Autonomous Wheelchair [53] has been developed in cooperation with the neurological clinic of a Bremen hospital. It provides the following functionality: During a teaching phase, the system explores
the routes and places pertinent for the future user(s). If, e. g., the wheelchair is
used in a rehabilitation center for amnesic patients, the routes to all relevant
places in the building could be learned and stored for later replay with the help
of the route generalization algorithm already mentioned. In the replay mode,
a nurse chooses a certain target for the patient in the wheelchair. Similar to a
GPS-based navigation system, the large-scale navigation is done by the Route
Assistant by giving instructions where to go at decision points, enabling the
267
Fig. 7. Multi Modal Driving Assistant: Illustration of navigable paths along with their
labelled target nodes (left), and a photograph of the corresponding environment (right).
While the left figure presents the current pose of the wheelchair by a detailed shape
that is heading to the top, the picture to the right depicts the position and orientation
of the wheelchair by a green dot and arrow.
patient to travel around on his or her own. The patient is independently responsible for controlling the vehicle with respect to local maneuvers such as obstacle
avoidance.
5.4
(Autonomous) Navigation Assistant
The Navigation Assistant goes further. It is based on a SLAM-based metric representation of the environment with an embedded route graph. The aim of the
system is to have a robust navigation component for an autonomous wheelchair
within a global frame of reference. The user is enabled to choose a specific destination – e. g. “Refrigerator”, “Entrance”, or “Wardrobe” – from a predefined list.
Given the destination, the system provides two different use cases: autonomous
navigation of the robot without any need for interventions by the user, or a
guidance mode which lets the user drive manually but provides information
about where to navigate next. The latter assistant is similar to the route assistant, but here the spatial representation is a map, not a route. The key idea
behind this assistant is to specify all possible destinations together with all possible routes between these destinations, i. e. the route graph, in advance. This
approach shifts parts of the intelligence needed from the robot to an external
human expert. Thereby, it can be assured that necessary routes always exist
and lead along paths that are actually usable by a wheelchair, i. e., they do
not contain any undetectable obstacles. Especially the latter cannot be assured
for autonomous route generation given the current sensorial equipment of the
wheelchair Rolland.
268
5.5
Multi Modal Driving Assistant
The essential idea behind the Multi Modal Driving Assistant [12] (cf. Fig. 7) is
a simple speech interface with which the operator of the wheelchair selects a
desired movement from a set of graphically presented navigable paths. Derived
from the route graph representation of the environment, the proposed paths
are selected by verbally calling their name. In order to improve the recognition
process of the applied speech recognizer VoCon, each proposed path is labelled by
a token coming from an alphabet with multisyllabic words that provide mutually
different sounds. Once established the connection between the commanded token
and the corresponding path, its endpoint is forwarded to the obstacle avoiding
navigation module, i.e. a specialized implementation of the Nearness Diagram
Navigation approach (cf. section 4.3).
6
Conclusion
It was a long way from Rolland I, the construction of which is so complex that
even healthy people have problems sitting down in it, to Rolland IV that only
stands out because of its functionality, but not because of its appearance. Hence,
the commercialization of the results is the next logical step. In cooperation with
our new partner Otto Bock Mobility Solutions, we will develop some of the
assistants presented in this paper to industrial prototypes. However, there are
still interesting areas for further wheelchair research, e. g., driving outdoors. In
addition, we are investigating another device meant for supporting the elderly:
the intelligent walker, i. e. the iWalker.
References
1. Lankenau, A., Röfer, T.: Smart wheelchairs - state of the art in an emerging market.
Künstliche Intelligenz. Schwerpunkt Autonome Mobile Systeme (4) (2000) 37–39
2. Mandel, C.: Navigation of the Smart Wheelchair Rolland. PhD thesis, University
of Bremen (2008)
3. Röfer, T.: Strategies for using a simulation in the development of the Bremen
Autonomous Wheelchair. In Zobel, R., Moeller, D., eds.: Simulation-Past, Present
and Future, Society for Computer Simulation International (1998) 460–464
4. Bühlmeier, A., Kollmann, J., Krieg-Brückner, B., Röfer, T., eds.: Studentisches
Projekt SAUS: Sensomotorik autonomer Systeme. Informatik Bericht. ISSN 07228996. Universität Bremen (1998)
5. Lankenau, A., Röfer, T.: A safe and versatile mobility assistant. Reinventing the
Wheelchair. IEEE Robotics and Automation Magazine (7) (2001) 29–37
6. Röfer, T., Lankenau, A.: Ensuring safe obstacle avoidance in a shared-control
system. In Fuertes, J.M., ed.: Proceedings of the 7th International Conference on
Emergent Technologies and Factory Automation (ETFA-99). (1999) 1405–1414
7. Röfer, T., Lankenau, A.: Ein Fahrassistent für ältere und behinderte Menschen. In
Schmidt, G., Hanebeck, U., Freyberger, F., eds.: Autonome Mobile Systeme 1999.
Informatik aktuell, Springer (1999) 334–343
269
8. Lankenau, A., Röfer, T.: Architecture of the Bremen Autonomous Wheelchair.
In Hildebrand, B., Moratz, R., Scheering, C., eds.: Architectures in Cognitive
Robotics. Technical Report. Number 98/13, SFB 360 ”Situierte Künstliche Kommunikatoren”. Universität Bielefeld (1998) 19–24
9. Gollub, J.: Umsetzung und Evaluation eines mit Laserscannern gesicherten Rollstuhls als interaktives Museumsexponat. Diplomarbeit, Fachbereich 3, Universität
Bremen (2007)
10. Mandel, C., Röfer, T., Frese, U.: Applying a 3DOF orientation tracker as a humanrobot interface for autonomous wheelchairs. In: Proceedings of the 10th International Conference on Rehabilitation Robotics. (2007)
11. Mandel, C., Frese, U., Röfer, T.: Design improvements for proportional control
of autonomous wheelchairs via 3DOF orientation tracker. In: Proceedings of the
9th International Work-Conference on Artificial Neural Networks (IWANN’2007).
Lecture Notes in Computer Science, Springer; Berlin (2007)
12. Mandel, C., Frese, U.: Comparison of wheelchair user interfaces for the paralysed:
Head-joystick vs. verbal path selection from an offered route-set. In: Proceedings
of the 3rd European Conference on Mobile Robots (ECMR 2007). (2007)
13. Müller, R., Röfer, T., Lankenau, A., Musto, A., Stein, K., Eisenkolb, A.: Coarse
qualitative descriptions in robot navigation. In Freksa, C., Brauer, W., Habel,
C., Wender, K.F., eds.: Spatial Cognition II. Number 1849 in Lecture Notes in
Artificial Intelligence, Springer (2000) 265–276
14. Mandel, C., Frese, U., Röfer, T.: Robot navigation based on the mapping of coarse
qualitative route descriptions to route graphs. In: Proceedings of the IEEE/RSJ
International Conference on Intelligent Robots and Systems (IROS 2006). (2006)
205–210
15. Shi, H., Mandel, C., Ross, R.J.: Interpreting route instructions as qualitative
spatial actions. In Barkowsky, T., Knauff, M., Ligozat, G., Montello, D.R., eds.:
Spatial Cognition V. Volume 4387 of Lecture Notes in Computer Science., 14197
Berlin, Springer Verlag (2008)
16. Graimann, B., Allison, B., Mandel, C., Lueth, T., Valbuena, D., Gräser, A.: Noninvasive Brain-Computer Interfaces for Semi-Autonomous Assistive Devices. In:
Robust Intelligent Systems. Springer Verlag (2007) to appear.
17. Röfer, T.: Panoramic Image Processing and Route Navigation. Number 7 in BISS
Monographs. Shaker-Verlag (1998)
18. Röfer, T.: Route navigation and panoramic image processing. In: Ausgezeichnete
Informatikdissertationen 1998. B. G. Teubner; Stuttgart, Leipzig (1999) 132–141
19. Röfer, T.: Route navigation and panoramic image processing. Künstliche Intelligenz. Schwerpunkt Autonome Mobile Systeme (1) (2000) 62–64
20. Krieg-Brückner, B., Röfer, T., Carmesin, H.O., Müller, R.: A taxonomy of spatial knowledge for navigation and its application to the Bremen Autonomous
Wheelchair. In Freksa, C., Habel, C., Wender, K.F., eds.: Spatial Cognition. Number 1404 in Lecture Notes in Artificial Intelligence, Springer (1998) 373–397
21. Röfer, T.: Routemark-based navigation of a wheelchair. In: Proceedings of the 3rd
ECPD International Conference on Advanced Robotics, Intelligent Automation
and Active Systems. (1997) 333–338
22. Röfer, T., Müller, R.: Navigation and routemark detection of the Bremen Autonomous Wheelchair. In Lüth, T., Dillmann, R., Dario, P., Wörn, H., eds.: Distributed Autonomous Robotics Systems, Springer (1998) 183–192
23. Röfer, T.: Routenbeschreibung durch Odometrie-Scans. In Wörn, H., Dillmann,
R., Henrich, D., eds.: Autonome Mobile Systeme 1998. Informatik aktuell, Springer
(1998) 122–129
270
24. Musto, A., Stein, K., Eisenkolb, A., Röfer, T.: Qualitative and quantitative representations of locomotion and their application in robot navigation. In: Proceedings
of the 16th International Joint Conference on Artificial Intelligence (IJCAI-99),
Morgan Kaufman Publishers, Inc; San Francisco, CA (1999) 1067–1073
25. Röfer, T.: Route navigation using motion analysis. In Freksa, C., Mark, D.M.,
eds.: Proceedings of the Conference on Spatial Information Theory (COSIT-99).
Number 1661 in Lecture Notes in Computer Science, Springer (1999) 21–36
26. Musto, A., Stein, K., Eisenkolb, A., Röfer, T., Brauer, W., Schill, K.: From motion
observation to qualitative motion representation. In Freksa, C., Brauer, W., Habel,
C., Wender, K.F., eds.: Spatial Cognition II. Number 1849 in Lecture Notes in
Artificial Intelligence, Springer (2000) 115–126
27. Lankenau, A.: The Bremen Autonomous Wheelchair ‘Rolland’: Self-Localization
and Shared Control. PhD thesis, Universität Bremen (2002)
28. Lankenau, A., Röfer, T., Krieg-Brückner, B.: Self-localization in large-scale environments for the Bremen Autonomous Wheelchair. In Freksa, C., Brauer, W.,
Habel, C., Wender, K.F., eds.: Spatial Cognition III. Number 2685 in Lecture Notes
in Artificial Intelligence, Springer (2003) 34–61
29. Lankenau, A., Röfer, T.: Mobile robot self-localization in large-scale environments.
In: Proceedings of the IEEE International Conference on Robotics and Automation
2002 (ICRA-2002), IEEE (2002) 1359–1364
30. Röfer, T., Lankenau, A.: Route-based robot navigation. Künstliche Intelligenz Themenheft Spatial Cognition (2002) 29–31
31. Kollmann, J., Röfer, T.:
Echtzeitkartenaufbau mit einem 180◦ -LaserEntfernungssensor. In Dillmann, R., Wörn, H., von M. Ehr, eds.: Autonome Mobile
Systeme 2000. Informatik aktuell, Springer (2000) 121–128
32. Röfer, T.: Konsistente Karten aus Laser Scans. In Levi, P., Schanz, M., eds.:
Autonome Mobile Systeme 2001. Informatik aktuell, Springer (2001) 171–177
33. Röfer, T.: Building consistent laser scan maps. In: Proceedings of the 4th European Workshop on Advanced Mobile Robots (Eurobot 2001). Volume 86., Lund
University Cognitive Studies (2001) 83–90
34. Röfer, T.: Using histogram correlation to create consistent laser scan maps. In:
Proceedings of the IEEE International Conference on Robotics Systems (IROS2002), EPFL; Lausanne, Switzerland (2002) 625–630
35. Röfer, T.: Controlling a wheelchair with image-based homing. In: Spatial Reasoning in Mobile Robots and Animals, AISB-97 Workshop, Manchester University
(1997) 66–75
36. Röfer, T.: Image based homing using a self-organizing feature map. In FogelmanSoulie, F., Gallinari, P., eds.: Proceedings of the International Conference on Artificial Neural Networks (ICANN-95). Volume 1., EC2 & Cie (1995) 475–480
37. Röfer, T.: Bildbasierte Navigation mit eindimensionalen 360◦ -Bildern. In Dillmann, R., Rembold, U., Lüth, T., eds.: Autonome Mobile Systeme 1995. Informatik
Aktuell, Springer (1995) 193–202
38. Röfer, T.: Controlling a robot with image based homing. In Krieg-Brückner,
B., Herwig, C., eds.: Tagungsband des Workshops ”Kognitive Robotik”. Number
3/95 in ZKW Bericht, Zentrum für Kognitionswissenschaften. Universität Bremen
(1995)
39. Kraetsch, P.: Entwicklung einer reaktiven steuerung für mobile roboter auf basis
der nearness-diagram-methode für navigation in innenraum̈en. Master’s thesis,
Universität Bremen (2007)
271
40. Mandel, C.: Trajektorienplanung und Trajektorienfolgeregelung im Konfigurationsraum nicht-holonomer Fahrzeuge. Diplomarbeit, Fachbereich 3, Universität
Bremen (2002)
41. Mandel, C., Huebner, K., Vierhuff, T.: Towards an autonomous wheelchair: Cognitive aspects in service robotics. In: Proceedings of Towards Autonomous Robotic
Systems (TAROS 2005). (2005) 165–172
42. Lankenau, A.: Avoiding mode confusion in service-robots. In Mokhtari, M., ed.:
Integration of Assistive Technology in the Information Age, Proc. of the 7th Int.
Conf. on Rehabilitation Robotics, IOS Press (2001) 162 – 167
43. Lankenau, A., Röfer, T.: The role of shared control in service robots - the Bremen
Autonomous Wheelchair as an example. In Röfer, T., Lankenau, A., Moratz, R.,
eds.: Service Robotics - Applications and Safety Issues in an Emerging Market.
Workshop Notes. (2000) 27–31
44. Bredereke, J., Lankenau, A.: A rigorous view of mode confusion. In: Proc. of Safecomp 2002, 21st Int’l Conf. on Computer Safety, Reliability and Security. Number
2434 in Lecture Notes in Computer Science, Springer-Verlag; D-69121 Heidelberg,
Germany (2002) 19–31
45. Krieg-Brückner, B., Frese, U., Lüttich, K., Mandel, C., Mossakowski, T., Ross, R.:
Specification of an ontology for route graphs. In Freksa, C., Knauff, M., KriegBrückner, B., Nebel, B., Barkowsky, T., eds.: Spatial Cognition IV. Volume 3343
of Lecture Notes in Artificial Intelligence. Springer-Verlag, D-69121 Heidelberg,
Germany (2005) 390–412
46. Krieg-Brückner, B., Shi, H., Fischer, C., Röfer, T., Cui, J., Schill, K.: Welche
Sicherheitsassistenz brauchen Rollstuhlfahrer? In: 2. Deutscher AAL-Kongress
2009, VDE-Verlag; Berlin-Offenbach, Germany
47. Röfer, T., Lankenau, A.: Architecture and applications of the Bremen Autonomous
Wheelchair. In Wang, P., ed.: Information Sciences. Volume 1-4. Elsevier Science
BV (2000) 1–20
48. Röfer, T., Lankenau, A.: Architecture and applications of the Bremen Autonomous
Wheelchair. In Wang, P.P., ed.: Proceedings of the 4th Joint Conference on Information Systems. Volume 1., Association for Intelligent Machinery (1998) 365–368
49. Lankenau, A., Meyer, O.: Formal methods in robotics: Fault tree based verification.
In: Proc. of Quality Week Europe. (1999)
50. Lankenau, A., Meyer, O.: Der autonome Rollstuhl als sicheres eingebettetes System. Diplomarbeit, Fachbereich 3, Universität Bremen (1997)
51. Lankenau, A., Meyer, O., Krieg-Brückner, B.: Safety in robotics: The Bremen
Autonomous Wheelchair. In: Proceedings of the 5th Int. Workshop on Advanced
Motion Control (AMC ’98). (1998) 524 – 529
52. Lankenau, A., Röfer, T.: Rollstuhl ¨‘Rolland¨’ unterstützt ältere und behinderte
Menschen. FIfF-Kommunikation. Informationstechnik und Behinderung (2) (2000)
48–50
53. Lankenau, A., Röfer, T.: Selbstlokalisation in Routengraphen. In Levi, P., Schanz,
M., eds.: Autonome Mobile Systeme 2001. Informatik aktuell, Springer (2001) 157–
163
272
Thomas Röfer received his diploma in computer science and his Dr.ing. degree under the supervision of
Bernd Krieg-Brückner from Universität Bremen, Germany, in 1993 and 1998, respectively. He is in the Executive Committee of the RoboCup Federation, and he
is member of the Transregional Collaborative Research
Center SFB/TR 8 “Spatial Cognition” at Bremen. He
is currently with the German Research Center for Artificial Intelligence at the research group for Safe and
Secure Cognitive Systems, Bremen, Germany. His research interests include rehabilitation robotics, robot
soccer, real-time computer vision, world modeling, and
humanoid robots.
Christian Mandel received the Dipl.Inf. and Dr.Ing.
degrees in computer science under the supervision
of Bernd Krieg-Brückner from Universität Bremen,
Germany, in 2002 and 2008, respectively. In 2008 he
worked for the German Research Center for Artificial
Intelligence at the research group for Safe and Secure
Cognitive Systems, Bremen, Germany. He is currently
with the University of Bremen at the Collaborative Research Center Spatial Cognition, doing research and
development on navigation techniques and advanced
interface devices for assistive devices, such as electrical wheelchairs.
Axel Lankenau received the Dipl.Inf. and Dr.Ing.
degrees in computer science under the supervision of
Bernd Krieg-Brückner from Universität Bremen, Germany, in 1997 and 2002, respectively. He was member
of the Rolland project from 1994 until 2002. Currently,
he is a manager for documentation process design in
the Research and Advanced Engineering department
of the Daimler AG in Sindelfingen, Germany.
Bernd Gersdorf received the Dipl.Inf. and Dr.Ing.
degrees in computer science under the supervision of
Bernd Krieg-Brückner from Universität Bremen, Germany, in 1986 and 1992, respectively. In 1995, he
started to work in the industry in the software development for Process Automation, and later for flight
safety systems. He joined the Rolland project early
2007 with the German Research Center for Artificial
Intelligence at the research group for Safe and Secure
Cognitive Systems, Bremen, Germany.
Udo Frese received his Ph.D. from University of
Erlangen-Nürnberg in 2004. From 2004 until 2008
he has been a post-doc researcher in Bernd KriegBrückner’s lab at Universität Bremen and in the
German Research Center for Artificial Intelligence
(DFKI). He is now assistant professor for real-time
computer vision at University of Bremen and further
on affiliated with the DFKI. His research topics are
real time computer vision, in particular simultaneous
localization and mapping (SLAM), and computer vision for sport robotics.
On Classical and Quantum Process Algebra as Abstract Data Types
(Detailed Abstract)
Dedicated to the 60th birthday of Prof. Bernd Krieg-Brückner
Ruqian Lu1,2,3, Lixing Li1, Yun Shang1, Xiaoyu Li4
1
CAS Key Lab of MADIS, Academy of Mathematics and Systems Science, CAS
2
Shanghai Key Lab of Intelligent Information Processing, Fudan University,
3
CAS Key Lab of IIP, Institute of Computing Technology, CAS
4
School of Information Engineering, Zhengzhou University
Keywords: process algebra, quantum process algebra, quantum programming language, formal
semantics, algebraic semantics, abstract data type
$ 1 Formal Semantics of Quantum Process Algebras
1.1 Quantum process algebra and its semantics
Together with the prospect of building quantum computers for real and practical use,
scientists have been also interested in studying theories and techniques of quantum programming
languages. Some have imperative character [Kni96][BCS03][Ome00] [Xu08] [Zul01] [Mln06], some others are
functional [Sel04b][Ton04][SV05][AG05]. For a survey of both directions see [Gay06] and [Sel04a].
Quantum programming languages of a particular kind are called quantum process
algebra[JL04][LJ04][GN05].[F07] [Y08]. There are some subtle points of quantum process algebra. First,
each quantum system is attached with a quantum state space. This will influence the form of
transition rules and of bisimulation. Second, physical constraints of quantum mechanics may
appear in different form than in other quantum programming languages. For example, the quantum
copying problem appears not in form of assignment statements, but in form of input, output and
communication rules.
Regarding the first point, most quantum process algebra proposed in the literature use the
concept of configuration <P, C> to describe the state of a quantum system, where P is the quantum
process and C is the context. For example, a context of QPALg[LJ04] is represented as
s, q
U , f ! , where s is the environment stack declaring all classical and quantum variable
names with types, q the set of quantum variables, U the density matrix characterizing the state of q,
f the classical environment assigning values to classical variables. On the other hand, according to
whether the authors adopt a probabilistic transition system, these operational semantics can be
classified as probabilistic or non-probabilistic. One advantage of the non-probabilistic technique is
avoiding dealing with probability calculation and making the explicit probability implicit, since
any probability is implied in the density matrix. As a trade-off, the non-probabilistic approach can
not process classical data and classical communication in case of quantum measuring. Apart from
[Y08], most authors[LJ04][GN05][F07] have made use of a probabilistic transition system.
Quantum bisimulation is not as widely studied as the transition rule based operational
semantics. It has been introduced in [Lal06], [F07] and [Y08]. The difference of transition systems
has a great impact on the definition of bisimulation. Those accepting probabilistic transitions adopt
naturally the concept of probabilistic bisimulation[Lal06] [F07], while the authors of [Y08], who
-------------------------------------------------*Partially supported by NSFC projects 60496324, 60736011, 60603002, 90818026
274
adopt the non-probabilistic approach, keep also their bisimulation definition non-probabilistic,
where the authors have introduced various kinds of bisimulation, including strong bisimulation,
strong reduction bisimulation and strong approximate bisimulation. Note that the strong
approximate bisimulation is an extension of Mingsheng Ying’s previous result in classical process
algebra[Y01]. Besides, the authors of [Y08] have proved a series of algebraic properties of these
different kinds of bisimulation.
1.2 Abstract data types as program semantics
The idea of abstract data type (ADT) was born in early seventies. Examples of early
important contributions to ADT research are [Zil74], [GH78], [GT75,78], [BW80a,b], for more
reference see [W89]. The advantage of using ADT as algebraic semantics is that it provides not
only a single model of semantics for a given program, but a set of different models, depending on
different choice of axiom systems.
Just as said by Wirsing: “Data stractures, algorithms and programming languages can be
described in a uniform implementation independent way by axiomatic abstract data types”[W89].
The idea of considering programming languages as ADT first appeared in 1980, where Broy &
Wirsing[BW80a,b] and Cleaveland[C80] proposed this idea separately almost at the same time.
Despite the fact that ADT had been used for defining algebraic semantics of various kinds
of classical programs, no work has been reported on using ADT to define algebraic semantics for
quantum process algebras. The authors even don’t know if there is any work done on using ADT to
define classical process algebra semantics. Therefore in this paper we first study ADT as algebraic
semantics for classical process algebra and then generalize the results to quantum process algebra.
1.3 Main contribution of this paper
In this paper we introduced an algebraic semantics for classical and quantum process algebra
in form of ADT. For that purpose, we developed a particular type of 6 algebra, the seed algebra,
which describes exactly the behavior of a process. Each non-hidden element of the seed algebra is a
state of the process. In the seed algebra, generating an element e1 from another element e2 by the
application of a non-hidden operation represents a state transition from e2 to e1 in process calculus.
We have proved that two processes bisimulate each other if and only if their corresponding seed
algebras have a deep rough seed isomorphism. Based on this result, we have given new proofs of
some known results of process algebra in form of theorems of seed algebras. We generalized these
results to study the algebraic semantics of quantum process algebra in section 3. We provided a
detailed ADT characterization of quantum processes and quantum operations and presented similar
results.
$2 ADT as semantics of Process algebra
Assume the usual definitions of abstract data types and 6 algebra are known. We introduce
additionally the ‘hidden’ concept for use in this paper.
Definition 2.1: We differentiate the operations of an ADT as hidden (hop) or non-hidden (op).
Each operation can be either hidden or non-hidden. But no operation can be both hidden and
non-hidden.
Their semantics will be shown later.
Definition 2.2: Given ADT D and 6 algebra A (D). Further assume a is an element of A (D). The
275
seed algebra A (D, a) is defined as follows:
1. a  A (D, a);
2. If for some i, bi  A (D, a) then f (b1 ,..., bi ,...bn )  A (D, a), where
f: s1 u s2 u … u si u … u sn Æ sn+1
is an operation of D, all b1 ,..., bi ,...bn  A (D), each b j is of sort sj.
3. a is called the seed of this algebra,
4. For the case where A (D) is the ground term algebra of D, the seed algebra A (D, a) is also
called ground seed algebra.
In this paper, we only consider ground seed algebra.
Proposition 2.3: Let A (D) be the ground term algebra of D, then:
1. A (D, a) is a sub-algebra of A (D);
2. A (D, a) may be a proper sub-algebra of A (D).
3. For fixed D and a, the seed algebra A (D, a) is uniquely defined.
N
We differentiate between deterministic and non-deterministic objects.
Definition 2.4: 1. All elements of a seed algebra are deterministic objects;
2. A non-deterministic object D of sort s is an abstract entity, which represents any (deterministic)
object of a subset of seed algebra elements of sort s.
3. There are two types of non-deterministic objects: hidden extension and non-hidden closure.
4. The seed of a seed algebra is a non-hidden closure representing itself. Each constant operation is
a non-hidden closure representing itself;
5. For any non-hidden closure
E of sort s, its hidden extension [ E ] represents any legal
element f1 ( f 2 (... f k ( c)...)) , where c is any element represented by E , k t 0 , f1 , f 2 ,..., f k are
hidden functions, all of which have the functionality s Æ s.
6. if f is a non-hidden function, then f ([ E1 ],..., [ E m ]) is a non-hidden closure representing any
legal seed algebra element f (b1 ,..., bm ) , where each bi is represented by [ E i ] , which is the
hidden extension of the non-hidden closure E i .
Definition 2.5: Let A (D, a) = (Ea, Fa) and A (D, b) = (Eb, Fb) be two seed algebras of the same data
type D, where Ea, Eb are element sets; Fb, Fb are function sets; the seeds a, b are of the same sort. A
mapping from the non-hidden closures of A (D, a) to the non-hidden closures of A (D, b), denoted
with hnh, is called a rough seed homomorphism from A (D, a) to A (D, b), where
1. hnh (a) = b;
#According to definition 2.4, all seeds are non-hidden closure#
2. f  Fa , where f is the interpretation of a non-hidden operation O  op, g  Fb , such that
h ( f ) = g , where g is the interpretation of the same O (Henceforth we will simply call f and g as
276
non-hidden operation or non-hidden function);
3. f  Fa , where f is a non-hidden function, we call f ([D1 ],..., [D n ]) a non-hidden closure
of [ D i], where for 1 d i d n , D i are non-hidden closures and [D i ] their hidden extensions, there
are non-hidden g  Fb , and non-hidden closures
E1 ,..., E n , hnh ( f ) = g , such that
hnh ( f ([D1 ],..., [D n ])) = g ([ E1 ],..., [ E n ]) ,
(2.1)
where g ([ E1 ],..., [ E n ]) is a non-hidden closure of [ E i ], 1 d i d n ,
and hnh ( D i ) = E i
#Read:
For any non-hidden closure f ([D1 ],..., [D n ]) , there is a non-hidden closure
g ([ E1 ],..., [ E n ]) , such that i, 1 d i d n , hnh ( D i ) = E i and hnh ( f ([D1 ],..., [D n ])) =
g ([ E1 ],..., [ E n ]) #
4. hnh is called a rough seed isomorphism if it can be uniquely reversed to hnh-1. More exactly, it
requires in addition:
i, h-1nh ( E i ) = D i ,
1
1
hnh (hnh ( f ([D1 ],...,[D n ]))) = hnh ( g ([ E1 ],...,[ E n ])) = f ([D1 ],..., [D n ]) ,
(2.2)
f ([D1 ],..., [D n ]) and its rough homomorph mapping
#Read: For any non-hidden closure
1
hnh ( f ([D1 ],..., [D n ])) , the reverse mapping hnh (hnh ( f ([D1 ],...,[D n ]))) provides the original
non-hidden closure f ([D1 ],..., [D n ])
#
We use the notation nhc (A (D, a)) to denote the set of non-hidden closures of the seed algebra
A (D, a). A rough seed homomorphism (isomorphism) from A (D, a) to A (D, b) is also called a
homomorphism (isomorphism) from nhc (A (D, a)) to nhc (A (D, b)).
Proposition 2.6: The rough seed isomorphism is an equivalence relation.
Proof: straightforward.
N
Consider a simple process grammar:
P ::= nil | y?x.P | y!x.P | P + Q | (P |Q)
x::= a | b | u
y::= c | d
with operational semantics:
Input: c ? u.P o P[a / u ] , Output: c!a.P o P
ca
c!a
277
c!a
ca
P o
P' , Q o
Q'
W
P|Q
o P'| Q'
Tau:
W
W .P 
P
o
Sum:
D
D
P o
P'
P o
P'
, Compo:
D
D
D
P Q o P' , Q P o P'
P | Q o P' | Q, Q | P o
Q | P'
, Commu:
D
where D  {c !a, ca,W } .Since we are using early semantics, the condition bv (D ) , fv (Q )
empty is not needed.
The following hierarchical data type describes this process grammar:
Type C-Process-Cons =
{sort token, chan, proc, trace, ptrace
op
a, b, u: Æ token
c, d: Æ chan
(): Æ trace
nil: Æ proc
tau: proc Æ proc
i-proc: chan u token u proc Æ proc
o-proc: chan u token u proc Æ proc
sum proc u procÆ proc
compo: proc u proc Æ proc
tconf: proc u trace Æ ptrace
}
End of Type C-Process-Cons
Type C-Process-Trans =
{ C-Process-Cons +
sort token, chan, proc, trace, ptrace
op
tinput: ptrace u chan u token Æ ptrace
toutput: ptrace u chan u token Æ ptrace
t-h-act: ptrace Æ ptrace
hop
it: chan u token Æ trace
ot: chan u token Æ trace
run: trace u trace Æ trace
input: proc u chan u token Æ proc
output: proc u chan u token Æ proc
h-act: proc Æ proc
subst: proc u token u token Æ proc
tleft: ptrace Æ ptrace
tright: ptrace Æ ptrace
tback: ptrace Æ ptrace
278
left: proc Æ proc
right: proc Æ proc
compol: proc u proc Æ proc
compor: proc u proc Æ proc
back: proc Æ proc }
In order to avoid too many details, we leave the axiom part to Appendix. Note that we have
introduced a ‘undefined’ predicate for describing partial semantics, e.g. t-h-act (tconf (i-proc (c, u,
p), tr)) = undefined. We do not list them in appendix to save space. Further we introduce the short
notation A’ (D, P)P in process grammar notation =def A (D, tconf (p, ()))p in element of seed algebra notation.
Definition 2.7: Two processes P and Q are trace equivalent if for any action sequence
P1P 2 ...P n and transition sequence:
P1 ... P n
P 
o P '
there is a transition sequence of Q:
P1 ... P n
Q 
o Q'
and vice versa, where we use Greek letters P ,Q ,... to represent actions like c?a, c!b, etc.
Proposition 2.8: Two processes are trace equivalent if and only if their corresponding seed
algebras are rough seed isomorph.
Proof: Given two trace equivalent processes P, Q and their seed algebras A’ (D, P), A’ (D, Q).
Assume
D
f n ([ f n1 (... f1 (tconf ( p, tr )...)...)]...)  nhc (A’ (D, P)),
where p is the algebraic form of P and tr is any trace. It corresponds to the transition sequence:
P1P 2 ... P n
( P, tr ) 

o( P' , tr ' )
where
Pi  {c ? u, c!a,W } are notations of process action for f i . By definition of trace equivalence,
there is a transition sequence
1P 2 ... P n
(Q, tr ) P

o(Q' , tr ' )
which corresponds to
E
f n ([ f n1 (... f1 (tconf (q, tr )...)...)]...)  nhc (A’ (D, Q)),
In this way we get a rough seed isomorphism
hnh: nhc (A’ (D, P)) Æ nhc (A’ (D, Q))
with hnh ( D ) =
E . It is really a rough seed isomorphism since it satisfies all conditions of
definition 2.5. This shows the truth of the ‘only if’ part of the proposition. Similarly we can prove
the ‘if’ part.
N
Definition 2.9: Given two seed algebras A (D, a) and A (D, b), which are rough seed isomorph.
Assume elements c  A (D, a), d  A (D, b). (c, d) is called a corresponding pair if and only if there
279
are non-hidden closures D of nhc (A (D, a)) and E of nhc (A (D, b)), such that hnh ( D ) = E and c
and d are represented by D and E , respectively.
Definition 2.10: Two seed algebras A (D, a) and A (D, b) are called deep rough seed isomorph (or
simply: deep isomorph) if and only if:
1. A (D, a) and A (D, b) are rough seed isomorph by a mapping hnh: nhc (A (D, a)) Æ nhc (A (D, b));
2. For each c  A (D, a), there is an element d  A (D, b), such that (c, d) is a corresponding pair, A
(D, c) and A (D, d) are deep rough seed isomorph.
The following proposition shows that the second condition of the above definition can be
weakened.
Proposition 2.11:
Two seed algebras A (D, a) and A (D, b) are deep isomorph if and only if
1. A (D, a) and A (D, b) are rough seed isomorph by the mapping hnh: nhc (A (D, a)) Æ nhc (A (D,
b));
2. For any element c  A (D, a) there is an element d  A (D, b), such that (c, d) is a corresponding
pair, A (D, c) and A (D, d) are rough seed isomorph with the same mapping hnh.
Proof: First we prove that the condition presented in the statement of this proposition is necessary.
That means if A (D, a) and A (D, b) fulfill the conditions in definition 2.5, then they also have the
property stated in (the second part of) this proposition. That there is a rough seed homomorphism
hnh’: nhc (A (D, c)) Ænhc (A (D, d)) is trivial, since deep isomorphism implies rough seed
isomorphism. What we have to do is to prove that hnh’ = hnh, i.e. the new seed algebra A (D, c) is a
sub-algebra of A (D, a) and A (D, d) a sub-algebra of A (D, b). Assume there are elements u  A (D,
c) and v  A (D, d), such that (u, v) is a corresponding pair. According to the definition of seed
algebra there are operation symbols {fui}, {gvj} of D, such that
u = fu1 (…, fuk (…, c, …), …)
v = gv1 (…, gvq (…, d, …), …)
Remember that we always assume that each operation has only one implementation. Therefore we
can identify operations with functions. Again according to the definition of seed algebra there are
operation symbols {fci}, {gdj} of D, such that
c = fc1 (…, fcn (…, a, …), …)
d = gd1 (…, gdm (…, b, …), …)
Combine the above equations, we have
u = fu1 (…, fuk (…, fc1 (…, fcn (…, a, …), …), …), …)
v = gv1 (…, gvq (…, gd1 (…, gdm (…, b, …), …), …), …)
This shows that u  A (D, a) and v  A (D, b), i.e.
A (D, c) Ž A (D, a), A (D, d) Ž A (D, b)
This shows that A (D, c) is a sub-algebra of A (D, a) and A (D, d) a sub-algebra of A (D, b).
Therefore we may use the same rough seed isomorphism hnh for the mapping nhc (A (D, c)) Æ nhc
(A (D, d)).
Now we prove that the condition presented in the statement of this proposition is also
sufficient. If the second condition of proposition 2.11 were not true, then either there is no
corresponding pair (c, d) such that A (D, c) and A (D, d) are rough seed isomorph, or for some
element u  A (D, c), there is no v  A (D, d), such that (u, v) is a corresponding pair and A (D, u)
280
and A (D, v) are rough seed isomorph. Reasoning repeatedly in this way, we would get two chains
of seed algebras. We use the notation A < B to denote that A is a seed algebra constructed with a
seed from B. Then we would have …… < A (D, cn) <A (D, cn-1) <…<A (D, c1) <A (D, c0),
and ……< A (D, dn) <A (D, dn-1) <…<A (D, d1) <A (D, d0), where c = c0, d = d0, such that for all
1 d i d n, ci  A (D, ci-1) and di  A (D, di-1). Since A (D, c0) and A (D, d0) are known to be rough
seed isomorph, there are two possibilities. Either for all i, A (D, ci) and A (D, di) are rough seed
isomorph, then the deep isomorphism is proved, against the assumption. Otherwise, there must be
some n, A (D, cn) and A (D, dn) are not rough seed isomorph. Use the same method as above we can
prove that A (D, cn) is a sub-seed algebra of A (D, cn-1) and A (D, dn) a sub-seed algebra of A (D,
dn-1). Thus they must be rough seed isomorph. This refutes the assumption.
N
Definition 2.12: A symmetric relation SBS on the set of all processes is called strong bisimulation
if for each (P, Q)  SBS, the following proposition is true:
P
P
P , P o
P ' , there is Q’ such that Q 
o
Q'
and (P’, Q’)  SBS, where
P = W or c?a or c!a for any channel c and any token a.
Two processes P and Q are called strongly bisimular (or: strongly bisimulate each other) if
there is a strong bisimulation relation SBS such that (P, Q)  SBS.
Theorem 2.13: A pair of processes (P, Q) strongly bisimulate each other iff their corresponding
seed algebras A (D, a) and A (D, b) are deep isomorph, where a = tconf (p, ()), b = tconf (q, ()). Note
that p and q are the seed algebra element form of the processes P and Q.
Proof:
Since the proof involves lots of details, we sketch the main idea with following steps:
Step 1: We first establish a one-one correspondence between the set of all processes (generated by
the process grammar) and the set of proc elements generated by C-Process-Cons. This is
guaranteed by the following relation:
i-proc (c, u, p) ÅÆ c?u .P
o-proc (c, a, p) ÅÆ c!a .P
sum (p, q) ÅÆ P + Q
compo (p, q) ÅÆ P | Q
nil ÅÆ nil
tau (p) ÅÆ W .P
where p and q are seed algebraic forms of P and Q.
Step 2: We then establish a one-one correspondence between transition sequences of P and
elements of seed algebra with p as seed. Note that a transition can be considered both as concrete
and abstract transition. Assume process P = c!b.S + c!a.Q | c!b.R outputs b via channel c. As a
concrete transition, it may be written as
c!b
c!b
c!b.S c!a.Q | c!b.R o
S or c!b.S c!a.Q | c!b.R o
c!a.Q | R (2.3)
depending on which c!b is choosen to be performed. But as an abstract transition, we just write
c!b
P o
P'
(2.4)
where all possibilities of performing c!b are included. Clearly an abstract transition represents a
group of concrete transitions.
281
In accordance to the two kinds of transitions, there are two kinds of correspondence: the
correspondence between abstract transitions and non-hidden closures; and that between concrete
transitions and seed algebra elements whose outermost function is non-hidden. In the above
example, the abstract transition (2.4) corresponds to the non-hidden closure
toutput ([tconf (p, ())], [c], [b])
where p = sum (o-proc (c, b, s), compo (o-proc (c, a, q), o-proc (c, b, r))), whereas the concrete
transitions (2.3) correspond to seed algebra elements:
toutput (tleft (tconf (p, ())),c, b) and toutput (tright (tright (tconf (p, ()))),c, b)
Both corrspondences are obviously one-to-one. Both correspondences are needed in the proof of
the theorem.
Let’s consider the first kind of correspondence, where the transition rules no more describe
process to process transition, but transitions from process state to process state. Replace process by
process state in transition rule is here only for convenience. But it is a necessity in transition rules
of quantum process algebra.
1. Each process P considered as zero transition corresponds to the seed tconf (p, ()), which is of the
sort ptrace, where p is the algebraic form of P and () the zero trace;
2. The transitions
P1
( P, tr ) o
( P' , (tr, P1 )) ,
P1
c?u or c!a
W
( P, tr ) 
o
( P' , tr ) ,
correspond to the non-hidden closure f 1 ([tconf ( p, tr )],...) , where f1 is tinput, toutput or t-h-act,
corresponding to the action c?u, c!a or W that P may perform, and the dots …either represent a
channel c and a token a in case of tinput and toutput, or represent nothing (empty) in case of
invisible action W .
3. If the transition sequence
P1 P 2 ... P n
( P, tr ) 

o( P' , tr ' ) ,
n t1,
i, Pi  { c?u, c!a, W }
corresponds to the non-hidden closure f n ([ f n 1 (... f 1 ([tconf ( p, tr )],...)...)],...) , where each
f i is the non-hidden function corresponding to action Pi and the dots … have the same meaning
as above, then the transition sequences:
P1P 2 ... P n P n 1
( P, tr ) 

o( P' , (tr' , P n 1 )) ,
P n 1  { c?u, c!a }
P1 P 2 ... P nW
( P, tr ) 
o( P' , tr ' )
correspond to the non-hidden closure f n 1 ([ f n (... f 1 ([tconf ( p, tr )],...)...)],...) , where f n 1 is
the non-hidden function representing the action
P n 1 or W , the dots … have the same meaning as
above.
In this way, we have found a one-one correspondence between the transition sequences (traces)
282
of the process P and the non-hidden closures of the process states ptrace of the seed algebra with
tconf (p, ()) as seed. At the same time we also found a one-one correspondence (2.5) between the
non-hidden closures of the process states ptrace of two seed algebras with tconf (p, ()) respectively
tconf (q, ()) as seed if the two processes bisimulate each other. More exactly, this is the one-one
correspondence:
f n ([ f n 1 (... f 1 ([tconf ( p, tr )],...)...)],...)
ÅÆ
f n ([ f n 1 (... f 1 ([tconf ( q, tr )],...)...)],...)
(2.5)
where the corresponding dots on both sides either denote the same channel and token or are both
empty.
Step 3: Prove that if P and Q are strongly bisimular, then A’ (D, P) and A’ (D, Q) are deep isomorph.
From strong bisimilarity of P and Q their trace equivalence follows. By proposition 2.8 the
rough seed isomorphism of A’ (D, P) and A’ (D, Q) is guaranteed. Now assume
a
f
p
m
(f
p
m 1
... f
p
1
(tconf ( p, ())...)...)  A’ (D, P),
It corresponds to a concrete transition sequence:
P1 ... P k
( P, ()) 
oconc ( P' , tr )
where
P1 ,..., P k corresponds to the non-hidden subsequence f nh1 ,..., f nh k of the function
sequence f
p
1
,..., f
p
m.
By strong bisimularity there is a concrete transition sequence
P1 ... P k
(Q, ()) 
oconc (Q' , tr )
Such that P’ and Q’ are strongly bisimular.
This trasition sequence corresponds to some element
f q n ( f q n1... f q1 (tconf (q, ())...)...)  A’ (D, Q),
b
where f
q
1
,..., f q n has the same non-hidden subsequence f nh1 ,..., f nh k . Obviously (a, b) is a
corresponding pair since they are represented by
D
f nh k (...[ f nh1 (tconf ( p, ())...)]...) and E
f nh k (...[ f nh1 (tconf (q, ())...)]...)
respectively. Remember that P’ and Q’ are strongly bisimular. Apply proposition 2.8 to the seed
algebras A’ (D, a) and A’ (D, b) once again, we get the completed proof.
Step 4: Prove that if A’ (D, P) and A’ (D, Q) are deep isomorph, then P and Q are strongly bisimular.
The proof of this part is almost the same as in the last step. We just need to reverse the proof
procedure of step 3 to get the result.
N
There is a known set of theorems about bisimulation of processes in classical theory of
process algebra. At this place, we will restate some of them, but in seed algebra form, and prove
them by using techniques of seed algebra.
283
Henceforth we will use the notation A’ (D, P) # A’ (D, Q) to denote their deep isomorphism.
Proposition 2.14: # is an equivalence relation.
N
Lemma 2.15: A (D, tconf (p, tr1)) # A (D, tconf (p, tr2))
N
Theorem 2.16: Given ADT D and two processes P and Q. If there is a deep isomorphism hnh : A’ (D,
P) # A’ (D, Q), then there are also the following deep isomorphisms:
1. A’ (D, W . P) # A’ (D, W . Q);
2. A’ (D, c!a.P) # A’ (D, c!a.Q);
3. A’ (D, c?u.P) # A’ (D, c?u.Q);
4. A’ (D, P + R) # A’ (D, Q + R) for any process R;
5. A’ (D, P | R) # A’ (D, Q | R) for any process R.
N
Proposition 2.17: Given ADT D and processes P, Q, R, T. From A’ (D, P) # A’ (D, Q) and A’ (D,
R) # A’ (D, T) it follows A’ (D, P + R) # A’ (D, Q + T).
N
Proposition 2.18: Given data type D and processes P, Q, R, T. From A’ (D, P) # A’ (D, Q) and A’ (D,
R) # A’ (D, T) it follows A’ (D, P | R) # A’ (D, Q | T).
N
$ 3 ADT as semantics of Quantum Process algebra
Due to the space limit we only report the main results about ADT semantics of quantum
process algebra in this section and omit many physical interpretations. The interested reader may
consult the listed literature. Consider the following grammar of quantum process algebra:
P ::= Act. T | (P + P’) | (P | P’)
Act ::= qc?qx | qc!qx | W | uni [un, X] | mes [X]
T:: = P | rp | nil
un ::= unitary matrix
X ::= qx | qx’, X
where uni means unitary transformation, mes is measuring operator, X is a sequence of quantum
variables, rp means recursive call. This grammar is attached with following context conditions to
prevent quantum copying.
1. Parallel processes should not contain the same free quantum variable, i.h.
If qx is free in P and qx’ is free in P’ then qx z qx’ for the process P | P’;
2. Process P outputting variable qx should not contain qx in its body (P’s body is P with its output
prefix removed), i.h, qc!qx.P is only then a legal process if qx is not free in P.
It has the following (early) semantics:
Input:
qv  fv( P )
qc!qv
, Output: ( qc! qv.P, U ) 
o( P, U )
qc ? qv
(qc ? qu.P, U ) 
o( P[qv / qu ], U )
(The restriction qv  fv ( P ) for input is only needed in quantum case to avoid quantum
copying. Note that neither input nor output will change the density operator U . When input or
284
output is performed, only the names of quantum variables will be moved. Only operators like
unitary operator or measure operator can change the content of U .)
Sum:
D
( P, U ) o
( P' , U ' )
D
( P Q, U ) o( P' , U ' ), (Q P, U ) o
( P' , U ' )
D
Compo:
D
( P, U ) 
( P' , U ' )
o
D
D
( P | Q, U ) 
o( P'| Q, U ' ), (Q | P, U ) 
o
(Q | P' , U ' )
(Just as in the case of classical process algebra in section 2, the extra condition
bv (D ) , fv (Q )
Commu:
D
D
( P, U ) 
o
( P' , U ), (Q, U ) 
o
(Q' , U )
W
, Tau: (W .P, U ) 
o
( P, U )
W
( P | Q, U ) 
o( P' | Q' , U )
,X ]
(uni[un, X ].P, U ) uni
[ un
o( P, uni[un, X ]( U ))
Unitary:
Measuring:
Recur:
empty is not needed)
X]
(mes[ X ].P, U ) mes
[
o( P, mes[ X ]( U ))
D
( P, U ) o
( P' , U )
D
(rp, U ) o
( P' , U )
rp
def
P
where for any channel qc and any variable qv, qc ? qv
qc! qv , qc! qv
qc ? qv , D  {qc!qv,
qc?qv}.
Theorem 3.1: There is a hierarchical ADT (we call it Q-Process-Trans), which defines the above
quantum process algebra, such that the seed algebras of this ADT correspond to its process
behaviors, and the (equation logic) axioms describe exactly the functions of unitary and measuring
operations of quantum mechanics.
N
Definition 3.2: A quantum state 3 = <P, U > is a pair of process P and its density operator U . Let
Qconf be the set of all quantum states. A symmetric relation SQBS Ž Qconf u Qconf is called
strong quantum bisimulation if for each ( 31 , 3 2 )  SQBS, the following proposition is true:
1.
D
D
D , 31 o
31 ' , there is 3 2 ' such that 3 2 o
32 '
and ( 31 ' , 3 2 ' )  SQBS, where
or
D = W or qc!qv for any channel qc and any quantum variable qv,
D = uni [un, X] or mes [X] for any unitary operator un and any subset of quantum variables X,
where the dimension of un is 2|X|.
2.
qv
qc, qv, 31 qc
?
o 31 ' , where qv  fv(31 ) 8 fv(3 2 ) , there is 3 2 ' such that
qc ? qv
3 2 
o 3 2 ' and ( 31 ' , 3 2 ' )  SQBS.
Two quantum states 31 , 3 2 are said to be strongly bisimular if and only if there is a strong
285
bisimular relation SQBS such that ( 31 , 3 2 )  SQBS.
Theorem 3.3: A pair of quantum systems ( 31 = (P1, U 1), 3 2 = (P2, U 2)) strongly bisimulate each
other iff their corresponding seed algebras A (Q-Process-Trans, a) and A (Q-Process-Trans, b) are
~
~
~
~
deep isomorph, where a = qconf ( 31 , ()), b = qconf ( 3 2 , ()) where 31 = (p1, p1, U 1), 3 2 = (p2,
p 2, U 1) .
Note that P1 and P2 are quantum processes constructed by the quantum process grammar,
whereas p1 and p2 are their seed algebraic forms constructed by Q-Process-Trans. The Greek letters
~
U 1 and U 1 denote density matrices. That 3 i are slightly different from 3 i is purely due to a
technical reason (of the need of processing recursion).
N
By modifying the axioms we can get different models of algebraic semantics for quantum
process algebra, such as weak bisimulation and late semantics. Due to space shortage we leave it to
the full paper version.
$ 4 Conclusion
In summary, the approach presented in this paper provides the following advantages:
The first advantage is its formality as compared with operational semantics and bisimulation
semantics. Other semantics need more or less the help of natural language to explain the exact
meaning they imply. For example, the difference of early and late semantics can hardly be made
clear without some literal explanation. But in algebraic semantics, the change of axioms represents
everyting rigorously.
The second advantage comes from the multiplicity of its models. One can consider different
models as different semantics. The ADT approach provides a unified view of all these semantics.
The third advantage is its unified form of representation. Traditionally, syntax and semantics
of a process algebra are represented by different formalism. In ADT, everything is represented with
the (sort, signature, axiom) three level structure. We have even ‘programmed’ the major part of the
core of quantum mechanics in this way.
Some problems discussed in this paper have been solved in the literature with different
techniques. For tackling non-determinism see e.g. [BW81] and [LER82]. In this paper we introduce
the concept of non-deterministic objects and their isomorphism. This approach not only reflects the
idea of nondeterminism in a more straightford way, but also avoids the explicit description of sets.
For hidden techniques and observability see e.g. [W89] and [ST87]. Our purpose of
introducing hidden operation is not to mask the details or private parts of a program module to
prevent it from unwanted intrusion. Rather, our purpose is to mask the unwanted details in the
isomorphic mapping of non-deterministic objects. The introduction of hidden symbols is actually a
necessity in establishing the correspondence between seed algebra elements and process states. The
granule of masking is also a difference between the various approaches: is it on the signature level,
the operation level, the sort level or the carrier set elements level? Our approach adopts an
operation level for masking.
It is worthwhile to mention that we have introduced the concept of subsort to describe the
hierarchy of linear operators on a Hilbert space for our quantum process algebra. It is not
286
paraphrased above to save the space. A subsort satisfies all axioms of some already defined sorts,
and at the same type satisfies some additional axioms. Subsort is a more delicate concept than
subtype and is close to the approach of embedding [ABK+02]. It makes our ADT not only a
many-sorted algebra but also an order-sorted algebra[ABK+02].
References
[ABK+02] E.Astesiano, M.Bidoit, H.Kirchner, B.Krieg-Brueckner, P.D.Mosses, D.Sanella,
A.Tarlecki. CASL: The Common Algebraic Specification Language. TCS. 286(2): 153-196, 2002.
[AG05] T. Altenkirch and R. Grattage. A Functional Quantum Programming Language. In
Proceedings of the 20th Annual IEEE Symposium on Logic in Computer Science(LICS). IEEE
Computer Society. 2005.
[BCS03] S.Bettelli, T. Calarco and L. Serafini. Toward an Architecture for Quantum Programming.
The European Physical Journal D. 25: 181-200, 2003.
[BPW82] M.Broy, P.Pepper, M.Wirsing, On the Algebraic Definition Programming Languages,
TUM-I8204, 1982.
[BW80a] M.Broy, M.Wirsing, Programming Languages as Abstract Data Types, In M.Dauchet
(Ed.), 5me Coll. Sur les Arbres en Algebre et Programmation, pp.160-177, 1980.
[BW80b] M.Broy, M.Wirsing, Partial Recursive Functions and Abstract Data Types, EATCS Bull.
11, pp.34-41, 1980.
[BW81] M.Broy, M.Wirsing, On the Algebraic Specification of nondeterministic programming
languages, E. Astesiano, C.Boehm (eds.), 6th CAAP, LNCS 112, pp.162-179, 1981.
[BW82] M.Broy, M.Wirsing, Partial Abstract Data Types, Acta Informatica, 18 (1): 47-64, 1982.
[C80] J.C.Cleaveland, Programming Languages Considered as Abstract Data Types, Proc. of the
ACM 1980 Annual Conference, pp. 236-245, 1980
[F07] Y. Feng, R. Y. Duan, Z. F. Ji and M. S. Ying. Probabilstic Bisimulations for Quantum
Processes. Information and Computation. 205: 1608-1639, 2007.
[Gay06] S. Gay. Quantum Programming Languages: Survey and Bibliography. Mathematical
Structures in Computer Science. 16(4): 581-600, 2006.
[GH78] J. Guttag and J Horning. The Algebraic Specification of Abstract Data Types. Acta
Information. 10: 27-52, 1978.
[GN05] S. Gay and R. Nagarajan. Communicating Quantum Processes. In Proceedings of the 32nd
ACM Symposium on Principles of Programming Languages, Long Beach, California, USA, ACM
Press, 145-157, 2005.
[GT75] J.Goguen, J. Thatcher, E. Wagner, J. Wright. Abstract Data Types as Initial Algebras and
the Correctness of Data Representation. In Proceedings of Conference on Computer Graphics,
Pattern Recognition and Data Structures, pp. 89-93, 1975.
[GT78] J. Goguen, J. Thatcher and E. Wagner. An Initial Algebraic Approach to the Specification,
Correctness and Implementation of Abstract Data Types. Currend Trends in Progr. Meth. 4: 80-149,
1978.
[Kni96] E. Knill. Conventions for Quantum Pseudocode. Technical Report LAUR-96-2724, Los
Alamos National Laboratory, 1996.
[Lal06] M. Lalire. Relations among Quantum Processes: Bisimulation and Congruence.
Mathematical Structures in Computer Science. 16: 407-428, 2006.
[Ler82] P.Lerscanne, Mod è ls non-deterministe de Types Abstraits, RAIRO Informatique
287
Theorique 16, pp.225-244, 1982.
[LJ04] M. Lalire and P. Jorrand. A Process Algebraic Approach to Concurrent and Distributed
Quantum Computing: Operational Semantics. In P. Selinger(Ed.) Proceedings of the 2nd
International Workshop on Quantum Programming Languages, pages 109-126.
[JL04] P. Jorrand and M. Lalire, Toward a Quantum Process Algebra, In Proceedings of the 1st
ACM Conference on Computing Frontiers. ACM Press, 2004.
[Jor05] P. Jorrand. A Programmer’s Survey of the Quantum Computing Paradigm. In Proceedings
of World Academy of Science, Engineering and Technology. Vol. 8, 2005.
[Lu87] R. Lu, Algebraic Semantics of Programs with Dynamic Structure, Science in China, A, 30
(1), pp. 32-44, 1987.
[Mln06] H. Mlnaík, Operational Semantics of Quantum Programming Language LanQ. Masaryk,
University Technical Reports. FIMU-RS-2006-10.
[Öme00] B. Ömer. Quantum programming in QCL. Master thesis, Institute of Information Systems,
Technical University of Vienna. 2000.
[Pap04] N.K.Papanikolaou, qSpec: A programming language for quantum communication systems design, in:
Proceedings of PREP2004 Postgraduate Research Conference in Electronics, Photonics, Communications &
Networks, and Computing Science (2004)
[Sel04a] P. Selinger. A Brief Survey of Quantum Programming Languages. Proceedings of the 7th
International Symposium on Functional and Logic Programming, Nara, Japan. Springer LNCS
2998, pp. 1-6, 2004.
[Sel04b] P. Selinger. Towards a Quantum Programming Language. Mathematical Structures in
Computer Science. 14(4): 527-586, 2004.
[ST87] D.T.Sanella, A.Tarlecki, On Observational Equivalence and Algebraic Specifications, J.
Comp. and Sys. Sceinces, 34, pp.150-178, 1987.
[SV05] P.Selinger and B.Valiron. A Lambda Calculus for Quantum Computation with Classical
Control. In Proceeding of the 7th International Conference on Typed Lambda Calculi and
Applications(TLCA), volume 3467 of Lecture Notes in Computer Science. Springer, 2005.
[SW83] D.T.Sanella, M.Wirsing. A Kernel Language for Algebraic Specification and
Implementation. LNCS 158, pp. 413-427. 1983
[Ton04] A. van Tonder, A Lambda Calculus for Quantum Computation, SIAM Journal of
Computing. 33(5): 1109-1135, 2004.
[W89] M. Wirsing, Algebraic Specification, MIP-8914, Uni.Passau, 1989.
[Xu08] J. Xu, F. Song, S. Qian, J. Dai and Y. Zhang. The Quantum Programming Language
NDQJava. Journel of Software. 19(1): 1-8, 2008. (In Chinses with English abstract)
[Y01] M. S. Ying, Topology in Process Calculus, Springer Verlag, 2001.
[Y08] M. S. Ying, Y. Feng, R. Y. Duan and Z. F. Ji. An Algebra of Quantum Processes. ACM
Transactions on Computational Logic. Vol. V. No. N. 1-12, 2008.
[Zil74] S. Zilles. Algebraic Specification of data types. Computation Structures Group Memo 119,
Lab for Computer Science, MIT, 1974.
[Zul01] P. Zuliani. Quantum Programming. PhD Thesis, University of Oxford, 2001.
Appendix
Axiom:
#non-hidden operations#
Axiom Part of C-Process-Trans
288
tinput (tconf (i-proc (c, u, p), tr), c, t) = tconf (subst (p, u, t), run (tr, it (c, t)))
toutput (tconf (o-proc (c, u, p) , tr), c, u) = tconf (p, run (tr, ot (c, t)))
tinput (tconf (compol (x, y), tr), c, t) = tconf (input (compol (x, y), c, t), run (tr, it (c, t)))
toutput (tconf (compol (x, y), tr), c, t) = tconf (output (compol (x, y), c, t), run (tr, ot (c, t)))
tinput (tconf (compor (x, y), tr), c, t) = tconf (input (compor (x, y), c, t), run (tr, it (c, t)))
toutput (tconf (compor (x, y), tr), c, t) = tconf (output (compor (x, y), c, t), run (tr, ot (c, t)))
t-h-act (tconf (p, tr)) = tconf (h-act (p), tr)
#hidden Operations#
input (compol (i-proc (c, u, p), z), c, t) = compol (subst (p, u, t), z)
output (compol (o-proc (c, u, p), z), c, u) = compol (p, z)
input (compor (z, i-proc (c, u, p)), c, t) = compor (z, subst (p, u, t))
output (compor (z, o-proc (c, u, p)), c, u) = compor (z, p)
h-act (tau (p)) = p
h-act (compo (i-proc (c, u, p), o-proc (c, v, q))) = compo (subst (p, u, v), q)
h-act (compol (x, y)) = compol (h-act (x), y)
h-act (compor (x, y)) = compor (x, h-act (y))
tleft (tconf (sum (x, y), tr)) = tconf (x, tr)
tright (tconf (sum (x, y), tr)) = tconf (y, tr)
tleft (tconf (compo (x, y), tr)) = tconf (compol (x, y), tr)
tright (tconf (compo (x, y), tr)) = tconf (compor (x, y), tr)
tleft (tconf (compol (x, y), tr)) = tconf (compol (left (x), y), tr)
tright (tconf (compol (x, y), tr)) = tconf (compol (right (x), y), tr)
tleft (tconf (compor (x, y), tr)) = tconf (compor (x, left (y)), tr)
tright (tconf (compor (x, y), tr)) = tconf (compor (x, right (y)), tr)
tback (tconf (x, tr)) = tconf (back (x), tr)
left (compo (x, y)) = compol (x, y)
right (compo (x, y)) = compor (x, y)
left (compol (x, y)) = compol (left (x), y)
right (compol (x, y)) = compol (right (x), y)
left (compor (x, y)) = compor (x, left (y))
right (compor (x, y)) = compor (x, right (y))
back (i-proc (c, u, p)) = i-proc (c, u, p)
back (o-proc (c, u, p)) = o-proc (c, u, p)
back (tau (x)) = tau (x)
back (nil) = nil
back (sum (x, y)) = sum (x, y)
back (compo (x, y)) = compo (x, y)
back (compol (x, y)) = compo (back (x), back (y))
back (compor (x, y)) = compo (back (x), back (y))
subst (i-proc (c, u, p), u, t) = i-proc (c, u, p)
subst (o-proc (c, u, p), u, t) = o-proc (c, t, subst (p, u, t))
subst (tau (x), u, t) = tau (subst (x, u, t))
subst (nil, u, t) = nil
subst (sum (x, y), u, t) = sum (subst (x, u, t), subst (y, u, t))
289
subst (compo (x, y), u, t) = compo (subst (x, u, t), subst (y, u, t))
sum (x, y) = sum (y, x)
sum (p, sum (x, y)) = sum (sum (p, x), y)
compo (x , y) = compo (y, x)
compo (p, compo (x, y)) = compo (compo (p, x), y)
#undefined operations ommited#
}
End of Type C-Process-Trans
290
Ruqian Lu received his Diplom in Mathematics from
Jena University in 1959 and is now a full professor
of Academy of Mathematics and Systems Science and
also a fellow of Chinese Academy of Sciences. He met
Bernd Krieg-Brückner for the first time in 1978 at a
workshop of IFIP 2.4 and has been keeping personal
relationship with Bernd for more than 30 years. His research focus is artificial intelligence, knowledge based
software engineering and formal semantics of quantum
programming.
Lixing Li received hisbachelor’s degree fromCentral
South Universityin 2006 majoring in computer science
and is now a PhD candidate under supervision of Professor Ruqian Lu. His research focus isformal semantics, quantum computation and quantum information.
Yun Shang received her PhD from Shannxi Normal
University in 2005 under supervision of Prof. Yongming Li. From 2005-2007, she worked as a Post-doc
in the group of Prof. Ruqian Lu. Since 2007, she has
been a lecturer at Academy of Mathematics and Systems Science. Her research focus is quantum information and quantum logic.
Xiaoyu Li received his PhD from the Institute of
Computing Technology, Chinese Academy of Sciences
in 2004 under supervision of Ruqian Lu. Then he went
to Zhengzhou University in 2005 and worked in School
of Information Engineering as lecturer. He got a position of associate professor in 2006. His research focus
is quantum information, especially quantum entanglement and non-locality, and quantum cryptography.
AG.com – Bremer Informatik-Arbeitsgemeinschaften in der
gymnasialen Sekundarstufe I
Prof. Dr. Holger Schlingloff
Humboldt-Universität zu Berlin & Fraunhofer FIRST
Holger Schlingloff war von 1996 bis 2002 Geschäftsführer des Bremer Instituts für Sichere
Systeme (BISS) im Technologie-Zentrum Informatik (TZI) der Universität Bremen, dessen
Direktor Bernd Krieg-Brückner war. Als Geschäftsführer akquirierte, leitete und bearbeitete er
viele Projekte für und zusammen mit dem Jubilar. Über eines dieser Projekte wird
nachfolgend berichtet.
Im Jahre 2000 beauftragte der Senator für Bildung und Wissenschaft des Landes Bremen,
Willi Lemke, den Dekan des Fachbereichs 3 (Mathematik / Informatik) der Universität
Bremen, Prof. Bernd Krieg-Brückner, mit der Erarbeitung eines Konzepts für BegabtenFörderkurse Informatik in den 7. und 8. Klassen an Bremer Gymnasien. In Folge dieses
Auftrags erstellte Prof. Krieg-Brückner in Zusammenarbeit mit dem Autor des vorliegenden
Berichtes eine Projektskizze und Angebot für solche Kurse, die als „AG7.com“ im Schuljahr
2001/2002 an 10 Bremer Schulen erstmals durchgeführt wurden und in leicht abgewandelter
Form als „AG.com“ noch heute (Schuljahr 2008/2009) laufen [1]. Dieser Artikel beschreibt die
pädagogischen Ziele, Inhalte und Erfahrungen, die sich bei der Durchführung ergeben
haben, und reflektiert das Projekt aus heutiger Sicht.
1.
Hintergründe und Voraussetzungen
Zur Zeit der Konzeption des Projektes gab es starke gesellschaftliche und politische
Motivationen, die Informatik an Bremer Schulen auszubauen. Wirtschaftlich befand sich die
Informatikindustrie zu dieser Zeit auf einem relativen Höchststand. In Bayern war gerade
beschlossen worden, Informatik als Pflichtfach an allen weiterführenden Schulen ab 2002
einzuführen. In Bremen war mit dem Projekt „Schulen ans Netz“ eine leistungsfähige
Infrastruktur geschaffen worden, die einer Nutzung bedurfte. Der neue Senator (seit Juli 99
im Amt) beabsichtigte eine Stärkung und Profilbildung der Bremer Schulen, gerade in
technischen Fächern. Andererseits standen dem Projekt einige Hindernisse im Wege. Schon
bei Bekanntwerden der Pläne gab es erhebliche Kompetenzstreitigkeiten, sowohl zwischen
dem Landesinstitut Schule und der Universität, als auch innerhalb des FB3 selbst. Über den
möglichen und sinnvollen Umfang gab es finanzielle Unklarheiten: während zunächst geplant
war, 50 Kurse mit je 10 Schülerinnen und Schülern und einem Finanzvolumen von 1,3 MDM
einzurichten, wurden in der ersten Phase lediglich 7 Informatik-Kurse mit je 15 Lernenden für
knapp 100 TDM eingerichtet. Es gab massive Interessenkonflikte mit anderen Protagonisten,
die die ursprüngliche Idee an weitere Ziele koppeln wollten (Mädchenförderung, Förderung
benachteiligter sozialer Gruppen, integrativer Unterricht, Stärkung der mathematischnaturwissenschaftlichen und / oder sprachlichen Kompetenzen, …). Dies führte letztlich
dazu, dass neben den 7 Informatik-AGs 3 Mathematik-, 3 Physik-, 3 Technomathematik- und
2 „Sprachen-im-Web“-AGs aus den vorhandenen Mitteln finanziert wurden und eine AG nur
für Mädchen offen war. Der Gedanke der Begabtenförderung wurde aus politischen
Überlegungen fallen gelassen; der Aufruf zur Teilnahme richtete sich an alle SekI-Schulen
und -Schüler. Er fand ein enormes Echo bei den Bremer Schulen, Lehrkräften und Kindern:
über 20 Schulen bewarben sich für die vorgesehenen 5 Standorte, trotz Mehrbelastung
waren viele Lehrkräfte freiwillig bereit, am Projekt teilzunehmen, und seitens der
Schülerinnen und Schüler gab es bis zu 10 mal so viele Anmeldungen, wie Plätze vorhanden
waren.
Das Projekt wurde mit einer Pressekonferenz am 12.2.2001 gestartet.
292
2.
Pädagogische Zielsetzung und Konzeption
Generell war und ist der Begriff „Informatik“ in der Sek 1 vielfach diskutiert und (fehl-)
interpretiert worden. Zur Abgrenzung ist zunächst zu unterscheiden zwischen der Nutzung
von Computern im Unterricht, der informationstechnischen Grundbildung (ITG) und der
Informatik als Fach.
x
Computer im Unterricht
Hier geht es darum, den Computer als Medium zur Unterstützung von Lernvorgängen
und zur Kommunikation im Unterricht zu verwenden. Beispiele sind
Sprachlernprogramme, Internetrecherchen zu bestimmten Themen oder
elektronische Abgabe von Aufgaben. Die Meinungen über die Sinnhaftigkeit des
Einsatzes des neuen Mediums sind geteilt: während viele Pädagogen der Auffassung
sind, in der nächsten Generation werde der PC ein selbstverständliches Arbeitsmittel
zu Hause und im Unterricht sein, gehen andere inzwischen schon wieder dazu über,
Hausaufgaben am Computer zu verbieten.
x
informationstechnische Grundbildung ITG
Hier geht es darum, den Schülerinnen und Schülern im Sinne eines
„Computerführerscheins“ Grundfertigkeiten in der Nutzung von Computern
beizubringen. Ziel ist die Vermittlung von vermeintlich oder auch tatsächlich
berufsrelevantem Wissen (vergleichbar dem Stenographieunterricht der 1950-er
Jahre), da inzwischen über 90% aller Berufe unmittelbar oder mittelbar mit
Computern zu tun haben. In der Erweiterung wird die informationstechnische
Grundbildung zum Programmierunterricht in einer bestimmten Sprache, z.B. Delphi
oder Java, wobei die Halbwertszeit des Gelernten oft kaum eine Rolle spielt.
x
Informatik als Fach
Hier geht es darum, Wirkprinzipien moderner Informationstechnologie zu erklären
und Systeme aus der Erfahrungswelt der Lernenden zu modellieren und zu
simulieren, um ein Verständnis für Funktionszusammenhänge als Grundlage für ein
allgemeines Weltbild zu wecken.
In der AG.com ging es nur um den dritten Aspekt, d.h. Informatik auf dem Niveau 13-jähriger
Schülerinnen und Schüler zu behandeln. Thema sollten die Möglichkeiten und Grenzen des
verantwortungsbewussten Einsatzes von Computern in ihrem Erfahrungs- und
Einflussbereich sein. Der Computer sollte als Werkzeug zur Informationsverarbeitung
verständlich gemacht werden, d.h. nicht Gegenstand, sondern Mittel zum Zweck sein.
Die Sek 1 eignet sich aus verschiedenen Gründen besonders gut für die Einführung von
Informatik als Fach. A priori ist bei den Schülerinnen und Schülern eine hohe Motivation
durch die besondere Attraktivität des Mediums Computer gegeben. Als wichtigstes Ergebnis
ist der Erwerb von Handlungskompetenz anzusehen, der die Lernenden zu
verantwortungsbewusstem Handeln verhilft. Durch den strukturierten Problemlöseprozess
ergibt sich eine allgemein anwendbare Methodenkompetenz für technischnaturwissenschaftliche Fächer. Die Erstellung „funktionierender“ Lösungen trägt zum Aufbau
von Selbstkompetenz und Selbstbewusstsein bei. Durch die Notwendigkeit von
Gruppenarbeit erwerben die Teilnehmenden erhebliche Sozialkompetenzen. Die Inhalte
vermitteln grundlegende Sachkompetenzen für den notwendigen Aufbau- Unterricht in der
Sekundarstufe II und in berufsbildenden Schulen.
Organisatorisch wurden die Kurse so eingeteilt, dass für ca. 15
Teilnehmende je eine Lehrkraft sowie eine studentische
„Lehrassistenzkraft“ zur Verfügung standen. Die AGs wurden zentral
betreut und verteilt in den Computerlabors der jeweiligen Schulen
durchgeführt. Wöchentlich gab es eine gemeinsame Vor- und
Nachbesprechung an der Uni, bei der eine direkte und indirekte
Rückkopplung zu den Konzepten und ein Austausch von Ideen statt
fanden. Von der zentralen Kursleitung wurden interaktive und
293
klassische Unterrichtsmaterialien bereitgestellt: Webseiten, Aufgabenblätter, Arbeitsbögen
und Anschauungsmaterialien. Besonders stolz ist der Autor auf das für die AG7.com
geschaffene Maskottchen „Saschi“, eine jugendliche Schildkröte, die als Identifikationsfigur
und Wegweiser im Stoffangebot fungiert. Als Lernsoftware wurde eine speziell angepasste
Version von „MicroWorlds Pro“ des kanadischen Herstellers LCSI eingesetzt; diese auf dem
klassischen LOGO basierende multimediale Programmierumgebung erlaubt ein weites
Spektrum von Anwendungen, von der Erstellung einfacher funktionaler und paralleler
Programme, Bild- und Textverarbeitung, bis hin zum komplexen Internet Design.
3.
Unterrichtseinheiten, Themen und Lerninhalte
Während für die Lernziele der AG7.com
Vorgaben in Gestalt von GIEmpfehlungen für Informatik in der SekI
[2] sowie Entwürfe für Lehrpläne
einzelner Länder vorlagen, musste eine
Strukturierung dieser Lernziele in
Unterrichtseinheiten (UE) neu erstellt
werden. Jede UE behandelte ein
bestimmtes Thema. Für die einzelnen
Themen wurden Lerninhalte erstellt, und
Fertigkeiten definiert, die die
Schülerinnen und Schüler durch die
Bearbeitung der Einheit erwerben
sollten. Zu jeder UE gab es
umfangreiche Materialien, sowohl für die
Lehrenden als auch für die Lernenden.
Bei der Auswahl der Lerninhalte wurden
verschiedene Kriterien berücksichtigt,
zuvorderst das der Übertragbarkeit:
während Inhalte der Klasse (a) für ein
konkretes IT-System (z.B. MicroWorlds
Pro) gelten, sind Inhalte der Klasse (b)
für eine Generation von IT-Systemen
anwendbar (z.B. Logo-Programmiersysteme). Inhalte der Klasse (c) sind für alle IT-Systeme
relevant (z.B. Programmiermethodik), während Inhalte der Klasse (d) auch außerhalb der
Informatik angewendet werden können (z.B. Organisation von Abläufen). Bei der
Präsentation des Stoffes wurde notwendigerweise von Inhalten der Klasse (a) ausgegangen;
es wurde jedoch Wert darauf gelegt, die Abstraktion zu den jeweils höheren Ebenen adäquat
darzustellen.
Ausgehend von den grundlegenden Aspekten der Informationsverarbeitung (Modellierung
bzw. Abstraktion eines Weltausschnittes, Verarbeitung bzw. Übermittlung von Information,
und Interpretation durch bzw. Interaktion mit Benutzern) wurden die zu vermittelnden Inhalte
jeder UE unterteilt in
x das Repräsentieren von Information und Wissen durch die Kinder,
x die Behandlung der Funktionsweise und Konstruieren von IT-Systemen, und
x die kritische Auswertung und Auseinandersetzung mit den Resultaten
Die Umsetzungsform war dabei immer projektbasiert, d.h. nach einer Problembegegnung
und -einordnung durch die Lehrkraft erfolgte eine individuelle informelle Beschreibung der
Aufgabe durch jeden einzelnen Lernenden, dann wurde eine gemeinschaftliche Modellierung
und Abstraktion durch Kleingruppen vorgenommen. Erst daran schloss sich die
Implementierung bzw. Realisierung am Computer an, in jedem Fall gefolgt durch eine
Bewertung und Auswertung durch die gesamte Gruppe, mit einem schulübergreifenden
Feedback (Wettbewerb) über die besten Lösungen.
294
Exemplarisch sei diese Vorgehensweise nachfolgend an zwei UE demonstriert.
UE3: Geschichten die das Leben schrieb...
Thema: Textverarbeitung I
Lernziele: Die Schüler sollen
die heutige Textverarbeitung
als Ergebnis einer
kulturgeschichtlichen
Entwicklung begreifen. Sie
sollen zwischen Textgehalt und
Textgestalt differenzieren und
die inhaltliche Erstellung eines
Textes vom Prozess des
Layouts separieren können.
Lerninhalte: Textverarbeitung
als häufigste Anwendung
heutiger Arbeitsplatzrechner;
kulturgeschichtliche
Entwicklung der Schrift aus
Bildern und Symbolen; Vorteil
kleiner Alphabete;
Vervielfältigung durch
manuelles (sequentielles)
Kopieren und (seitenweises)
Drucken; Analogie zwischen
der Aufteilung eines Wortes in
einzelne Buchstaben und der
Aufteilung einer Druckseite in
einzelne Lettern; nichtdruckende Lettern und
Trennzeichen; Zeichenattribute
(Font, Farbe, usw.);
Seitenlayout und Design am Beispiel einer Einladungskarte
Kompetenzen:
x Handlungskompetenz: Computer als selbstverständliches Schreibwerkzeug
x Methodenkompetenz: Inhalt versus Layout
x Selbstkompetenz: Schrift als Kulturgut
x soziale Kompetenz: Kopierproblematik
x Sachkompetenz: Erzeugen von Dokumenten
Fertigkeiten: MicroWorldsPro als Textverarbeitungsprogramm (Erzeugen von Textboxen,
Kombination von Text und Grafik, Rechtschreibprüfung).
Unterstützt wurden die Inhalte durch interaktive Webseiten und Arbeitsbögen für die Schüler,
Handreichungen für die Lehrer und weitere Materialien. Das obige Beispiel zeigt einen
Auszug aus den Internetseiten.
295
Die Handreichungen für die Lehrer umfassten neben Ontologie und pädagogischen
Hinweisen auch konkrete Tipps zum Umgang mit der Software (Inhalte Klasse (a)).
AG7.com
2.
2.1
Geschichten die das Leben schrieb...
Entwicklung der Textverarbeitung
Inhalt und didaktische Hinweise
Begriffe
Vor der Logo-Textverarbeitung soll die geschichtliche Entwicklung der Schrift
Textverarbeitung
und der Druckkunst behandelt werden Schriftsatz (siehe Arbeitsbögen und
WWW-Seiten). Insbesondere geht es um den Wandel des Arbeitslebens durch
Computereinsatz (Sachbearbeiter, Autoren, Reporter, usw.)
Die Schülerin/der Schüler soll eine Einladung zur Geburtstagsfeier schreiben.
Er/sie soll lernen, mit einfachen Textbearbeitungsmitteln zu hantieren und Bild
und Text zu kombinieren. Er/sie soll Vor- und Nachteile traditioneller Techniken
(Buntstifte, Schere, Papier) mit der elektronischen Vorgehensweise vergleichen.
Er soll weiterhin den Unterschied zwischen Textgehalt (Text als Folge von
Zeichen) und Textgestalt (Attribute jedes Zeichens: Schriftart, Schriftstil,
Schriftgröße, Schriftfarbe) begreifen. Insbesondere soll klar gemacht werden,
was beim Layout maschinell gemacht wird (Zeilenumbruch, Rechtschreibprüfung) und was nie automatisch gemacht werden kann (Textinhalt, Stil)
Textgehalt versus -gestalt
Starten des Programms und Einstellen der Projektgröße:
Datei – Projektgröße (je nach intendiertem Wiedergabemedium: Webplayer,
Postkarte, Din A4-Papier usw. Einstellen anderer Projektgrößen mit der
Anweisung
Projektgröße [1414 2000] für A4 Hochkant max.)
Import eines (Hintergrund-)bildes; Datei – Importieren – Bild.
Bilder befinden sich im englischsprachigen MW-Logo unter “Pictures”.
Projekteinstellungen
Alternativ: SchülerInnen erstellen mit den Grafikwerkzeugen eigenes
Hintergrundbild. Aus der Symbolleiste das Textwerkzeug “A” auswählen,
in “Seite” klicken, möglichst groß aufziehen und “Einladungskarte”
(oder Ähnliches) schreiben.
Schriftart und Farbe verändern: mit der Maus markieren, dann
Text – Schriftart – Schriftfarbe
Position und Größenänderung: Mit der Maus Textfeld umreissen
(Aktivierungspunkte).
Diskussion logische Zeilen (Absätze) vs. dargestellte Zeilen
(automatischer Zeilenumbruch).
rechter Mausklick ergibt Kontextmenü der Textbox, (Mac: Ctrl-Klick):
transparent / Namen / Rechtschreibung usw.
Dies machen wir nun noch ein paar Mal, in einer anderen Schriftart und Farbe.
Zum Ausdrucken: Menü “Seite einrichten”. Projekt speichern!
Bild Importieren
Textwerkzeug
Markieren
Nichtdruckbare Zeichen
296
Als Idee für eine Schülerarbeit wurde eine interaktive Geschichte angegeben (aus einer
früheren AG, Design Laura Schlingloff):
Zur Operationalisierung wurden die Schülerinnen und Schüler aufgefordert, eine interaktive
Einladungskarte zu erstellen. Ein Beispiel war auf den Internetseiten angegeben.
297
Nachfolgend sind einige Schülerarbeiten zu dieser UE wiedergegeben. Bemerkenswert war
dabei, dass die Aufgabe bei den Lehrerinnen und Lehrern nicht so gut ankam, weil sie ihnen
nicht „attraktiv“ genug erschien. Einige Pädagogen fragten, warum man „nicht gleich
Microsoft Word“ verwenden könnte, das sei ja viel praxisnäher, oder warum man nicht gleich
verlinkte Hypertext-Strukturen mit HTML einführte. Andere änderten die Aufgabenstellung
dahingehend ab, dass nicht eine normale Geburtstagseinladung erstellt werden sollte,
sondern eine „interaktive Web-Visitenkarte“, dies sei ja viel fortschrittlicher. Die Schülerinnen
und Schüler hingegen nahmen die (ursprüngliche) Aufgabenstellung (Erstellung einer
Einladungskarte für eine Party) begeistert auf, weil sie aus ihrer Erfahrungs- und Erlebniswelt
entstammte und sich auf ein Problem bezog, welches jeder selbst schon einmal mit anderen
Mitteln gelöst hatte.
298
Das zweite Beispiel betrifft die Einführung algorithmischer Konzepte (Begriff eines
„Programms“, Ablaufsteuerung, Strukturierung durch Makros).
Zur Vorbereitung war in einer vorhergehenden UE die Analogie zwischen einer Melodie (als
Folge von Tönen) und einem Programm (als Folge von
Einzelschritten) hergestellt. Diese Analogie wird von
MicroWorlds Pro recht gut unterstützt, da es einen
„Musikeditor“ enthält, mit dem man auf einem „virtuellen
Klavier“ Tonfolgen interaktiv eingeben kann. Man kann aber
auch die Befehle zum Abspielen der Noten als Befehle
eingeben. Daran lässt sich sehr gut das Erfordernis / die
Möglichkeit einer eindeutigen Notation und Eingabesyntax
vermitteln. Darüber hinaus dürfte diese Analogie einigen Jugendlichen durch das
„Programmieren“ von Klingeltönen für Mobiltelefone
vertraut sein.
Als umfangreicheres Beispiel für eine Folge von Schritten
wurde das internationale Kinderlied „Bruder Jakob“
vorgestellt, bei dem sich die Notwendigkeit einer
Strukturierung durch Makros leicht erkennen lässt.
299
Weitere Analogien für Computerprogramme sind Kinoprogramme als Folge der gezeigten
Filme, und politische Wahlprogramme als Liste der intendierten Gesetzentwürfe. Mit diesen
Vorbereitungen konnten Programme als Ablaufbeschreibungen allgemeinerer Natur
eingeführt werden.
UE6 Schritt für Schritt
Thema: Programme und Algorithmen
Lernziele: Die Schüler sollen eine algorithmische (iterative) Vorgehensweise zur Lösung von
Problemen kennenlernen, bei der eine Gesamtaufgabe in ein Folge von sequentiell zu
erledigenden Teilaufgaben zerlegt wird.
Sie sollen ein Programm als eine solche Folge von wohldefinierten Enzelanweisungen
begreifen und die Notwendigkeit einer eindeutigen sprachlichen Grundlage zur
Kommunikation mit Computern erkennen.
Lerninhalte: Begriff Programm, Algorithmus, Programmiersprache; Notwendigkeit der
Eindeutigkeit der Einzelschritte; Ausführung von Anweisungen als Abarbeitung eines
Programms (Beispiel aus dem Alltag); Definition von eingeschränkten Sprachmitteln;
Grundbefehle und Befehle mit Eingaben; Syntax von Logo; rechtwinklige Bewegungen der
Schildkröte anhand einfacher Spiele
Kompetenzen:
x Handlungskompetenz: Beschreibung von Handlungsabläufen
x Methodenkompetenz: Fähigkeit zur Definition elementarer Schritte eines Verfahrens
x Selbstkompetenz: eigene Planung von Aktionsfolgen für Arbeiten
x soziale Kompetenz: Kommunikation mit Menschen und Maschinen
x Sachkompetenz: Programmablauf als Ausführung einer Anweisungsfolge
Fertigkeiten: Beschreibung von Arbeitsabläufen als Folge wohldefinierter Einzelanweisungen
in einer wohldefinierten Sprache.
300
Zur Operationalisierung sollten die Schülerinnen
und Schüler an Hand eines Stadtplanes den Weg
von der Schule zum Bahnhof beschreiben, und
aus dieser Beschreibung einen Basissatz von
elementaren Anweisungen („100 Meter
geradeaus“, „rechts abbiegen“ usw.) extrahieren.
Mit diesen Basisanweisungen sollten dann
verschiedene Bewegungsspiele gestaltet werden
(mit verbundenen Augen um Hindernisse laufen,
den Weg zum Direktorat beschreiben usw.). Als
nächste Stufe wurde ein Spiel vorgegeben, bei
dem eine Schildkröte per Knopfdruck
Basisbewegungen durchführt („einen Schritt
vorwärts“, „rechts um“ usw.). Aufgabe ist es, die
Schildkröte aus einem einfachen Labyrinth
heraus zu führen. Die Folge der Knopfdrücke
lässt sich dabei, ähnlich wie die Folge der Tastendrücke auf dem virtuellen Klavier der
vorhergehenden UE aufzeichnen und insgesamt abspielen. Auf diese Weise entsteht
interaktiv ein erstes „Programm“ als Folge von Einzelschritten. In der nächsten Stufe ging es
dann darum, die Folge der Aktionen im Voraus zu planen, d.h., eine Folge von Anweisungen
zu finden, die die Schildkröte aus dem Labyrinth bringen, ohne dass diese während der
„Programmierung“ durchgeführt werden. Da diese Aufgabe signifikant schwieriger ist, wird
den Lernenden die Notwendigkeit von Strukturierungen und Makros erneut deutlich. In der
letzten Stufe dieser UE sollten die Schülerinnen und Schüler dann mit den aus vorherigen
UE bekannten Zeichenwerkzeugen eigene Labyrinthe erstellen. Die Schildkröte lässt sich
dabei so einstellen, dass sie sich kontinuierlich vorwärts bewegt; mit den Knöpfen „rechts“
und „links“ wird das Ergebnis dann zum interaktiven Action-Spiel, welches an kommerzielle
Produkte (Autorennen o.ä.) erinnert. Nachfolgend sind zwei solcher von Schülern
programmierten Spiele abgebildet.
301
4.
Erfahrungen, Reflektion und Ausblick
Die Erfahrungen mit dem Projekt lassen sich in verschiedene Ebenen einteilen. Auf Ebene
der Schülerinnen und Schüler lässt sich konstatieren, dass die Lerninhalte insgesamt sehr
gut und altersgerecht vermittelbar waren. Die Kinder waren zum größten Teil mit Feuereifer
bei der Sache, waren sehr engagiert und nahmen den Stoff wissbegierig auf. Sie waren stark
daran interessiert, den Stoff zu Hause weiter zu bearbeiten; insbesondere der interne
Wettbewerb und der Wettbewerb zwischen den Schulen erwies sich als
motivationsförderlich. Auch die Nachhaltigkeit war gegeben: Die meisten Teilnehmer gaben
an, Informatik weiter machen zu wollen. Jahre später traf der Autor auf einer RoboCupMeisterschaft einen der ehemaligen Teilnehmer wieder, der in den höchsten Tönen von dem
damaligen Kurs schwärmte. In Bezug auf die Bearbeitung der Aufgaben ließen sich auch
geschlechtsspezifische Unterschiede feststellen: während die Jungen eher mit den
Werkzeugen und Methoden experimentierten und die Aufgaben oft nicht fertig stellten, waren
die Mädchen eher an einer „korrekten“ Lösung interessiert und betrachteten Werkzeuge und
Methoden als Mittel zum Zweck. Die Arbeit in der nur für Mädchen reservierten AG war
insgesamt entspannter, in allen anderen AGs bildeten sich geschlechtsspezifische Gruppen,
die kaum interagierten.
Eine Haupt-Erfahrung ist, dass es erhebliche Schwierigkeiten bereitet, zusätzliche Inhalte
und Aufwände in den Schulalltag einzubringen. Das verfügbare Zeitbudget sowohl von
Schülern wie auch von Lehrern ist heute weitgehend ausgeschöpft; die Frage ist, welche
Prioritäten von Kindern, Eltern, Schule und Gesellschaft gesetzt werden. Die begabtesten
Schüler sind oft mit vielen weiteren Aktivitäten (Musik, Sport etc.) ausgelastet. Die
engagierten Lehrerinnen und Lehrer sind meist schon reichlich mit ähnlichen Aktivitäten
eingedeckt. Auch als Weiterbildung konkurriert solch eine AG mit anderen Angeboten und ist
aus Sicht der Lehrer nur von relativem Wert. Insbesondere waren die Lehrerinnen und
Lehrer in den wöchentlichen Besprechungen in erster Linie an Inhalten der Klasse (a) und
(b), d.h. an Informationen über die spezielle Programmierumgebung bzw. den Umgang mit
der Rechnerinfrastruktur der jeweiligen Schule interessiert.
Auf der Ebene der Lehrer ist weiterhin festzustellen, dass ein solcher anspruchsvoller
Unterricht sehr betreuungsintensiv ist; 15 Lernende auf eine Lehrkraft plus eine
Lehrassistenzkraft sind als Obergrenze anzusehen. Die Durchführung als AG war eher
zeitproblematisch, da die Kurse mit anderen Aktivitäten der Lernenden und Lehrenden
konkurrierten. Hier zeigten sich auch signifikante Unterschiede in der Kultur der einzelnen
Schulen: Während in einigen Schulen gegen Ende die Teilnehmerzahl deutlich abnahm,
blieben in anderen Schulen alle Schülerinnen und Schüler von Anfang bis zum Schluss
dabei. Die Mischung verschiedener Schulformen (Haupt/Realschule und Gymnasium)
funktionierte nicht, da die Motivation der Gymnasiastinnen und Gymnasiasten im
Allgemeinen höher und vor allem stetiger war.
In einzelnen Schulen ergaben sich ziemliche Schwierigkeiten mit der Infrastruktur, mit falsch
verstandenen Sicherheitskonzepten und mit administrativen Dingen. Die technischen
Probleme waren jedoch meist hausgemacht, durch falsche Systemkonfigurationen oder
unzureichende Administration verursacht. Die AGs verhalfen den Schulen zur Profilbildung,
die in den meisten Fällen noch heute (2009) das Profil mitprägen.
Insgesamt lässt sich konstatieren, dass mit der AG.com gezeigt wurde, dass Informatik als
Sek1-Fach gut geeignet ist. Die Auswahl der richtigen Lehr- und Lernsoftware ist wichtig,
jedoch nicht von entscheidender Bedeutung. Wichtiger ist es, Technik und Inhalte sauber zu
trennen und die Stoffauswahl der Erlebenswelt der Ziel-Altersgruppe anzupassen. Die
AG.com wurde, nachdem der Autor die Projektleitung abgegeben hatte, mit Schwerpunkt im
Bereich Robotik fortgesetzt. Die wesentlichen Ideen tragen dabei noch heute und werden
voraussichtlich im Schuljahr 2009/2010 in einer vom Autor organisierten und von der
Telekom-Stiftung finanzierten „Junior-Ingenieur-Akademie Informatik“ in Berlin weitergeführt.
302
Literatur
[1] AGcom – Arbeitsgemeinschaften am Computer in Bremer Schulen;
http://www.agcom.uni-bremen.de/
[2] GI-Rahmenempfehlung für die Informatik im Unterricht der Sekundarstufe 1;
Informatik-Spektrum, 9 (1986) 2, S. 141-143
[3] Lehrpläne Informatik an Gymnasien;
Deutscher Bildungsserver http://www.bildungsserver.de,
siehe auch Bundeswettbewerb Informatik, http://www.bwinf.de/index.php?id=128
Using Orientation Calculi and Route Graphs
to Interpret Route Descriptions
Hui Shi
Universität Bremen and DFKI-Lab Bremen, Germany
{shi}@informatik.uni-bremen.de
Abstract. This paper summarizes two papers Prof. Bernd Krieg-Brückner
and myself published at the 4th International Conference on Geographic
Information Science 2006 [7] and in the International Journal of Software
and Informatics 2008 [14], in memory of a successful, pleasant and hard
joint work with him.
Keywords: Qualitative Spatial Calculi, Route Graph, Navigation Space,
Human-Robot Interaction, Conceptual Model, Algebraic Specification
1
Motivation
The major research aim here is to develop effective and natural communication between humans and robots about spatial issues, e.g. dialogues between a
driver and the Bremen semi-autonomous wheelchair Rolland [10]. A particular goal is the resolution of shared-control problems caused by spatial knowledge
mismatches between humans and robots through clarification dialogues.
Following the main ideas of the Spatial Semantic Hierarchy [8] we introduce
two distinct spatial representations in the human-robot interaction on route navigation. The “metric level” represents the spatial knowledge of a robot; and the
“conceptual level” the spatial knowledge of a user. In this context we are aiming for conceptual modelling humans’ spatial knowledge by integrating qualitative orientation calculi, in particular Freksa’s Double-Cross calculus, with Route
Graphs, to interpret users’ route descriptions for natural communication, and to
obtain a semantically well-founded interface to the robot’s internal spatial representation, such that an efficient mapping between them can be realized. As an
example, we would like to give a semantic interpretation to the route description
in our office building, cf. Fig. 1.
This paper is structured as follows: We introduce qualitative spatial calculi
and Route Graphs in Section 2. The conceptual model for interpreting route
descriptions is introduced in Section 3. In Section 4 a brief formalization of the
model in the algebraic specification language CASL is presented.
2
Qualitative Spatial Calculi and Route Graphs
Route descriptions are mental constructions that are schematized [17]. Certain
information, such as exact metric data, is often systematically simplified and
304
Fig. 1. Integrating the routes from the lift and the secretariat to Prof. Krieg-Brückner’s
office into a simple Route Graph
even distorted. Several empirical studies show that the critical elements of route
descriptions are landmarks and paths, links and nodes [16, 3, 15], thus topological
maps are often used to model such graph-like structures (e.g., [8, 17, 18]). In
[6] Krieg-Brückner et al. give a standard ontological definition of the Route
Graph for navigation by various agents in different applications; we will use it to
represent route descriptions, such that structural relations of route navigation
can be inferred.
On the other hand researches on qualitative spatial representation and reasoning try to model and to abstract common sense knowledge about space, such
that an automation could model, explain and inform about a human’s behaviour
in space, even if precise quantitative information is not available or is too complicated to be computed [2]. Considering the empirical evidence collected from
some our empirical experiments [15], we chose Freksa’s qualitative spatial calculus [5, 19] to represent and to reason about orientation information involved in
route descriptions.
2.1
The Route Graph Model
RouteGraphs have been introduced as a general concept for navigation by various agents in a variety of scenarios [18, 6]. They can be used as metrical maps
with sensory input and additional metrical computation models to control the
navigation of robotic agents in the environment, or be used at the cognitive level
to model abstractly humans’ topological knowledge while they act in the space.
The most important character of Route Graphs is that different routes can be
integrated into a graph-like structure in which the information concerning these
routes are composed.
Route Graphs are a special class of graphs. A Node of a Route Graph, called
a Place, has a particular position and has its own “reference system”; it may
be rooted in a (more) global reference system. An Edge of a Route Graph is
directed from a source to a target node. We call an edge a RouteSegment; it
always has three additional attributes: an Entry, a Course and an Exit. Exactly
what infomation is associated with these attributes is specially defined for each
305
directions in natural language:
Leave the lift into the stairway
Turn left, go straight ahead to the glass door
Go through the door
Turn right, follow the corridor until the last door on the lefthand-side
– Face it, enter the room Cartesium 2.041
–
–
–
–
Fig. 2. Sample Route: from the lift to Prof. Krieg-Brückner’s office
enter the room Cartesium 2.041 mapped to a RouteSegment
–
–
–
–
–
Source: corridor
Entry: towards the door of the room Cartesium 2.041
Course: go through the door
Exit: [against the door]
Target: the room Cartesium 2.041
Fig. 3. A Route Segment and its components
Route Graph instantiation. For example, an entry/exit at the cognitive level may
contain qualitative orientation information (e.g., to the left/right), the course is
then the path between two reorientations. Consider the example in Fig. 2 and
Fig. 3. A Route is a concatenation of a sequence of route segments from one
place to another. For more information please see [18, 6].
2.2
Qualitative Spatial Reasoning using Orientation Information
In the Double-Cross calculus [5, 19], the concept orientation grid is introduced
to represent qualitative orientation information. The grid is aligned to the orientation determined by two points in 2-dimensional space, the start point and
the end point of a movement represented as an oriented edge. Combining the
front/back and the left/right dichotomy, the Double-Cross calculus may distinguish 15 meaningful disjoint orientation relations, see Fig. 4.
The edge between the two points gives rise to the orientations front and back
on this line, also leftAtExit, rightAtExit, leftAtEntry, and rightAtEntry, whose
terminology we have borrowed from Route Graphs in this paper; left and right
correspond to two areas w.r.t. the edge, similarly leftFront, rightFront, leftBack,
and rightBack; finally, the special positions atEntry, atExit and onCourse are
introduced.
306
front
rightFront
front
leftFront
rightFront
leftFront
left
right
leftBack
rightBack
back
(a)
atExit
onCourse
atEntry
leftAtExit
left
leftAtEntry
rightAtExit
right
rightAtEntry
leftBack
rightBack
back
(b)
Fig. 4. Orientation grid with 8 (a) and 15 (b) relations for the Double-Cross calculus
3
A Conceptual Model for Route Navigation
We are now going to develop a spatial model, called conceptual model, to represent humans’ route descriptions, which combines the Double-Cross calculus
and the Route Graph. Its structure and operational properties are then formally
specified in Section 4.
Let O be the type of orientations, L of locations, M of landmarks and V of
vectors. The function at defines the location of a given landmark, thus at(m) = l
means that landmark m is at location l. The conceptual model is in fact a special
kind of Route Graph P, S, where P ⊆ L is the set of reachable places and S
⊆ V the set of route segments. Furthermore, the following functions are defined:
origin:
P → V, the reference frame of a place
oCourse: S → V, the course of a route segment
oEntry: S → O, the entry of a route segment
oExit:
S → O, the exit of a route segment
The elementary relation of the conceptual model is the orientation relation ori, which defines the orientation of a location with respect to a vector: ori(v, l, rightFront), for example, means l is on the right front of v, i.e.,
v : l = rightFront in the Double-Cross calculus.
People often use spatial relations in their route descriptions, for example,
“the main stairs are on the left” just states the spatial relation between the
“main stairs” and the current position (i.e., the current place and direction);
“through the door” includes a movement and a relation: after the movement
the door is behind. Based on the relations ori and at , a set of relations can be
introduced for defining such spatial relations.
via(v, m) ⇔ ∃l. at(m) = l ∧ ori(v, l, onCourse)
passOnLeft(v, m) ⇔ ∃l. at(m) = l ∧ ori(v, l, left)
passOnRight(v, m) ⇔ ∃l. at(m) = l ∧ ori(v, l, right)
passBy(v, m) ⇔ ∃l. at(m) = l ∧ (passOnLeft(v, l) ∨ passOnRight(v, l))
(1)
(2)
(3)
(4)
As defined, a route is a sequence of connected route segments. To integrate
two routes into a route graph, one has to come up with a common location in the
environment, and all related entries and exits on one route at that location must
be recomputed to conform with the other route at that location. Furthermore,
the consistency of the spatial relations related to the common location should
be proved according to the Double-Cross calculus [5].
307
4
Formalizing the Conceptual Model in Casl
In this section we give an overview to the specification of the conceptual model
(see Section 3) using the Common Algebraic Specification Language, Casl [1,
12]. A detailed and complete specification can be found in [14]. A Casl specification basically includes declarations, to introduce components of signatures,
such as sorts to introduce types, operations or predicates, as well as axioms written in first-order logic, to give properties of those structures. Anybody, who has
some knowledge about algebraic specifications, should have no serious problem
to understand the specifications given in this section.
4.1
Ego Orientation Calculus
While the Cardinal Direction calculus (cf. [4, 11]) and the Star calculus [13]
are based on a global reference frame, the ego orientation calculus uses local
reference frame associated to each point. Depending on the application, we may
want to distinguish 2, 4, and 8 orientations (or more); the more orientations we
consider, the finer our statements become, but we generate more computational
complexity in this way.
The Casl specification EgoOrientation introduces the sort Orientation
and its three subsorts. Orientation2 has two orientations front and back; Orientation4 extends Orientation2 with left and right; finally Orientation8 extends
Orientation4 with additional four orientations: rightFront, leftFront, rightBack
and leftBack. Several operations are defined for the orientations, inverse (∼ o)
and complement ( o ); plus (+) and minus (−), denoting addition and subtraction of orientations. + is associative, commutative and has f ront as a unit
element, thus (Orientation, f ront, +) constitutes a commutative monoid.
All the properties defined in EgoOrientation hold generally for all ego
orientation calculi, no matter how fine the orientations are. Here are some examples:
•a=a
• ∼ front = back
4.2
• a + a = front
•∼∼a =a
• back + back = front • left = ∼ right
Locations and Vectors
The definition of locations extends the pre-defined specification Identities with
a new sort Location and an operator id associating each location with an identifier (Fig. 5). As an extension of EgoOrientation and Locations, the specification Vectors introduces the sort Vector that denotes directed connections
between two locations.
Each vector has a source and a target location, defined by two functions
source and target. Vectors are constructed by the operation “−→ ”. Note that
“→?” is a partial constructor operation in Location × Location →? V ector;
the additional axiom definedness states that the vector is defined, only if the
operation −→ is applied to two different locations.
308
spec Locations = Identities then
sort Location
op
id : Location → Id
end
spec Vectors = EgoOrientation and Locations then
sort Vector
ops
−→ : Location × Location →? Vector
source, target : Vector → Location
• def x −→ y ⇔ ¬ x = y
ops − : Vector → Vector ; %% inverse
:Vector × Vector → Orientation
∀ v, w : Vector
• v w = ∼ (v − w )
•v −w =−v w
•−−v =v
• v w = (w v)
then %implies
∀ v, w : Vector
•v w =−v −w
• − v w = ∼ (v w )
end
%% definedness
Fig. 5. Specification of locations, vectors, and orientations between them
The operation “−” inverts the direction of a vector; the operation “ ” yields
an orientation between two vectors: v w is called the orientation of w with respect to v. The algebraic properties nicely relate the orientation relation between
vectors ( ), vector inversion (−), and the orientation operations inverse (∼ o)
and complement (o).
4.3
The Double-Cross Calculus
We now want to define the Double-Cross calculus in terms of the Ego Orientation calculus. Note that we defined a variant of the Double-Cross calculus. The
orientation o of a point c with respect to a vector ab (i.e., ab : c = o in the
Double-Cross calculus) is denoted as the orientation relation between the two
vectors ab and bc (i.e., a −→ b b −→ c o). Additionally, we can define the
orientation between any two vectors if the source or target of one vector equals
either the source or the target of another one, specified by the axiom (O0) in
Fig. 7.
We start with the Double-Cross calculus with only eight orientation relations.
Fig. 6 shows the principle of defining each relation of the Double-Cross calculus
as a conjunction of two relations in the Ego Orientation calculus, a pattern that
will repeat in the specification. Example (a) in Fig. 6 shows the definition of left;
the other 7 relations follow the same pattern, cf. the axiom (O8). The pattern
is still the same when we extend the Double-Cross with relations at the entry,
exit on course, and at the position entry and exit, respectively, yielding fifteen
relations, the related Casl axioms are omitted here. For the predicate onCourse
we also need two ego orientation frames in the same pattern (Fig. 6(b)).
309
front
leftFront
rightFront
b
left
c
leftBack
right
left
rightBack
exit
back
left
right
a
leftBack
entry
rightFront
Double−Cross
(a)
c
front
c
leftFront
leftBack
Ego−Orientation
exit
c
back
left
rightBack
rightBack
leftBack
left
c
front
leftFront
c
front
rightFront
b
right
leftFront
rightFront
onCourse
entry
right
a
rightBack
Ego−Orientation
Double−Cross
(b)
Fig. 6. Relation between two Ego and corresponding Double-Cross orientations: (a)
the definition of left as the conjunction of leftBack and leftFront; (b) oCourse as the
conjunction of back and front
spec DoubleCrossCalculus = Vectors
then
sorts Orientation8 < OrientationDCC13 ;
OrientationDCC13 < OrientationDCC
OrientationDCC < Orientation
pred
: Vector × Vector × OrientationDCC
∀ a, b, c, d : Location; o : OrientationDCC
• a −→ b c −→ d o ⇒ a = c ∨ b = c ∨ a = d ∨ b = d
%% (O0)
∀ entry, exit, c : Location; u, v, w : Vector
• v = entry −→ exit ∧ w = exit −→ c ∧ u = entry −→ c ⇒
(v w leftFront
⇔ v w = leftFront ∧ v u = leftFront)
∧
(v w left
⇔ v w = leftBack ∧ v u = leftFront)
∧
(v w leftBack
⇔ v w = leftBack ∧ v u = leftBack )
∧
(v w front
⇔ v w = front
∧ v u = front)
∧
(v w back
⇔ v w = back
∧ v u = back )
∧
(v w rightFront
⇔ v w = rightFront ∧ v u = rightFront) ∧
(v w right
⇔ v w = rightBack ∧ v u = rightFront) ∧
(v w rightBack
⇔ v w = rightBack ∧ v u = rightBack ) %% (O8)
then
free type OrientationDCC13 ::= sort Orientation8 |
leftAtEntry | rightAtEntry | leftAtExit | rightAtExit | onCourse
free type OrientationDCC ::=
sort OrientationDCC13 | atEntry | atExit
···
end
Fig. 7. The Double-Cross calculus with 8, 13, and 15 orientations
310
4.4
Route Graphs
To introduce Route Graphs we are interested in special kinds of locations and
vectors to define concepts like places and route segments. A place is a location
and has its own local reference system, as specified in Fig. 8. The operation
oOrigin defines the inherent reference direction as a vector at a place, i.e., the
local reference frame (or origin) at that place. The origin may also be used to
ground a place in a global reference system. A route segment of the sort Segment
is a vector from a source place to a target place, thus it is in fact a subsort of
Vector, such that the source and target of a segment belong to the sort Place. In
the context of route descriptions here, an entry or an exit is in fact represented as
an orientation with respect to a given origin at this place. The operation oEntry
defines the orientation of a segment with respect to the origin at its entry (see
axiom (S1)), and oExit the orientation of the origin at its exit with respect to
the segment (see axiom (S2) in Fig. 8).
The second part of the specification Segments contains five axioms describing some properties of the orientations between segments (“ ”) and the entries
and exits of segments.
spec Places = DoubleCrossCalculus then
sort Place < Location
op
oOrigin : Place → Vector
end
spec Segments = Places then
sort Segment = {s : Vector • source(s) ∈ Place ∧ target(s) ∈ Place}
ops oEntry, oExit : Segment → OrientationDCC ;
oCourse : Segment → Vector ;
−→ : Place × Place →? Segment
∀ x, entry, exit : Place
• source(oOrigin(x )) = x
• oEntry(entry −→ exit) = oOrigin(entry) entry −→ exit
%% (S1)
• oExit(entry −→ exit) = entry −→ exit oOrigin(exit)
%% (S2)
then %% recomputation of entry or exit
∀ a, b, c, d : Place
• a −→ b b −→ c = a −→ b b −→ d + b −→ d b −→ c
%% (S3)
• a −→ b b −→ c = oExit(a −→ b) + oEntry(b −→ c)
%% (S4)
then % implied
• oExit(a −→ b) = front ⇒ a −→ b b −→ c = oEntry(b −→ c)
• oEntry(b −→ c) = front ⇒ a −→ b b −→ c = oExit(a −→ b)
• a −→ b b −→ c = a −→ d d −→ c ⇒
oExit(a −→ b) + oEntry(b −→ c) = oExit(a −→ d ) + oEntry(d −→ c)
end
Fig. 8. Specification of places, route segments
311
spec RoutePredicates = Segments then
via : Segment × Location
pred
∀ a, b, c: Place
• a−→ b via c ⇔ a−→ b b−→ c onCourse
then
preds passBy : Segment × Location;
pass on : Segment × Location × OrientationDCC
∀ a, b, c: Place; lr : OrientationDCC
• a−→ b pass c on lr ⇔
(∃ d : Location • a−→ b via d ∧ a−→ d d −→ c lr )
• a−→ b passBy c ⇔
a−→ b pass c on left ∨ a−→ b pass c on right
then
sort Landmark < Location
ops start, doorLift, doorGlass, door2041, p1, p2, p3 , p4, p5 : Place;
doorOfLift, glassDoor, doorOf2041, lift, stairWay, room2041 : Landmark
• oEntry(start−→ doorLift) = front
• at(lift)=start
• start−→ p1 via doorLift
• at(doorOfLift)=doorLift
• start−→ p1 p1 −→ p2 left
• at(stairWay)=p1
• p2 −→ p3 via doorGlass
• at(glassDoor )=doorGlass
• p2 −→ p3 p3 −→ p4 right
• p3 −→ p4 p4 −→ door2041 left
• at(doorOf2041 ) =door2041
• oEntry(p4 −→ p5 ) = left
• p4 −→ p5 via door2041
• at(room2041 )=p5
end
Fig. 9. Route predicates and the sample route
Based on the Casl specification Places and Segments the relations necessary for the interpretation of route descriptions are presented in Fig. 9, where
the sample route description (see Fig. 2) is specified by a set of axioms.
In the conceptual model discussed and specified in this paper landmarks are
treated as points and their spatial extensions are ignored. Hence, route descriptions like “go into the park” and “go along the river” can not be interpreted
in the current model. A possible extension of the model deal with region-like
landmarks is suggested in [9].
Acknowlegment. My thanks go to Dr. Klaus Lüttich and Dr. Christian Maeder
for helping us to check the consistency of CASL specifications.
References
1. E. Astesiano, M. Bidoit, B. Krieg-Brückner, H. Kirchner, P. Mosses, D. Sannella,
and A. Tarlecki. CASL - The Common Algebraic Specification Language. In
Theoretical Computer Science, number 286. Springer-Verlag, 2002.
2. A. G. Cohn, B. Bennett, J. Gooday, and N. M. Gotts. Qualitative spatial representation and reasoning with the region connection calculus. Geoinformatics, 1:1–44,
1997.
312
3. M. Denis. The description of routes: A cognitive approach to the production of
spatial discourse. Cahiers de Psychologie Cognitive, 16:409–458, 1997.
4. A. U. Frank. Qualitative spatial reasoning with cardinal directions. In Proceedings
of the Seventh Austrian Conference on Artificial Intelligence. Springer, 1991.
5. C. Freksa. Using orientation information for qualitative spatial reasoning. In
Theories and Methods of Spatio-Temporal Reasoning in Geographic Space, volume
639 of Lecture Notes in Computer Science, pages 162–178. Springer-Verlag, 1992.
6. B. Krieg-Brückner, U. Frese, K. Lüttich, C. Mandel, T. Mossakowski, and R. J.
Ross. Specification of an ontology for route graphs. In C. Freksa, M. Knauff,
B. Krieg-Brückner, B. Nebel, and T. Barkowsky, editors, Proceedings of Spatial
Cognition IV, Chiemsee, Germany, volume 3343 of Lecture Notes in Artificial Intelligence, pages 989–995. Springer, 2004.
7. B. Krieg-Brückner and H. Shi. Orientation calculi and route graphs: Towards
semantic representations for route descriptions. In Proceedings of GIScience 2006,
Münster, Germany, volume 4197 of Lecture Notes in Computer Science, 2006.
8. B. Kuipers. The spatial semantic hierarchy. Artificial Intelligence, 119:191–233,
2000.
9. Y. Kurata and H. Shi. Rfdl: Models for capturing directional and topological
characteristics of path-landmark arrangements. In N. Van de Weghe, R. Billen,
and P. Kuipers, B. Bogaert, editors, International Workshop on Moving Objects,
2008.
10. A. Lankenau and T. Röfer. Mobile robot self-localization in large-scale environments. In Proceedings of the IEEE International Conference on Robotics and Automation, pages 1359–1364, 2002.
11. G. Ligozat. Reasoning about cardinal directions. Journal of Visual Languages and
Computing, 9:23–44, 1998.
12. P. D. Mosses and M. Bidoit. Casl — the Common Algebraic Specification Language: User Manual, volume 2900 of Lecture Notes in Computer Science. Springer,
2004.
13. J. Renz and D. Mitra. Qualitative direction calculi with arbitrary granularity. In
Proceedings PRICAI 2004, volume 3157 of Lecture Notes in Artificial Intelligence,
pages 65–74. Springer, 2004.
14. H. Shi and B. Krieg-Brückner. Modelling route instructions for robust human-robot
interaction on navigation tasks. International Journal of Software and Informatics,
2(1):33–60, 2008.
15. H. Shi and T. Tenbrink. Telling Rolland where to go: HRI dialogues on route navigation. In WoSLaD Workshop on Spatial Language and Dialogue, Delmenhorst,
Germany, October 23-25, 2005.
16. L. Talmy. How language structures space. In H. L. Pick and L. P. Acredolo, editors,
Spatial Orientation: Theory, Research and Application. Plenum, NY, 1983.
17. B. Tversky and P. Lee. How space structures language. In C. Freksa, C. Habel,
and K. Wender, editors, Spatial Cognition: An interdisciplinary Approach to Representation and Processing of Spatial Knowledge, volume 1404 of Lecture Notes in
Artificial Intelligence, pages 157–175. Springer-Verlag, 1998.
18. S. Werner, B. Krieg-Brückner, and T. Hermann. Modelling navigational knowledge
by route graphs. In Spatial Cognition II: Integrating Abstract Theories, Empirical
Studies, Formal Methods, and Pratical Applications, volume 1849 of Lecture Notes
in Artificial Intelligence, pages 259–316. Springer-Verlag, 2000.
19. K. Zimmermann and C. Freksa. Qualitative spatial reasoning using orientation,
distance, and path knowledge. Applied Intelligence, 6:49–58, 1996.
313
Hui Shi received her master degree in Computer Science from the Changsha Institute and Technology,
China and her Dr.-Ing. degree from Universität Bremen, Germany, in 1987 and 1994. respectively. She
is a senior researcher at DFKI Bremen, Safe & Secure Cognitive Systems and member of the DFGfunded SFB/TR 8 “Spatial Cognition”. Among others
she is working in the following research areas: Formal
Methods, Shared Control, User Modeling and Dialogue
Modeling.
Top-down vs. Bottom-up:
Transformationelle Methoden im
wissenschaftlichen Wettbewerb
Ein essayistischer Überblick über den Stand der Kunst
Burkhart Wolff
Université de Paris-Sud,
Laboratoire de Recherche Informatique,
Parc Club, 4 Rue Jaques Monod
F-91405 Orsay Cedex, France
Mail: [email protected], WWW: http://www.lri.fr/~wolff
Zusammenfassung In diesem Beitrag wird versucht, einen Überblick
über den Wettbewerb zweier konkurrierender Forschungsprogramme —
genannt Top-down oder Transformationelle Entwicklungsmethodik versus Bottom-up oder post-hoc Programmverifikation — im Bereich “Formaler Methoden” zu geben. Als Einordnungsrahmen benutze ich Lakatos’s Konzept des “wissenschaftlichen Forschungsprogramms”. Es ergibt
sich ein bewusst altmodischer Versuch — im Gegensatz zu modischen
bibliometrischen Kriterien (vulgo: Zahlen von Veröffentlichungen) — inhaltliche Kriterien für die Qualität und damit den Fortschritt der wissenschaftlichen Arbeit in unserem Gebiet zu entwickeln.
1
Einleitung
Man könnte diesen Artikel mit dem Glaubensbekenntnis beginnen lassen: “Formale Methoden helfen dabei, die Sicherheit von Software zu erhöhen”. Oder auch
inhaltlicher, aber eklektischer: “Formale Methoden sind eine Sammlung von mathematischen Beschreibungsmitteln und Analyseverfahren, die ihre Grundlage in
diskreter Mathematik haben”. Wir alle haben solche Sätze vermutlich schon oft
gehört, und — wenn auch mit Unbehagen — geschrieben.
Statt dessen wird in diesem Beitrag versucht, “Formale Methoden” (FM) als
“Wissenschaftliches Forschungsprogramm” zur Fundierung des Software-Engineerings zu begreifen und zwei seiner wichtigsten Unterprogramme — Transformationelle Programmentwicklung vs. post-hoc Programmverifikation — gegenüberzustellen. Als Begriffsrahmen für die Grundkonzepte “Kern”, “wissenschaftliches
Programm”, “Anomalie”, etc., dient dabei die Erkenntnistheorie Lakatos’s [64].
In der Informatik fängt bekanntlich alles mit Alan Turing an. Als ausgebildeter Mathematiker mit starker Ausrichtung auf logische Grundlagen quälte
er sich allerdings nicht mit der neueren Frage, ob er “Formale Methoden” nun
benutzen sollte oder nicht bei dem Tun, das wir heute Informatik nennen —
er tat es einfach, so wie man Luft zum Atmen braucht. Er entwickelte damit
315
Konzepte wie Maschine, Zustand, und sogar Zustandsinvariante, die unbestritten zum Grundinventar moderner Programmiersprachentheorien gehören, und
beschäftigte sich mit den Zusammenhängen funktionaler und zustandsorientierter Programme. Allerdings galt Turing auch bald als akademischer Nestbeschmutzer, interessierte er sich doch (Pfui!!!) auch für die konkreten Probleme
des Computerbaus und seiner Programmierung.1
Während die Bedeutung der mathematisch-logischen Grundlagen für die Geschichte der Informatik und für den akademischen Betrieb akzeptierter Fakt ist,
ist ihre Relevanz für die ingenieursmässige Konstruktion von Software-Systemen,
also das Software-Engineering, höchst umstritten, auch im akademischen Betrieb
selbst. Sichtbarster Ausdruck für die interne Geringschätzung ihrer Grundlagen
ist die Tatsache, das die “curriculum recommandations” der ACM für das Fach
Informatik im Jahr 2005 keinerlei mathematische Fähigkeiten mehr erwähnen,
und lediglich einen Kurs zur Theorie der Programmiersprachen vorsehen[32].
Das akademische Misstrauen an FM wird zudem vertieft durch die Tatsache,
das es aus Sicht des mathematischen Mainstreams als Extrem, als Gipfelpunkt
eines älteren Forschungsprogramms wahrgenommen wird, das nach allgemeinen Dafürhalten als gescheitert gilt: dem Programm des Formalismus nämlich.
Dies begann mit Leipnitz’ Idee einer “lingua universalis”, einer Sprache in der
sich alle wissenschaftlich-mathematischen Probleme formulieren und mittels deduktiven Schliessen beantworten liessen, über erste technische Beiträge durch
Boole, Peano und Frege bis hin zu Whitehead und Russels “Principia Mathematica”, die das Programm erstmalig konkret ausformulierten. Die Unmöglichkeit
der Vollständigkeit und Entscheidbarkeit hinreichend ausdrucksstarker logischer
Sprachen, und damit das Ende des Hilbertschen Traums von “Wir müssen wissen, und wir werden wissen!”, diskreditierte in den Augen vieler auch den Formalismus als Grundlage der Mathematik. Die Debatte mag in jüngster Zeit wiederaufgeflammt sein [55,56,39,40], angefeuert durch die zunehmende Schwierigkeit,
die sich häufenden Monsterbeweise in mathematischen Journalen tatsächlich
nachzuprüfen, und diesbezügliche Erfolge durch mathematische Beweisassistenten, d.h. interaktive Theorembeweisumgebungen. Doch zunächst einmal wird
sich an der allgemeinen Formalismusskepsis nichts ändern, auch aus Gründen
der Bequemlichkeit und Opportunität.2
1
2
Hardy verhinderte noch Turings Promotionsvorschlag einer Interpreterimplementierung für den getypten Lambdakalkül mit den Worten “Nothing is worth a Phd that
can be done with a computer”.[42] Wir wissen heute recht gut Bescheid über das
Problem der Normalisierung des getypten Lambda-Kalküls und seiner Bedeutung für
die Theorie der funktionalen Programmierung als auch den Bau von Beweissystemen
für die Logiken höherer Stufe; auch die technischen Herausforderungen solcher Programme auf der Hardware der fünfziger Jahre würden heute noch Respekt abnötigen.
Nur — wer zum Teufel war doch gleich dieser Hardy???
Ich darf den Kollegen Ralf Küsters zitieren mit seinem Ausspruch: “Für mich ist ein
Beweis ein Beweis, wenn ich den Gutachter davon habe überzeugen können, das sich
daraus vielleicht ein formaler Beweis machen liesse. Das machen alle im Betrieb so!”
[48].
316
2
Exkurs: Imre Lakatos “Kritischer Rationalismus”
Die Abgrenzung der Wissenschaft von Pseudowissenschaft ist ein grundlegendes
erkenntnistheoretisches Problem: sollten wir etwa, um ein extremes Beispiel zu
gebrauchen, der Erforschung des Einflusses der Bachblütentherapie auf Programmiererteams den Zugang zu öffentlich geförderten Forschungsmitteln gewähren?
Der leicht-fertige Einwand: “Nein, das ist pseudo-wissenschaftlich”, führt unmittelbar zur Frage, was solche “Ansätze”, die sich bei Bedarf mit klingenden
Adjektiven wie “ganzheitlich-partizipativ” schmücken lassen, denn inhaltlich von
wissenschaftlicher Arbeit unterscheidet. Und weitergehend auch die Frage, was
wissenschaftlichen Fortschritt denn ausmacht und wie er sich denn bewerten liesse — eine Frage, die im Zeitalter zunehmender Ökonomisierung, in dem Erfolg
in Geld umgerechnet wird und gleichzeitig Geld Mass des Erfolges ist, leicht ein
existentielles Beigeschmäckle hat.
Die Frage ist uralt, und hat, wie wir sehen werden, keine Beil-scharfen Lösungen. Bekanntlich schlug Popper[58] folgendes Kriterium vor: “Eine Theorie ist
wissenschaftlich, wenn man bereit ist, im Voraus ein entscheidendes Experiment
(oder eine Beobachtung) anzugeben, das sie falsifizieren würde, und sie ist unwissenschaftlich, wenn man zu einem solchen ‘Experimentum Crucis’ nicht in der
Lage ist”. Demnach kann eine Theorie wissenschaftlich sein, obwohl überhaupt
keine Daten für sie sprechen (eine Situation, die gerade die Diskussion um die
String-Theorie beschäftigt), oder unwissenschaftlich, obwohl alle Daten für sie
sprechen. Zweifellos ist es interessant, einen Marxisten nach dem Experimentum
Crucis zu fragen; er wird u.U. die Frage als unmoralisch ablehnen. In der Tat
kann also diese Frage einen Anhaltspunkt dafür liefern, ob eine Theorie zu einem
Dogma erstarrt ist oder zu echter Erkenntnis beiträgt.
Ist nun das Poppersche Falsifikationskriterim die Lösung des Problems der
Abgrenzung von Wissenschaft und Pseudowissenschaft? Nein, denn es berücksichtigt nicht die bemerkenswerte Zählebigkeit wissenschaftlicher Theorien; Poppers Kriterium funktioniert einfach nicht in der wissenschaftlichen Praxis. Wissenschaftler geben eine Theorie nicht auf, nur weil ihr Tatsachen widersprechen
oder Schwierigkeiten auftauchen, weil Anomalien existieren. Mal diffamieren sie
die Anomalie als Monstrosität oder einen irrelevanten Ausnahmefall (Lakatos erfand für diese Verteidigungsstrategie den schönen Namen Monster-barring), oder
mal finden sie eine rettende Hypothese, die die Anomalie erklären soll (Lakatos’
Monster-adaption) oder, wenn das nicht möglich ist, vergisst man die Schwierigkeit einfach und wendet sich anderen Problemen zu. Das wohl berühmteste
Experimentum Crucis der Wissenschaftsgeschichte, das Michelson-Morley Experiment, das ursprünglich konzipiert war, als Richter zwischen zwei konkurrierenden Äthertheorien zu fungieren, verschwand 25 Jahre in der Schublade, bevor
es von Einstein hervorgeholt wurde um damit die spezielle Relativitätstheorie
(kein Äther) gegenüber der Newton’schen Mechanik auszuzeichnen.
Was ist also das Kennzeichen von guter Wissenschaft? Müssen wir kapitulieren und zugeben das eine wissenschaftliche Revolution nichts als eine irrationale
Änderung der Überzeugung, eine religiöse Bekehrung sein? Der amerikanische
Wissenschaftstheoretiker Thomas Kuhn kam zu diesem Ergebnis, nachdem er
317
die Naivität des Popperschen Falsifikationismus erkannt hatte. Da er Kriterien der Unterscheidung zwischem wissenschaftlichem Fortschritt und geistigem
Niedergang prinzipiell nicht mehr anerkennen will, gibt es für ihn auch keine
allgemeinen Kriterien der Rationalität mehr. Sein Schüler Paul Feyerabend [30]
hat diese Position zu einem erkenntnistheoretischen Anarchismus des “Anything
Goes” ausgebaut.
Lakatos Beitrag in der Debatte [64] ist es, die Geschichte der Wissenschaft
ernst zu nehmen und den Begriff des Fortschritts in der erkenntnistheoretischer
Debatte nicht nur auf der philosophischen Spielwiese definieren zu wollen. Für
Lakatos macht es keinen Sinn, eine wissenschaftliche Theorie nur als ein konkretes Hypothesengebäude zu begreifen, das man durch eine Anomalie widerlegen
könne. Denn alle wissenschaftlichen Theorien sind für ihn Inseln in einem “Meer
von Anomalien”, und Erkenntnisfortschritt zeichnet sich dadurch aus, das diese
Inseln Landgewinne machen. Die genaue Formulierung der Hypothesen kann sich
auch über die Zeit verändern, man hat es also eigentlich mit einer Familie von
Theorien, T1 , T2 , ... zu tun, eben dem “Forschungsprogramm”. Ein Forschungsprogramm zeichnet sich aus durch:
1. einen harten Kern, d. i. eine konkrete mathematische, stark formalisierte
Theorie, die zäh verteidigt wird, sowie
2. ein System von Hilfshypothesen, wozu insbesondere auch die Beobachtungstheorien gezählt werden. Modifikationen im System der Hilfshypothesen gehören zum wissenschaftlichen Tagesgeschäft. Und außerdem:
3. einen heuristischen Apparat — oder moderner gesagt: eine Pragmatik — mit
der Anomalien im Sinne des Programms gedeutet, also “erklärt” werden.
Zum Kern der Newtonschen Mechanik zum Beispiel gehören die vier Grundaxiome über die Zusammenhänge von Masse, Kraft, Beschleunigung und Gravitation, etc., sowie die mathematischen Theorien, die zu ihrer Fundierung benötigt
werden (z.B. Axiomatische Mengenlehre, Theorie der Funktionen, Theorie der
Integral- und Differentialrechnung, Differentialgleichungen). Zu den Hilfshypothesen gehören die Theorien über Planetenbewegungen, aber auch Theorien über
die Messinstrumente und Beobachtungstheorien über Experimentalanordnungen. Nehmen wir an, eine Anomalie taucht auf: “Teeblätter in einer Teetasse
sammeln sich nicht, wie von den Gesetzen der Zentrifugalkraft gefordert, aussen,
sondern in der Mitte. Die Newtonsche Mechanik ist folglich unwissenschaftlicher
Unsinn!” Natürlich — und sinnvollerweise — passiert nicht das, was laut naivem
Falsifikationismus zu fordern wäre: alle Newtonianer fallen vom Glauben ab, die
Physik verfällt in einen Zustand von Lethargie und Verzweiflung, und wir treten
in ein neues Zeitalter des Mystizismus ein. Stattdessen wird ein Newtonianer —
je nach Jahrhundert und Entwicklungsstand der Hilfshypothesen — argumentieren: “Mit Flüssigkeiten geben wir uns nicht ab!” (Monster-barring) oder eine
Strömungsmechanik entwickeln, nach der sich aufgrund der Reibung das Maximum der Strömungsgeschwindigkeit einer rotierenden Flüssigkeit nicht am Rand,
sondern bei etwa einem Drittel des Radius vom Rotationszentrum liegt (Monsteradaption). Tatsächlich finden neue Messungen statt — und die Teeblätter
sammeln sich gar nicht in der Mitte, sondern bewegen sich auf Kreisbahnen um
318
das Zentrum herum, und mit etwas Massage gelingt sogar die genaue empirische Vorhersage. Und Newtonianer sprechen nun erhobenen Hauptes von einer
glänzenden Bestätigung des Programms. Tatsächlich ist letzteres — eine Anomalie führt zu neuen Vorhersagen und neuen Beobachtungsergebnissen — für Lakatos Ausdruck eines progressiven Forschungsprogramms, Monster-barring sowie
Monsteranpassung ohne Erweiterung der empirischen Datenbasis sind dagegen
Ausdruck eines degenerativen. Wissenschaftliche Revolutionen finden vor allem
dann statt, wenn neuere Forschungsprogramme nicht nur progressiv sind, sondern auch die empirische Basis älterer Programme mit erklären und übernehmen
können (Ptolemäus vs. Kepler, aber auch Newton vs. Einstein). Im Gegensatz zu
Popper fordert Lakatos also keine “Sofortrationalität” von Forschungsprogrammen, was dazu führt, das sich der Fortschritt z.T. erst im Nachhinein und mit
einigem zeitlichem Verzug erkennen lässt. Dies hat wiederum Feyerabend und
viele Wissenschaftspraktiker zu einer Kritik am Lakatosschen kritischen Rationalismus veranlasst.
So ist der zeitgenössische Glaube an “moderne Evalutationstechniken” in
der Wissenschaftspraxis auch ein Versuch, eine praktikable Antwort auf das
Abgrenzungsproblem zu finden. Leider ist festzustellen, das die Evaluationitis
in Berufungskommissionen und Forschungsförderungsgremien nach wie vor ungehemmt ist und bisherige spektakuläre Fehlentscheidungen — wie jüngst die
Schliessung des fachlich gesunden Fachbereichs Chemie an der Universität Exeter — noch nicht wirklich schmerzen. So bleibt vorerst nur Sarkasmus: Der
ungebrochene Glauben einiger Wissenschaftler und Forschungspolitiker an bibliometrische Evaluations-Kriterien erinnert stark an die religiösen Gefühle von
Bankern beim Anblick der Aktienkurse kurz vor dem Platzen der Blase. Denn
er produziert einen Typus von Wissenschaftler, der Schwalle zweitklassiger Publikationen ausstößt, aber selbst aus Zeitmangel nicht mehr liest; er produziert
eine Forschungskultur, die inhärent nicht mehr fortschrittsfähig ist. Wen eine
ernsthafte, systematische Kritik an bibliometrischen Kriterien und ihrer Manipulierbarkeit wie [2,28] nicht abschreckt, dem ist nicht zu helfen.
3
Des Pudels Kern und das “Meer der Anomalien”
Man kann im Fall der Informatik ohne Schwierigkeit drei Fundamental-Systeme
identifizieren:
1. Klassische Mengentheorie (ZFC),
2. Typtheorie, und
3. Kategorientheorie.
Dies sind Systeme, die zumeist älter als die Informatik selbst sind, trotzdem von
der Informatik in Form von Personal und Fragestellungen (vulgo: Forschungsgeldern) profitieren und auch die Ausprägung gewisser Forschungsunterprogramme
wie etwa “algebraisch top-down” (siehe Abschnitt 5) bis in den heutigen Tag
bestimmen.
319
Auf dieser Grundlage werden Programmier-, Architektur- und Spezifikationssprachen semantisch begründet, die zum Ziel haben, in technisch unterstützter
Form die Praxis der Softwareentwicklung zu ändern — das ist als des Pudels
Kern des FM zu betrachten.
Bei der Forderung nach “technischer Unterstützung” einer Sprache/Methode
ist zwar auch keine “Sofort-Rationalität” zu erwarten, aber wenn ein Teil-Forschungsprogramm fortdauernd keine Erfolge bei der Entwicklung von programmiersprachlichen Konzepten und entsprechenden Compilern, oder von technischunterstützten Entwicklungsmethoden, oder von Beweis-oder Analyseverfahren
aufzuweisen hat, dann ist stark zu vermuten, das es sich um ein ziellos-degenerierendes Programm handelt. Denn Theorien zu bilden allein ist bei der Stärke
der Fundamental-Systeme keine besondere Schwierigkeit und hat beträchtliche
Beliebigkeit, ob sie dagegen tatsächlich etwas zum Bau von Systemen beitragen
können, macht die Substanz ihres Beitrags aus. Dieses Ziel computer-technisch
Unterstützung einer Sprache oder Methode ist meiner Auffassung das fachgerechte Experimentum Crucis was die Abgrenzung zu einem “Anything Goes”
erlaubt.
Von Anbeginn an waren dabei zwei Lager im FM zu unterscheiden:
1. Top-down: Der Ansatzpunkt dieses Teil-Programms sind Spezifikations formalismen, die semantisch dicht an einem der Fundamental-Systeme liegen,
und konsequenterweise weit entfernt von Programmiersprachen und insbesondere zustandsorientierten Programmiersprachen. Der Kern gründet auf
den semantischen Theorien um Sprachen wie Z, VDM, HOL, ACT-ONE,
LOTUS, CSP und neuerdings Event B. Die Entwicklungsmethodik zielt auf
formales Refinement oder Transformationelle Entwicklung.
2. Bottom-up: Der Ansatzpunkt dieses Teil-Programms sind Programmiersprachen und Programmiercode, der möglichst realistisch erfasst werden soll. Der
Kern gründet auf den operational-semantischen Theorien dieser Sprachen.
Die Entwicklungsmethodik zielt einerseits auf optimierte Übersetzung und
Programmtransformation, andererseits auf Typsysteme und prädikative oder
temporale Annotierungssprachen.
Ein typischer Top-downer (wie z.B. der geschätzte Kollege Jean-Raymond Abrial) hält einen Annotierungsansatz der Programmiersprache C für das “Gott-seibei-uns”, stellvertretend für viele Andere, für die die semantische Behandlung
“realer” Programmiersprachen lange Zeit unvorstellbar war und mit dem der
Entwicklungsprozeß nicht belastet werden sollte. Aus Sicht der Top-downer ist
eine Sprache wie C schlichtweg “dreckig”, Vertreter dieses Programms begegnen
dem Problem also zur Zeit mit klassischem Monster-barring.
Ein typischer Bottom-upper (wie z.B. der geschätzte Kollege Peter Müller) 3
hält die Idee für utopisch, das man den Entwicklungsprozeß jemals mit abstrakten Systemmodellen beginnen werde; Informatiker in der industriellen Praxis
3
Mit beiden Kollegen hatte ich über viele Jahre regelmässige, anregende Debatten zu
diesem Thema im “Formal Methods Club” an der ETH Zürich ...
320
sind zunächst einmal Prgrammierer, und an den realen Entwicklungsprozessen
müsse man ansetzen.
Interessanterweise haben beide Forschungsprogramme industrielle Anwender
— sie sind also beide wachsende Inseln mit einem mehr oder weniger breiten
Uferstreifen. Der größte Software-Produzent weltweit hat sich in den letzten
Jahren massiv in der Forschung von annotierungsorientierten Spezifikationssprachen (Spec#, VCC) engagiert, wobei abgeschwächte Formen dieses Ansatzes im großen Stil (z.B. SAL) auch bei der Produktion der aktuellen Version
des Hauptprodukts verwendet worden sind. Währenddessen engagiert sich der
zweitgrößte Software-Konzern insbesondere für modellbasierter Systementwicklung; hier interessieren vor allem Modelle von Geschäftsprozessen, Code dagegen
wird in Wesentlichen generiert und stellt oft kein Primärprodukt mehr dar.
Je nach Anwendungsdomaine haben also beide Forschungsprogramme zunächst einmal Erfolge auszuweisen, auf die im Folgenden näher eingegangen wird.
4
Modellierung
FM bietet mathematisch fundierte Beschreibungsformalismen zur Modellierung
von Aspekten, wie beispielsweise Systemanforderungen, Architekturen, oder von
lokalen Eigenschaften. Durch die Definition dieser Systeme in einem FundamentalSystem werden Mehrdeutigkeiten im Verständnis einer Modellierung ausgeschlossen, weil jedes Modell eine im mathematischen Sinne präzise, eindeutige Bedeutung, die so genannte formale Semantik hat. Diese Fundierung erlaubt darüberhinaus typischerweise die Ableitung von Kalkülen, also Mengen von Regeln, mit
deren Hilfe sich Eigenschaften einer Modellierung berechnen lassen.
4.1
Gibt es Datenorientierte Spezifikationssprachen?
Da Modellierung an für sich im Software-Entwicklungsprozeß keinen Selbstzweck
darstellt, sondern wie gesagt im Hinblick auf technisch unterstützte Analyseverfahren genutzt wird, hat sich die Unterscheidung in daten-orientierte und verhaltensorientierte Spezifikationssprachen eingebürgert, je nachdem, ob die Analysetechnik traditionell eher deduktionsbasiert oder automatenbasiert war (Vergleiche Abschnitt 7). Klassische Vertreter des datenorientierten Modellierungsstils
sind Z, VDM und sein “objekt-orientierter” Nachfahr UML/OCL (an Klassendiagrammen), diverse Hoare Logiken, Dijkstra/Floyds wp-Kalkül als auch diverse algebraische Ansätze wie ACT ONE und CASL. Klassische Vertreter des
verhaltensorientierten Ansatzes sind Prozesskalküle wie CSP als auch temporale
Logiken wie LTL, CTL, Promela oder JavaPathFinder/JML.
Datenorientierte Modellierungen zielen auf die Darstellung einer Transition eines Systemzustands, in der Regel durch Vor- und Nachbedingungen, ab.
Demgegenüber werden bei verhaltensorientierten Modellierungen Transitionsrelationen (ggf. mit Zusatzinformationen wie “Events” angereichert) zur Menge
aller möglichen Abläufe eines Systems (Traces) vervollständigt; über Mengen
321
von Traces können dann Eigenschaften formuliert werden, wie etwa dass alle erreichbaren Zustände des Systems eine gegebene Sicherheitseigenschaft erfüllen.
In Verhaltensmodellen können zudem Nebenläufigkeit, Parallelität und Zeitverhalten von Systemmodellen adäquat beschrieben werden.
Die Übertragung der Unterscheidung datenorientiert/verhaltensorientierter
Modellierung auf Spezifikationssprachen dagegen ist, obwohl weithin üblich,
blanker Traditionalismus, wie man schon an der Willkürlichkeit der Einteilung
von JML und Promela vs. Hoare-Logiken erkennen kann. Es ist kein besonderes
Problem, auch in Z (oder HOL) die Logiken LTL oder CSP zu codieren [65,10]
und dies auch analysetechnisch auszunutzen [10]. Zudem haben Sprachen wie
CSP oder Promela mit Typ-Konstruktoren für Listen, Mengen und Tupel außerordentlich ähnliche Sprachausdrucksmächtigkeiten wie z.B. Z. Bei neueren
Formalismen wie VCC-I[53], VCC-II[22] oder ACSL[11], die mittels Programmannotierungen Aussagen über das nebenläufige Verhalten von C Programmen
machen und hierzu sowohl den SMT-Beweiser Z3 als auch Isabelle/HOL als
Beweiser-Backend benutzen [14,15], wird diese Unterscheidung endgültig ad absurdum.
Sinnvoller als eine Einteilung in datenorientiert bzw. verhaltensorientiert
auf Sprachen erscheint eine Einteilung in Top-down vs. Bottom-up bzw. Konzeptionsorientiert vs. Ausführungsorientiert. Erstere Sprachklasse zeichnet sich
durch einen geringen konzeptionellen Abstand zu einem Fundamentalsystem aus,
während letzere einen geringen Abstand zu einem Ausführungsmodell einer “realen Programmiersprache”, also im Extremfall C auf einem konkreten Prozessor
anstrebt. In erstere Sprachklasse können somit Z, CSP und UML/OCL eingeordnet werden, in letztere z.B. Spec#/Boogie[8,7], JML/Krakatoa[31], JML/ESCJava[45], ACSL/Caduceus[31,11] und VCC-I/II/Boogie[53,22].
4.2
Architekturen und Muster
Musterbasierte Ansätze finden bei der Entwicklung sicherheitskritischer Systeme
an mehreren Stellen Verwendung: Beispielsweise haben Sicherheitsanforderungen
oftmals sich wiederholende Formen (z.B. Ereignis x darf erst stattfinden, nach”
dem Ereignis y stattgefunden hat“). Der Vorteil der Verwendung von Mustern
liegt hier (wie auch in der Softwareentwicklung ganz allgemein) in der Wiederverwendung etablierten Wissens. Es sind daher sog. safety patterns definiert worden,
die helfen, Sicherheitseigenschaften vollständig und präzise zu formulieren. Die
so erhaltenen Anforderungen lassen sich sowohl in einer quasi-natürlichen Fachsprache als auch in einer geeigneten Temporallogik oder einer Prozessalgebra
ausdrücken.
Weiterhin lassen sich wiederkehrende Architekturen sicherheitskritischer Systeme identifizieren. Ein Beispiel für eine solche Referenzarchitektur sind zyklische
Systeme, wo in regelmäßigen Zeitabständen die Sensorwerte abgefragt und Aktorkommandos generiert werden. Aber auch interruptgesteuerte Architekturen
sind weit verbreitet, nicht zuletzt in Mainstream-Prozessoren mit ihrem harwaretechnisch unterstützten Interruptmechanismen.
322
Ein weiterer Top-down Trend lässt sich mit der zunehmenden Verbreitung
der Unified Modeling Language (UML, der derzeitige de-facto-Standard in der
industriellen Software-Modellierung) erkennen, die bei der Modellierung von
Geschäftsprozessen, bei der Entwicklung sicherheitskritischer Systeme und bei
Datenbanksystemen immer weitere Verbreitung findet. UML stellt eine Vielzahl von Teilsprachen und assoziierten diagrammatischen Notationen bereit, mit
der verschiedene Aspekte von Systeme geeignet dargestellt werden können. Hier
sind insbesondere Klassenmodelle zur Modellierung von Daten sowie Sequenzdiagramme und Zustandsmaschinen zur Modellierung von Verhalten zu nennen. Auf
der Ebene der Spezifikationen wird UML verwendet, um sicherheitskritische Systeme zu spezifizieren und diese Spezifikationen formal auf Sicherheitslücken zu
untersuchen. Auf Basis von Expertenwissen in der Entwicklung sicherer Software sind Erweiterungen der UML definiert worden, die das einfache Einfügen von
Sicherheitsanforderungen erlauben (zum Beispiel UMLsec [44] und SecureUML
[9,17,17]). Diese Sicherheitseigenschaften können dann mit Beweiswerkzeugen
(vgl. Abschnitt 7) formal verifiziert werden. Diese Verifikation basiert jeweils auf
einer formalen Ausführungs-Semantik für das verwendete Fragment von UML.
Insbesondere können auf dieser formalen Grundlage wissenschaftliche Fragen beantwortet werden, die unabhängig vom Kontext der modell-basierten Entwicklung von Interesse sind, wie die Verfeinerung von Spezifikationen sicherheitskritischer Systeme, Sicherheitsaspekte geschichteter Architekturen (wie Protokollschichten), und die (De-)Komponierbarkeit und Modularität sicherheitskritischer Systeme.
Die Forschung nach Architekturen oder Mustern gehört also zu den Kernstücken des Top-down Programms; nach einer Zeit der Euphorie nach dem GoFBook [33] ist allerdings insgesamt eine gewisse Stagnation zu verzeichnen.
5
Programmiersprachen
Naturgemäß ist die Forschung nach Programmiersprachen, und insbesondere ihrer semantischen Begründung eine der zentralen Interessengebiete im FM. Komplementär zum Abschnitt 4.2, wird die Forschung allerdings primär von Bottomuppern vorangetrieben.
Unschwer lassen sich Top-down und Bottom-up Strömungen unterscheiden,
seitdem es Sprachen gibt: Funktionale und logische Sprachen einerseits und imperative Sprachen und Assemblersprachen mit mehr oder weniger starker Abstraktion gegenüber unterliegenden Maschinenmodellen andererseits. Aus dieser
Perspektive nehmen objektorientierte Sprachen in gewisser Weise eine Zwitterstellung ein: einerseits sind sie tiefgreifend zustandsorientiert (Objektreferenzen
bilden Objektgraphen mit mutierbaren Objektattributen...), andererseits stellen
objektorientierte Zustandsmodelle wie in Java eine beträchtliche Abstraktion
gegenüber selbst einfachsten Speichermodellen wie linearem Speicher dar. Der
Zwittercharakter wird deutlich, wenn man die Hybridsprachen von der funktionalen Seite wie OCaml, Pizza oder SCALA betrachtet, zu denen die Unterschiede
sich zunehmend verwischen.
323
Die Fundierung von Top-down - als auch Bottom-up mittels formaler, maschinengeprüfter Sprach- und Maschinenmodelle hat in den letzten 10 Jahren
große Fortschritte gebracht. Mehrfach wurde die Semantische Begründung der
funktionalen Sprachen, nämlich des getypten Lambda-Kalküls mit partieller Rekursion (LCF, Domaintheorie) oder mit total-Rekursiven Funktionen (Theorie
der Wohlfundierten Ordnungen und darauf basierender Rekursion) ihrerseits in
Logiken höherer Stufe entwickelt, und zwar sowohl in einem klassischem [54] als
auch konstruktivistischen Setting [60]. Vor Kurzem gelang sogar die semantische Fundierung von komplexen Typsystemen mit Typkonstruktorklassen durch
Reflektion auf Funktionen über dem reflexiven Bereich D ∼
= A + [D → D][43];
hierdurch können die üblichen Monaden-Kombinatoren in Haskell-artigen Bibliotheken endgültig im Rahmen einer Logik höherer Stufe wie HOL (oder je
nach Geschmack: Cartesian Closed Categories (CCC)) interpretiert werden.
Die folgende Übersicht — eine subjektive Auswahl — listet wesentliche Erfolge des Bottom-up Programms des letzten Jahrzehnts:
1. konkrete, “reale” Maschinenmodelle für Prozessoren (AMD,INTEL) und
abstrakte Maschinen (JVM) werden routinemäßig entwickelt und im industriellen Maßstab verwendet. Neue Anwendungen wie genaue Laufzeitabschätzungen werden absehbar [70],
2. konservative, d.h. nur auf definitorischen Axiomen basierende Modelle für
operationale und axiomatische Semantik für Java-artige Sprachen sind möglich
geworden [46,68],
3. JVM Modelle und darauf basierene, korrekt bewiesene Analysemethoden
wie [46],
4. axiomatische objektorientierte Speichermodelle wie für Spec# [6],
5. konservative objektorientierte Speichermodelle wie für HOL-OCL# [21],
6. axiomatische lineare Speichermodelle wie für C in VCC-I# [51,53,15],
7. axiomatische lineare Speichermodelle wie für verteiltes C in VCC-II# [22],
8. separation logics und ähnliche Abstraktionen über linearem Speicher zur
Formalisierung von Framing Conditions, [66],
9. etc., etc.,
Die Programm-Verifikationswerkzeuge wie Spec#, VCC oder Simpl[62] basieren darauf, das für gegebene operationale Semantiken ein wp-Kalkül abgeleitet
werden kann, der vollautomatisch aus einem mit Zusicherungen und Invarianten
annotierten Programm eine Formel in Logik erster Stufe generiert. Gilt diese,
dann existiert auch ein Beweis in einer Hoare-Logik, der die Gültigkeit der Zusicherungen etabliert (also insbesondere auch Vor- und Nachbedingungen). Da
die Formel unabhängig vom Programm ist, kann sie einfach als Beweisziel eines
Verifikationsverfahrens (siehe Abschnitt 7) behandelt werden. Schränkt man die
Logik geeignet ein (etwa auf Aussagen über Speicherressourcen, wie in SALAnnotations[23]), können im Hinblick auf spezielle Eigenschaften hochautomatisierte Verfahren zur (partiellen) Programmverifikation konstruiert werden.
Die Liste markiert nicht nur eine beträchtliche Dynamik, die durch die Verfügbarkeit von formalen Modelle von “realen” Prozessoren und maschinen-nahen
324
Programmiersprachen in interaktiven Theorembeweisumgebungen ausgelöst wurde, sondern auch die entscheidende Bedeutung für das Bottom-up Programm:
Sie widerlegt die Hauptkritik, welche die Möglichkeit bestreitet, die Semantiken von “real languages” formal zu handhaben und damit eine wissenschaftlich
rationale Fundierung für Werkzeuge und Methoden zu liefern.
Wesentliche andere Forschungslinien der Programmiersprachenforschung sind
natürlich Typsysteme und die Möglichkeit, sie für statische Programmanalysen zu nutzen. Für Top-down lässt sich der Trend erkennen, größere Fragmente aus System F herauszuschneiden und für neuere, mächtigere Typsysteme
in zumeist funktionalen Sprachen einzusetzen (Ausnahme: SCALA); der Preis
sind neuartige, typ-orientierte, vom Benutzer hinzuzufügende Annotierungen
(die stark semantischen Charakter haben können), und pragmatisch schlechter kontrollierbare Typinferenz-Systeme. Der Vorteil sind neue Darstellungen
von Typ-Konstruktorklassen oder, im Falle der Anwendung dieser Typsysteme
in Logiken Höherer Stufe [67], poly-typische Beweistechniken für ProgrammTransformationen im Stile der Squiggol-School (or “Algebra of Programming”)
[12] (d.h. die üblichen “Catamorphismen” und “Hylomorphismen” werden in
diesen Logiken tatsächlich First-Class-Citizens).
Übrigens ist dieser späte Erfolg der Squiggol School, zusammen mit der
Funktor-Semantik für parametrisierte Datentypen (die immerhin inspirierend für
Sprachen wie SML und das Parametrisierungskonzept in Java gewesen ist), einer
der wenigen echten Beiträge des Algebraischen Top-down Programms, was sich
im Wesentlichen durch die Verwendung von Kategorientheorie als Fundamentalsystems auszeichnet. Es scheint sich zu rächen, das in diesem Teilprogramm
niemals ernsthaft versucht worden ist, Beweisumgebungen direkt kategorientheoretisch zu begründen und einen entsprechenden Fundus an formalisierter
Mathematik aufzubauen (von wenigen Ausnahmen, etwa im MIZAR Journal,
abgesehen). Dadurch haben viele neuere Arbeiten lediglich den Charakter, andere bekannte Logiken oder Rechenmodelle kategorientheoretisch nachzuformulieren — der Vergleich mit dem Neo-Marxismus, der primär damit beschäftigt
ist, Erklärungen für ’56 und ’89 zu finden, liegt da nahe.
6
Formale Entwicklungsmodelle
Im Folgenden behandeln wir die Frage, wie man mit formalen Techniken von
der abstrakten Spezifikation und seinen formalisierten Anforderungen zu einem
ablauffähigen Programm, einer Implementierung kommen kann, das die Anforderung erfüllt. Im Gegensatz zu Abschnitt 7 geht also hier darum, den Entwicklungsprozess selbst formal zu untermauern und technisch zu unterstützen.
Grundsätzlich kann man drei Ansätze unterscheiden: Programm-Entwicklung
durch Verfeinerung, durch Programmtransformation und durch Model-Driven
Engineering.
325
6.1
Verfeinerung
Jede formalen Spezifikation ist ein mathematische Modell und damit notwendigerweise eine Abstraktion der Realität. Spezifikationen können aber auch Abstraktionen von komplexeren Systemmodellen sein: Modelle auf höherem Abstraktionsgraden sind in der Regel erheblich einfacher zu verstehen oder maschinell zu analysieren als detailliertere.
Die Verfeinerungsbeziehung ist eine formale Relation P % Q zwischen Modellen, die ausdrückt, ob ein Modell Q eine korrekte Implementierung des Modells
P ist. Der dieser Beziehung unterliegende Korrektheitsbegriff hängt einerseits
vom verwendeten Formalismus und andererseits von der Art der Anforderungen
ab, die in der Spezifikation beschrieben werden.
Aus dem Bereich Algebraisch-Top-down sind Ansätze bekannt, die zumeist
auf Theoriemorphismen basieren (von klassischem “Forget-Restrict-Identify” [41]
über injektive Theoriemorphismen à la KIDS und seinen Nachfolgesystemen), in
den letzten Jahren aber kaum weiterentwickelt worden sind. Vorherrschend sind
Systeme, die Transitionssysteme über Forward- oder Backwardsimulation (Vergleiche [50] für eine Übersicht über die verschiedensten Simulationsbegriffe) über
eine Abstraktionsrelation R zwischen abstrakten und konkreten Zuständen in
Beziehung setzen; die Verfeinerungsbeziehung wird also als Tripel P %R Q aufgefasst. Solche Transitionssysteme werden üblicherweise durch Vor- und Nachbedingungen einer Operation und dann durch Module oder Packages von Operationen spezifiziert. Ein Verfeinerungssystem kann aus den vom Benutzer gegebenen
Transitionssystemen sowie der Abstraktionsrelation Beweisverpflichtungen generieren, die dann verifiziert werden müssen (siehe Abschnitt 7). Systeme, die dieser Methodik unterliegen, sind z.B. das KIV Tool, das KeY Tool, HOL-Z[18,69],
sowie B-Tools und sein Nachfolger Rodin[1].
CVS client 1
cvs
CVS client 1
cvs
working
copy
login
| add
| update
| commit
CVS-Server
CVS client n
cvs
working
copy
login
| add
| update
| commit
(a) “system”
CVS-Server
CVS client n
cvs
repository
login
| add
| update
| commit
login
| add
| update
| commit
copy, mv, chmod, chown, . . .
filesystem
repository
and
working copy
copy, mv, chmod, chown, . . .
(b) “implementation”
Abbildung 1. Verfeinerungen einer Client-Server Sicherheitsarchitectur
Eine beispielhafte Fallstudie für eine Systementwicklung basierend auf einer
Forwardsimulation ist die CVS-Server Studie basierend auf HOL-Z. Wie in Abbildung 1 gezeigt, verfeinert sie ein ein datenorientiertes (Sicherheits-)Modell
326
einer Datenbank, deren Zugriffskontrolle durch ein role-based access control model (RBAC) modelliert wurde, auf ein konkrete Konfiguration eines Dateisystems, das mit UNIX/Posix typischen Lese-und Schreibrechten attibutiert ist.
Dabei werden die abstrakten Operationen wie login, add, update und check-in
berücksichtigt. “korrekt” ist diese Verfeinerung dann, wenn die Implementierung
eine Verfeinerung des abstrakten Modells ist, sich also so verhält wie sie sich laut
Spezifikation verhalten sollte (vgl. Abschnitt 6.1). Dieses Datenmodell kann in
ein Verhaltensmodell engebettet werden: Ein Anwendungsszenario hierfür ergibt sich, wenn man über dieser lokalen Systemtransition (eine Transaktion der
Datenbank) die Menge aller Systemläufe konstruiert und hierfür verlangt, dass
niemand in einem Systemablauf Daten bekommen kann, für die er sich vorher
nicht für hinreichend stark authentifiziert hat (durch login). Solche temporallogischen Aussagen lassen sich durch Induktion über die Position in den Läufen
zeigen (z.B. [19,10]).
Selbstverständlich gibt es auch eine reichhaltige Literatur für verhaltensorientierte Verfeinerungen; insbesondere für Prozessalgebren wie CSP oder CCS
sind in vielfältige Verfeinerungsbegriffe definiert und Modelchecker wie FDR implementiert worden. Der Ansatz stagniert aber seit einer Weile; zum Einen macht
es sich von der Modellierung her für große Systeme nicht bezahlt, Zustände nur
als Prozesse zu betrachten; das resultierende exponentielle Wachstum der Zustandsräume macht allzu viele praktische Probleme “untractable”, und Kombinationen mit Zustandsabstraktionsrelationen haben sich nicht durchsetzen können. Auch aus diesem Grunde sind Implementierungen von Protokoll-Analysetools, die an sich in natürlicher Weise durch prozessalgebraische Verfeinerungen
ausgedrückt werden können, von prozessalgebraischen Techniken abgerückt und
verwenden zumeist problemspezifisches Modelchecking (siehe z.B. der OFMC
aus AVISPA).
6.2
Transformation
Kernidee des transformationellen Entwicklungsansatzes ist, Spezifikationen oder
Programme syntaktisch so umzuformen, das sich ihre Semantik nicht oder nur
verfeinernd verändert. Transformationen sind oft regelbasiert beschrieben, können
aber im Grundsatz beliebige Programme sein, die die Übersetzung durchführen.
Frühe Top-down, d.h. spezifikations-orientierte Transformationssysteme waren CIP-S, PROSPECTRA, KIDS, sowie in HOL realisierte Implementierungen
von Back und Wrights Refinement Calculus [4,5] oder von KIDS-artiger Standardtransformationen wie GlobalSearch und DivideAndConquer [47]. Diese Forschungsrichtung ist aufgegangen in dem Teilprogramm Model-driven Engineering
der im nachfolgenden Abschnitt diskutiert wird.
Bottom-up orientierte Programmtransformationsansätze haben eine aktive
Entwicklung erfahren, insbesondere im Bereich der Compilerverifikation. Denn
auch wenn man im allgemeinen auf ihre Korrektheit vertraut, so weisen Compiler
doch immer wieder Fehler auf, vor allem wenn komplexe Optimierungsstrategien
angewendet werden, oder wenn insbesondere sehr hardware-naher Code compiliert wird (bei kombiniertem C und Assemblercode, beispielsweise). Die schon
327
erwähnte Verfügbarkeit von konkreten, realistischen Maschinenmodellen und immer leistungsfähigeren Beweisumgebungen hat auch hier entscheidende Durchbrüche gebracht. Dabei können Programme nach einem Divide et Impera-Prinzip
sowohl lokal transformiert als auch lokal verifiziert werden und anschließend ihre
Übersetzungen bzw. Beweisteile zusammen gefügt werden. Verifikationen nach
dieser Methode finden sich z.B. in [29,61]. Verifikationen, die auch optimierende, strukturverändernde Transformationen behandeln, sind in [13] beschrieben.
Zusammenfassungen zur Compiler-Verifikation von sind in [37,36] zu finden.
Eine besondere Problematik von optimierenden Übersetzungstransformationen ergibt sich im Zusammenhang mit modernen, stark nebenläufigen Speichermodellen, z.B. für die JVM, wo lediglich eine partielle Ordnung auf der Ausführung von Load- und Store-operationen durch die Semantik definiert wird. Dadurch ist sowohl die Modellierung der Korrektheit als auch der konkrete Beweis
einer optimierenden Transformation für nebenläufige Java-Programme eine Herausforderung [63].
6.3
Model-Driven Engineering
Model-Driven Engineering (MDE) bezeichnet die systematische Verwendung von
Modellen als primären Entwicklungsartefakten durch den Entwicklungsprozess
hindurch. In seinem sehr liberalen Verständnis bezeichnen “Modelle” hier einfach Beschreibungen in einem maschinengestützten Format (im Grenzfall also: einer abstrakten Syntax), während sich der Terminus “systematisch” auf
maschinengestütze Transformationen zwischen Modellen und von Modellen zu
Code bezieht. Für die Instanz von MDE, die sich auf uml bezieht und die
durch die Object Management Group (omg) definiert wird, hat sich der Begriff
model-driven architecture (mda) etabliert (Vergleiche Abschnitt 4.2). In uml,
unterschiedliche Modellelemente wie Klassen oder Zustände können zu Klassensystemen oder Zustandsmaschinen kombiniert werden. Modellelemente können
mit logischen Constraints der Object Constraint Language (ocl) annotiert werden. uml als Sprache überlässt es bewusst eher Werkzeugbauern, verschiedenen
Modellelementen oder Modelltransformationsprozessen eine formale Semantik zu
unterlegen; für einige Teile der Sprache gibt es keine (Use-Cases), für manche
Modelle eine (Klassenmodelle), für andere Teile mehrere formale Semantiken
(Sequenzdiagramme, Statemachines); für alle gibt es aber einheitliche abstrakte
Syntaxen (“Meta-modelle”), technische Infrastrukturen und Diagramm-Typen
um Aspekte der Modelle graphisch darzustellen und zu kommunizieren — Code
oder Code-Fragmente sind übrigens auch Modellelemente in diesem Paradigma.
Ein wesentlicher Aspekt dieses Ansatzes ist die Möglichkeit, Modelle zu transformieren — AndroMDA oder ArcStyler sind entsprechende generische Rahmenwerke, in denen Modelle gespeichert und die Implementierung von Modelltransformationen auf ihnen unterstützt werden. Industriell relevante Anwendungen
solcher Rahmenwerke erzeugen z.B. verteilte Applikationen für Corba- und EJB
Frameworks, dynamischen Testcode (durch entsprechende Wrapper um CodeFragmente herum), oder Zugriffs-“Watchdogs” um Zugriffskontrolle in einem
verteilten Softwaresystem durchzusetzen.
328
Class
+ Public Method
# Protected Method
− Private Method
Test Data
Proof
Class
attribute: Type
attribute: Type
1..∗
+ Public Method
Role
# Protected Method
− Private Method
Obligations
Validation
Class
attribute: Type
ArgoUML
+ Public Method
# Protected Method
− Private Method
SecureUML/OCL
or
UML/OCL
(XMI)
Model
Transformation
Model−Analysis
and Verification
(HOL−OCL)
Test
Harness
C#
+OCL
AC
Config
Model
Repository
(su4sml)
Code
Generator
...
Abbildung 2. Das HOL-OCL mda Framework
In Abbildung 2 wird das HOL-OCL framework [17] dargestellt: ein UML
Frontend wie ArgoUML wird mit einem Kernsystem verbunden, das auf einer
Einbettung von UML/OCL in Isabelle/HOL basiert, ein in SML implementiertes Repositorium für UML Modelle vorsieht (samt Parsern, Typecheckern
und import-export-Funktionen), sowie einem generischen Code-Generator. Ein
Spezifikum dieses Rahmenwerks ist es, das Beweisverpflichtungen ebenfalls Modellelemente sind und während einer Modell-oder Codetransformation erzeugt
werden können; diese können dann im Isabelle/HOL-OCL Kern mit interaktiven und automatischen Methoden bewiesen werden. Dadurch können ForwardSimulation Verfeinerungen implementiert werden genauso wie Transformationen,
die ein kombiniertes System- und Sicherheitsmodell à la SecureUML in ein angereichertes Standard-Systemmodell umwandeln und dabei Seitenbedingungen
generieren, die die Lifeness des generierten gesicherten Systems etablieren [17].
7
Verifikationstechnik
Es ist zwar grundsätzlich möglich, aus Verfeinerungen oder aus Programmannotierungen generierte Beweisverpflichtungen “von Hand” (etwa durch Inspektionen) zu beweisen. Allerdings stößt die Lösung dieser Aufgabe ohne computer-gestützte Techniken an Grenzen der Praktikabilität und Zuverlässigkeit: Erfahrungen mit VCC-I und seinem sehr feinkörnigen Speichermodell zeigen, das generierte Beweisverpflichtungen ohne weiteres ein Megabyte groß werden können
(für eine C Funktion mit ca. 100 Zeilen länge, gemessen im SMT Austauschformat das zwischen Boogie, Z3 sowie HOL-Boogie[14,15]). Es ist somit fair zu
sagen, daß Verfahren zur Lösung von mathematischen Formeln das Herz des
FM darstellen, und zwar gleichermaßen für Top-down wie für Bottom-up. Es ist
dabei bemerkenswert, das klassische Unterscheidungen wie Safety und Security
aus Verifikationssicht kaum von Bedeutung sind4 .
4
... und übrigens auch auf Modellierungsebene immer weniger Sinn machen: Egal
ob durch einen Protokoll-Fehler oder durch einen Exploit eines Buffer-Overflows
329
Grundsätzlich können drei Verfahren der Verifikation von Systemeigenschaften unterschieden werden: deduktionsbasiertes Theorembeweisen, automatenbasiertes Modelchecking und modell-basiertes Testen — letzteres verwendet, wie
wir noch sehen, nicht nur dieselben Basistechniken wie Theorembeweiser oder
Modelchecker und ähnliche Meta-Annahmen wie Modelchecking, sondern kann
eine “approximierende Verifikation” darstellen.
7.1
Theorembeweisen
Deduktionsorientiertes Theorembeweisen benutzt Kalküle, um Beweisziele (also
z.B. Beweisverpflichtungen aus einer Verfeinerung oder einer Programmverifikation) zu zerlegen. Konzeptionell sind Beweisziele Formeln in Logiken wie z.B.
Aussagenlogik (PL), Logik erster Stufe (FOL), oder Logik höherer Stufe (HOL),
die zumeist mit Hilfstheorien erweitert werden müssen (Arithmetik, Bitvektoren,
aber auch Speichermodellen mit Theorien über Load-Store-Operationen, etc.).
Grundsätzlich kann man interaktive Beweissysteme wie Isabelle/HOL und
automatische wie Vampire oder Z3[25,26] unterscheiden; bei ersteren wird die
Liste der Beweisziele und ihrer Zerlegungen zu einem Beweiszustand zusammengefaßt, der mittels Taktiken immer weiter verfeinert wird zu Endzuständen,
aus denen sich die logische Wahrheit syntaktisch ergibt. Letztere verwandeln
die Formeln in clevere Datenstrukturen (Tableaux, SAT), in denen hocheffizient
implementierte Heuristiken die Beweissuche realisieren. Erstere sind einfach zu
erweitern und haben einen beträchtlichen Fundus an formalisierter Mathematik
in ihren Bibliotheken, verfügten jedoch vor einiger Zeit noch über einen relativ
niedrigen Automatisierungsgrad. Letztere lösen Formeln von z.T. staunenerregender Komplexität in kürzester Zeit — Beweisfehlversuche sind allerdings notorisch schwer zu interpretieren und zu analysieren, und bis vor einiger Zeit war
die Erweiterbarkeit dieser Systeme ausserordentlich problematisch.
Die Situation hat sich in beiden Lagern in den letzten Jahren entscheidend
verbessert: durch die Möglichkeit, Taktiken programmgesteuert anzuwenden,
sind mächtige Beweisprozeduren (Termersetzungsprozeduren, Tableaux-Beweiser,
Arithmetische Entscheidungsprozeduren) in interaktiven Beweisern entwickelt
worden, die zudem konstruktionsbedingt eine außerordentliche hohe Zuverlässigkeit und Vertrauenswürdigkeit erreicht haben (zumindest bei Beweisern, die eine
“LCF-style Architecture” aufweisen wie Isabelle/HOL oder HOL-Light; eine Diskussion findet sich in [39]). Bei den Vollautomaten wie Z3 oder Alt-Ergo hat sich
vorläufig die DPLL(X) - Architektur [34] durchgesetzt, eine generische Erweiterung der SAT-Beweiser mit eingebauten Algorithmen für Kongruenzabschlüsse,
die mit Entscheidungsprozeduren für Arithmetik, Bitvektoren, Quantoreninstantiierungen etc. kombiniert werden kann. Durch geeignete Codierungstechniken
von Hilfsattributen ist zumindest für den Fall der Programmanalyse das Problem der Rückinterpretation von Beweisfehlversuchen erleichtert worden[49,14].
ein Virus Millionen Rechner und damit IT-Infrastukturen eines Landes zum Ausfall
bringt: es besteht in jedem Fall Gefahr für Leib und Leben von Menschen.
330
Zudem wird intensiv an Kombinationen von automatischen und interaktiven Beweistechniken gearbeitet, sei es auf Beweiser-Ebene wie bei zchaff oder Vampire
in Isabelle oder auf der Methoden-Ebene wie bei VCC-I/HOL-Boogie[15].
Interaktive Beweiser haben — wie schon erwähnt — erheblichen Anteil an
der Beherrschung der Grundlagen von “real machine models”, also Prozessoren,
Virtuelle Maschinen oder Ausführungsmodelle von Sprachen wie C gehabt[24];
diese waren für die Durchbrüche des Bottom-up Programmes von wesentlicher
Bedeutung. (Daneben gibt es beachtliche Erfolge wie die Formalisierung großer
mathematischer Beweise wie Vierfarbentheorem, Primzahl-Verteilungstheorem
oder der Kepler-Vermutung). Automatische Beweiser, deren Effektivität um
Grössenordnungen zugenommen hat, sind für den Bau von Programmanalysesystemen wie Krakatoa oder Spec# sowieso kriegsentscheidend.
Interaktives oder automatisches Theorembeweisen wird in der industriellen
Praxis zur Zeit vor allem bei hoch-kritischen Anwendungen im militärischen
Bereich, bei Java-Smartcards, bei Mikroprozessoren oder bei maschinennaher
Software wie Treiber und Betriebssysteme angewandt[57]. Die Firma Intel hat
beispielsweise mehr als 40 % des Pentium IV bis auf das Gatterlevel hinunter
so verifiziert. Prozessoren sind also vermutlich die am weitesten verbreiteten
handelsübliche Produkte, für die Verifikation in der Herstellung bedeutsam ist.
Für die schwierigen Probleme werden hier Theorembeweiser eingesetzt, weil die
im folgenden beschriebenen Modelchecker sie aus Aufwandsgründen praktisch
nicht mehr bewältigen können.
7.2
Modelchecking
Modelchecking (auch: Modellprüfverfahren) versucht zumeist, für Widerlegungsvollständige Logiken (PL, FOL, Monadische Logik 2. Stufe,LTL, CTL,...) für das
negierte Beweisziel ein Modell zu finden; wenn das gelingt, ist das Beweisziel widerlegt. Durch geeignete Datenstrukturen können Modelle vollautomatisch und
z.T. hocheffizient gesucht werden. Das Konzept ist auch auf verhaltensorientierte
Modelle in LTL und CTL, ja sogar auf Prozessalgebren anwendbar (die nichts
mit Widerlegungsvollständigkeit zu tun haben), und läuft dort zumeist auf die
Exploration von kompakt repräsentierten Automaten oder “Labelled Transition Systems” heraus. Für die kompakte Repräsentation dienen zumeist binary
decision diagrams (BDD), die boolesche Funktionen eindeutig (kanonisch) repräsentieren; sie setzen eine Variablenordnung voraus, aufgrund der die Boolesche Funktion mehr oder weniger kompakt repräsentiert werden können.
In etwas mehr Detail läßt sich das Verfahren so beschreiben; Ist das zu modellierende System endlich, kann die Systemtransitionsrelation als aussagenlogische
Formel φ dargestellt werden kann — sämtliche Daten müssen nur binär codiert
werden. Stellt man die Frage, ob φ eine Anforderung ψ erfüllt, bedeutet dies,
dass man das Beweisziel φ → ψ negieren und dafür ein Modell suchen muss; gibt
es keins, gilt das Beweisziel. Das Verfahren ist für Aussagenlogik vollständig,
allerdings oft unpraktikabel (“Zustandsexplosion”). BDD’s erlauben für viele
relevante Boolesche Funktionen (z.B. die Addition auf binär codierten Zahlen)
sehr kompakte Darstellungen, die die Zustandsexplosion umgehen.
331
Verallgemeinert man diese Techniken zur Behandlung von verhaltensorientieren Systemeigenschaften, so kann z.B. die transitive Hülle der Systemtransitionsrelation bilden (für Aussagen der Form: “alle erreichbaren Zustände erfüllen ψ”)
oder aus ihr einen endlichen Automaten erzeugen. In letzterem kann man auch
Aussagen der Form entscheiden: “irgendwann werden auf allen Pfaden Zustände
erreicht, auf denen ψ gilt”. Mit endlichen Automaten können auch nebenläufige
Systeme analysiert werden; hierbei werden Automaten konstruiert, die die Menge
aller möglichen Verzahnungen lokaler Teilsystemtransitionsrelationen enthalten.
Besonders erfolgreich kann Modelchecking für die Programmanalyse nebenläufiger Programme mit trivialem Datenanteil herangezogen werden (SPIN, Java/Pathfinder, ...), oder in der Protokollanalyse herangezogen werden, da man es
hier mit nebenläufigen Protokollabläufen zu tun hat, deren Interaktionen schwer
zu überblicken sind. Es stehen mittlerweile Spezialwerkzeuge zur semi- und vollautomatischen Analyse kryptographischer Protokolle zur Verfügung [3], die erfolgreich zum Aufdecken von Fehlern bzw. Nachweisen der Sicherheit veröffentlichter Protokolle eingesetzt worden sind und zudem den Entwurf und die Standardisierung von Protokollen unterstützen [52].
Während die Erfolge von Modelcheckern in sehr spezialisierten Anwendungsdomänen unbestreitbar sind, muss ihre Verwendbarkeit als allgemeine Modellierungs- und Analysetechnik skeptisch beurteilt werden. Es gibt dafür auch theoretische Gründe: Für BDD’s ist bekannt, das keine Variablenordnung existiert,
die gleichermaßen für Multiplikation und Addition kompaktifiziert; bekannte
Challenge-Probleme wie (x div y) ∗ y + (x mod y) = x können mit aktuellen High-End Modelcheckern auf BDD-basis lediglich für maximal fünfbittige
Zahldarstellungen für x und y bewiesen werden[59] ... Pragmatisch sind die Constraints für die Modellierung realer Probleme derart einschneidend, das auch die
Arbeitseffektivität der Verifikation gering ist, trotz eines automatischen Verifikationsverfahrens (siehe auch [10]). Erfolgreiche und etablierte Verifikationsabteilungen in der Industrie (Intel,AMD,Microsoft) setzen daher auf eine geschickte
Kombination von Testen, Modelchecking und interaktivem Theorembeweisen.
7.3
Testen
Wie beim Modelchecking muss beim systematischen, modell-basierten Testen der
Zustandsraum eines realen Systems sowie die Menge der möglichen Eingabedaten endlich beschränkt werden; während man pragmatisch bei modelcheckingbasierter Verifikation den Bereich der Quantoren einer Systemeigenschaft beschränkt, sucht man beim systematischen Testen andere endliche Teilbereiche
von Eingabedaten, die sich nach Testadäquatheitskriterien richten. Solche Kriterien, die eine Aussage darüber machen wann eine Eigenschaft eines Systems,
also ein Beweisziel, genug getestet worden ist, lassen sich teilweise durch Fehlermodelle von Schaltwerken, Programmen oder Spezifikationen begründen. Solche
Adäquatheitskriterien können spezifikationsbasiert sein (z.B. Partition-Coverage;
ein Test für jedes Glied der DNF der Spezifikation) oder programmbasiert sein
(z.B. Path-Coverage; ein Test für jeden Pfad durch den Kontrollflussgraphen
332
eines Programms) und sich auf das Systemverhalten beziehen (z.B. TransitionCoverage der Spezifikation; alle Transitionen der Spezifikation in Form eines LTS
müssen in einer Testsequenz vorkommen).
Vollständige, z.B. mit einem Theorembeweiser durchgeführte Verifikation ist
so gesehen kein Gegensatz zum Testen. Systematische, modellbasierte Tests verwenden nicht nur dieselben Basistechniken wie Theorembeweiser oder Modelchecker, sondern können eine “approximierende Verifikation” darstellen [20]; in
dem dort beschriebenen System HOL-TestGen wird eine Testspezifikation T S
automatisch in ein Testtheorem der Form:
Case1 =⇒ . . . =⇒ Casen =⇒ T HY P (H1 ) =⇒ . . . =⇒ T HY P (Hm ) =⇒ T S
äquivalent zerlegt, wobei die Casei (die Testfälle) der Form
Constraint1 x =⇒ . . . Constraintk x =⇒ post x (sut x)
sind. Constraint-Solving-Techniken erlauben nun die Konstruktion von Zeugen
für x, die die Constraints erfüllen. Aus post x (sut x) generiert man dann ein
Testorakel, das das System-Unter-Test sut mit diesem Zeugen aufruft und die
Zulässigkeit seines Outputs mittels des Codes für post prüft. Der Aspekt, das
gegebener Code gegenüber einem Modell abgeprüft wird und das man mit dem
Modell solange experimentieren kann, bis man durch das Modell den Code “erfaßt” hat, ist ein Spezifikum des Testansatzes und macht Modell-basiertes Testen
für das Reverse-Engineering unverzichtbar, auch im industriellen Maßstab [38].
HOL-TestGen ist für sehr ambitionierte Fallstudien angewendet worden;
diese nutzen auch die Möglichkeit, Scenarien für Sequenztest und reaktivem
Sequenztest in diesem Rahmenwerk darzustellen [16].
Die T HY P (Hi ) im obigen Testtheorem sind übrigens die explizite Testhypothesen, d.h. Formeln, die die genaue “logische Differenz” zwischen Test und
Verifikation darstellen. Das obige Testtheorem bedeutet demnach, daß wenn sut
den Test für alle Testfälle besteht und wenn die Testhypothesen gelten, dann
erfüllt sut genau die Testspezifikation. Klassische Testhypothesen [35] sind z.B.
Uniformität (wenn der Test für einen Testdatum der Klasse gelingt, gelingt er
für alle der Klasse) oder Regularität (es werden alle Daten bis zu einer gewissen
Komplexitätsstufe getestet).
Erschöpfendes Testen aller möglichen Eingabesequenzen ist selbst im Hinblick auf die üblichen Adäquatheitskriterien wegen der exponentiell wachsenden
Anzahl der Testabläufe in der Praxis oft nicht durchführbar. Wünschenswert
sind somit Adäquatheitskriterien, die bei einem gegebenen Aufwand zu einer
Menge von Testabläufen führen, die mit hoher Wahrscheinlichkeit sicherheitskritische Fehler aufdecken, und somit ein größtmögliches Vertrauen in die Sicherheit der Implementierung liefern. Dies führt zur Idee der probabilistischen
Adäquatheitskriterien, (z.B. probabilistische Uniformität[27]), bei der für jede
Testklasse verlangt wird, daß sie mit der gleichen Wahrscheinlichkeit wie alle
anderen getestet werden.
333
8
Schlussfolgerungen
Man kann einen Trend zu einem dichteren Netz von Anwendungsszenarien für
FM beobachten, das immer weitere Bereiche der Softwaretechnik durchzieht.
Dies gilt für Top-down, aber z. Zt. in scheinbar stärkerem Maße für Bottomup. Es fragt sich allerdings, ob sich die starke Dynamik, die sich insbesondere
auf Durchbrüche bei den vollautomatischen Beweissystemen wie Z3 stützt, nicht
abschwächt und abstraktere Modellierungstechniken bald wieder stärker im Vordergrund stehen.
Neben dem prinzipiellen Nutzen von FM, der in einer weitergehenden wissenschaftlichen Fundierung der Softwaretechnik als solcher besteht—sei es in der
Bedarfsanalyse, dem Entwurf, der Implementierung oder der Verifikation—wird
auch ein immer größerer praktischer Nutzen erkennbar. Drei praktisch relevante
Anwendungsszenarien lassen sich identifizieren:
– FM “under-the-hood”. In immer mehr Entwicklungswerkzeugen und Compilern werden intern Formale Methoden Techniken wie Abstrakte Interpretation, Modelchecking, Extended Typechecking und Theorembeweisen verwendet, ohne dass der Programmentwickler das direkt sieht (Java BytecodeVerifier, PolySpace, . . . ).
– FM “light-weighed”. Dies bezeichnet eine Reihe von Migrationsversuchen
von Formalen Methoden wie Runtime-Testing von Annotierungssprachen wie
OCL, JML oder Spec#, oder wie hochautomatische Verfahren basierend auf
eingeschränkten Annotierungen à la SAL oder modelcheck-fähige temporale Constraints aus Bibliotheken. FM “light-weighed” umfasst auch Techniken wie MDA (auch spezialisiert für gewisse Zwecke wie die systematische
Einführung von Sicherheitsmechanismen) oder Reverse-Engineering durch
Modell-basiertes Testen.
– FM “hard-core”. Damit werden theorembeweiser-basierten Ansätze wie
Verfeinerung und Codeverifikation zusammengefasst, die einen speziell ausgebildeten Verifikationsingenieur erfordern und aus diesem Grund in der
industriellen Praxis zur Zeit auf hardware-nahe, sicherheitskritische oder
höchste Zertifizierungsstufen erfordernde Systeme beschränkt ist.
Es sei hierbei explizit festgestellt, dass alle drei Bereiche sich gegenseitig bedingen; Techniken, die in FM “hard-core” entwickelt werden, finden mitunter ihren
Weg in FM “under-the-hood”-Szenarien. Umgekehrt ist ein höherer Kenntnisstand und die Verbreitung von FM “light-weighed”-Techniken in der industriellen Praxis für Projekte, die FM “hard-core” etwa aufgrund einer angestrebten
Zertifizierung anstreben, von Vorteil: Von Softwareentwicklern erstellte Modelle
etwa in UML/OCL können eine wertvolle Grundlage für ein verfeinertes und von
Verifikationsingenieuren erstelltes Systemmodell sein.
Danksagung: Ganz herzlich möchte ich mich bei Bernd Krieg-Brückner bedanken, der mich mit FM vertraut gemacht hat, mir viel Raum gab, mich zu
entwickeln und mich ein Stück meines Weges begleitet hat.
334
Literatur
1. Jean-Raymond Abrial. Modeling in Event-B: System and Software Engineering.
Cambridge University Press, 2009. to appear.
2. Nancy J. Adler and Anne-Wil Harzing. When Knowledge Wins: Transcending the
Sense and Nonsense of Academic Rankings. Learning & Education, 8(1), 2009.
3. A. Armando, D. Basin, M. Bouallagui, Y. Chevalier, L. Compagna, S. Mödersheim,
M. Rusinowitch, M. Turuani, L. Vigano, and L. Vigneron. The AVISS Security
Protocol Analysis Tool. In CAV 2002, LNCS 2404, pages 349–353, 2002.
4. R.-J. Back and J. von Wright. Refinement Calculus. 1998.
5. Ralph-Johan Back, Jim Grundy, and Joakim von Wright. Structured calculational
proof. Formal Asp. Comput., 9(5-6):469–483, 1997.
6. Michael Barnett, Robert DeLine, Manuel Fähndrich, K. Rustan M. Leino, and
Wolfram Schulte. Verification of object-oriented programs with invariants. Journal
of Object Technology, 3(6):27–56, 2004.
7. Mike Barnett, Manuel Fähndrich, K. Rustan M. Leino, Francesco Logozzo, Peter
Müller, Wolfram Schulte, Herman Venter, and Songtao Xia. Spec#. Microsoft
Research, Redmond, 2008. http://research.microsoft.com/specsharp.
8. Mike Barnett, K. Rustan M. Leino, Michal Moskal, and Philipp Rümmer. Boogie
program verification. Microsoft Research, Redmond, 2008. http://research.
microsoft.com/boogie/.
9. D. Basin, J. Doser, and T. Lodderstedt. Model driven security: from UML models
to access control infrastructures. ACM Transactions on Software Engineering and
Methodology, 2005.
10. David Basin, Hironobu Kuruma, Kunihiko Miyazaki, Kazuo Takaragi, and Burkhart Wolff. Verifying a signature architecture: a comparative case study. Formal
Aspects of Computing, 19(1):63–91, March 2007. http://www.springerlink.com/
content/u368650p18557674/?p=8851693f5ba14a3fb9d493dae37783b8&pi=0.
11. Patrick Baudin, Jean-Christophe Filliatre, Claude Marché, Benjamin Monate, Yannick Moy, and Virgile Prevosto. ACSL: ANSI/ISO C Specification Language –
version 1.3. Technical report, INRIA Saclay, 22. Oct, 2008.
12. Bird, Moore, Backhouse, Gibbons, et al. Algebra of programming research group.
Home Page. http://www.comlab.ox.ac.uk/research/pdt/ap/pubs.html.
13. J. Blech, S. Glesner, J. Leitner, and S. M¸lling. Optimizing Code Generation
from SSA Form: A Comparison Between Two Formal Correctness Proofs in Isabelle/HOL. In ETAPS 2005.
14. Sascha Böhme, Rustan Leino, and Burkhart Wolff. HOL-Boogie — an Interactive
Prover for the Boogie Program Verifier. In Sofiene Tahar, Otmane Ait Mohamed,
and César Muñoz, editors, 21th International Conference on Theorem proving in
Higher-Order Logics (TPHOLs 2008), LNCS 5170. Springer-Verlag, Montreal, Canada, 2008.
15. Sascha Böhme, Michal Moskal, Wolfram Schulte, and Burkhart Wolff. HOL-Boogie
— an Interactive Prover-Backend for the Verified C Compiler. Journal of Automated Reasoning (JAR), 2009. submitted.
16. Achim D. Brucker, Lukas Brügger, and Burkhart Wolff. Model-based firewall conformance testing. In Kenji Suzuki and Teruo Higashino, editors, Testcom/FATES
2008, LNCS 5047, pages 103–118. Springer-Verlag, Tokyo, Japan, 2008.
17. Achim D. Brucker, Jürgen Doser, and Burkhart Wolff. An MDA framework supporting OCL. Electronic Communications of the EASST, 5, 2007.
335
18. Achim D. Brucker, Frank Rittinger, and Burkhart Wolff. HOL-Z 2.0: A proof
environment for z-specifications. Journal of Universal Computer Science, 9(2):152–
172, February 2003.
19. Achim D. Brucker and Burkhart Wolff. A verification approach for applied system
security. International Journal on Software Tools for Technology Transfer (STTT),
7(3):233–247, 2005.
20. Achim D. Brucker and Burkhart Wolff. Test-sequence generation with hol-testgen
– with an application to firewall testing. In Bertrand Meyer and Yuri Gurevich,
editors, TAP 2007: Tests And Proofs, number 4454 in Lecture Notes in Computer
Science, pages 149–168. Springer-Verlag, Zurich, 2007.
21. Achim D. Brucker and Burkhart Wolff. An extensible encoding of object-oriented
data models in hol with an application to IMP++. Journal of Automated Reasoning (JAR), Selected Papers of the AVOCS-VERIFY Workshop 2006, 2008. Serge
Autexier, Heiko Mantel, Stephan Merz, and Tobias Nipkow (eds).
22. Ernie Cohen, Michal Moskal, Wolfram Schulte, and Stephan Tobies. A Precise
Yet Efficient Memory Model For C. In Proceedings of the European Conference of
Object-Oriented Programing-Languages (ECOOP ’09), Lecture Notes in Computer
Science. Springer-Verlag, July 2009. submitted paper.
23. Manuvir Das. Formal specifications on industrial-strength code-from myth to reality. In Thomas Ball and Robert B. Jones, editors, CAV, volume 4144 of Lecture
Notes in Computer Science, page 1. Springer, 2006.
24. Matthias Daum, Jan Dörrenbächer, and Burkhart Wolff. Proving fairness and
implementation correctness of a microkernel scheduler. Journal of Automated Reasoning (JAR), page 29 pages, 2009. Accepted (with minor revisions) by G. Klein,
R. Huuck and B. Schlich: Special Issue on Operating System Verification (2008).
To appear.
25. Leonardo de Moura and Nikolaj Bjørner. Efficient incremental E-matching for
SMT solvers. In 21st Conference on Automated Deduction (CADE 2007), LNCS
4603, 2007.
26. Leonardo de Moura and Nikolaj Bjørner. Z3: An efficient SMT solver. In TACAS
2008, volume 4963 of Lecture Notes in Computer Science, pages 337–340. Springer,
March–April 2008.
27. A. Denise, M.-C. Gaudel, S.-D. Gouraud, R. Lassaigne, and S. Peyronnet. Uniform random sampling of traces in very large models. In 1st International ACM
Workshop on Random Testing, pages 10 –19, July 2006.
28. Commission d’Evaluation 2007. Une évaluation des indicateurs bibliométriques.
Technical report, INRIA, 2007. http://www.inria.fr/inria/organigramme/
documents/ce_indicateurs.pdf.
29. A. Dold, F. W. von Henke, and W. Goerigk. A Completely Verified Realistic
Bootstrap Compiler. International Journal of Foundations of Computer Science,
14(4):659–680, 2003.
30. Paul Feyerabend. Wider den Methodenzwang. Suhrkamp (stw 597), Frankfurt am
Main, 1975. ISBN 3-518-28197-6.
31. Jean-Christophe Filliâtre and Claude Marché. The why/krakatoa/caduceus platform for deductive program verification. In Werner Damm and Holger Hermanns,
editors, CAV, volume 4590 of Lecture Notes in Computer Science, pages 173–177.
Springer, 2007.
32. Joint Taskforce for Computing Curricula. Computing curricula 2005: The overview
report. Technical report, ACM/AIS/ IEEE, 2005, 2005. www.acm.org/education/
curric_vols/CC2005-March06Final.pdf.
336
33. Erich Gamma, Richard Helm, Ralph Johnson, and John M. Vlissides. Design
Patterns, Elements of Reusable Object-Oriented Software. Addison-Wesley International, 1994.
34. Harald Ganzinger, George Hagen, Robert Nieuwenhuis, Albert Oliveras, and Cesare Tinelli. Dpll( t): Fast decision procedures. In Rajeev Alur and Doron Peled,
editors, CAV, volume 3114 of Lecture Notes in Computer Science, pages 175–188.
Springer, 2004.
35. Marie-Claude Gaudel. Testing can be formal, too. In Peter D. Mosses, Mogens
Nielsen, and Michael I. Schwartzbach, editors, TAPSOFT 95, volume 915 of Lecture
Notes in Computer Science, pages 82–96. Springer-Verlag, Aarhus, Denmark, 1995.
36. S. Glesner and J. Blech. Logische und softwaretechnische herausforderungen bei
der verifikation optimierender compiler. In SE 2005. LNI, März 2005.
37. S. Glesner, G. Goos, and W. Zimmermann. Verifix: Konstruktion und Architektur verifizierender Übersetzer (Verifix: Construction and Architecture of Verifying
Compilers). it - Information Technology, 46:265–276, 2004. Print ISSN: 1611-2776.
38. Wolfgang Grieskamp, Nicolas Kicillof, Dave MacDonald, Alok Nandan, Keith Stobie, and Fred L. Wurden. Model-based quality assurance of windows protocol
documentation. In ICST, pages 502–506, 2008.
39. Thomas C. Hales. Formal proof. Notices of the American Mathematical Society,
55(11):1370–1382, 2008.
40. John Harrison. Formal proof-theory and practice. Notices of the American Mathematical Society, 55(11):1305–1407, 2008.
41. Rolf Hennicker. Context induction: A proof principle for behavioural abstractions
and algebraic implementations. Formal Asp. Comput., 3(4):326–345, 1991.
42. Andrew Hodges. Alan Turing: the Enigma. London: Burnett; New York: Simon
and Schuster, 1983.
43. Brian Huffman, John Matthews, and Peter White. Axiomatic constructor classes
in isabelle/holcf. In Joe Hurd and Thomas F. Melham, editors, TPHOLs, volume
3603 of Lecture Notes in Computer Science, pages 147–162. Springer, 2005.
44. J. J¸rjens. Secure Systems Development with UML. 2004. ISBN: 3-540-00701-6.
45. Joseph Kiniry. ESC/Java2 Home Page, 2005. http://kind.ucd.ie/products/
opensource/ESCJava2.
46. Gerwin Klein and Tobias Nipkow. Verified bytecode verifiers. Theor. Comput. Sci.,
3(298):583–626, 2003.
47. Kolyang, T. Santen, and B. Wolff. Correct and user-friendly implementation of
transformation systems. In M.-C. Gaudel and J. Woodcock, editors, FME 96 —
Industrial Benefits and Advances in Formal Methods, LNCS 1051, pages 629–648.
Springer Verlag, 1996.
48. Ralf Küsters. ETH Zürich, personal communication, 2006.
49. K. Rustan M. Leino, Todd D. Millstein, and James B. Saxe. Generating error
traces from verification-condition counterexamples. Sci. Comput. Program., 55(13):209–226, 2005.
50. Nancy A. Lynch and Frits W. Vaandrager. Forward and backward simulations: I.
untimed systems. Inf. Comput., 121(2):214–233, 1995.
51. Stefan Maus, Michal Moskal, and Wolfram Schulte. Vx86: x86 Assembler Simulated
in C Powered by Automated Theorem Proving. In AMAST, pages 284–298, 2008.
52. C. Meadows, P. Syverson, and I. Cervesato. Formalizing GDOI group key management requirements in NPATRL. In Proceedings of the 8th ACM Computer and
Communications Security Conference (CCS 2001), pages 235–244. ACM Press,
2001.
337
53. Michal Moskal, Wolfram Schulte, and Herman Venter. Bits, words, and types:
Memory models for a verifying c compiler. Science of Computer Programming
(SCP), 2008. to appear.
54. Olaf Müller, Tobias Nipkow, David von Oheimb, and Oscar Slotosch.
Holcf=hol+lcf. J. Funct. Program., 9(2):191–223, 1999.
55. Special issue on formal proof, 2008. http://www.ams.org/notices/200811/.
56. Proof by computer: Harnessing the power of computers to verify mathematical
proofs. http://www.physorg.com/news145200777.html.
57. Wolfgang Paul, Thomas Santen, and Stefan Tobies. Verifying 50000 lines of code.
Futures — Microsoft’s European Innovation Magazine, pages 42–43, June 2008.
58. Karl Popper. Die Logik der Forschung. Verlags Mohr Siebeck, 1934.
59. Nicole Rauch and Burkhart Wolff. Formalizing Java’s two’s-complement integral
type in Isabelle/Hol. In Electronic Notes in Theoretical Computer Science, volume 80. Elsevier Science Publishers, 2003.
60. Bernhard Reus and Thomas Streicher. General synthetic domain theory - a logical
approach. Mathematical Structures in Computer Science, 9(2), 1999.
61. G. Schellhorn and W. Ahrendt. Reasoning about Abstract State Machines: The
WAM Case Study. Journal of Universal Computer Science, 3(4):377–413, 1997.
62. Norbert Schirmer. Verification of Sequential Imperative Programs in Isabelle/ hol.
PhD thesis, Technische Universität München, 2006.
63. Jaroslav Sevcı́k and David Aspinall. On validity of program transformations in the
Java memory model. In Jan Vitek, editor, ECOOP, volume 5142 of Lecture Notes
in Computer Science, pages 27–51. Springer, 2008.
64. Imre Lakatos (Übersetzung Arpád Szabó). Die Methodologie der wissenschaftlichen Forschungsprogramme. Vieweg Braunschweig; Wiesbaden, 1982. ISBN 3528-08429-4.
65. H. Tej and B. Wolff. A corrected failure-divergence model for CSP in Isabelle/HOL.
In J. Fitzgerald, C.B. Jones, and P. Lucas, editors, Proceedings of the FME 97 —
Industrial Applications and Strengthened Foundations of Formal Methods, LNCS
1313, pages 318–337. Springer Verlag, 1997.
66. Harvey Tuch, Gerwin Klein, and Michael Norrish. Types, bytes, and separation
logic. In Martin Hofmann and Matthias Felleisen, editors, POPL, pages 97–108.
ACM, 2007.
67. Norbert Völker. HOL2P - a system of classical higher order logic with second order
polymorphism. In Klaus Schneider and Jens Brandt, editors, TPHOLs, volume
4732 of Lecture Notes in Computer Science, pages 334–351. Springer, 2007.
68. David von Oheimb and Tobias Nipkow. Machine-Checking the Java Specification:
Proving Type-Safety. In Jim Alves-Foss, editor, Formal Syntax and Semantics of
Java, volume 1523 of Lecture Notes in Computer Science, pages 119–156. Springer,
1999.
69. Makarius Wenzel and Burkhart Wolff. Building formal method tools in the isabelle/isar framework. In Klaus Schneider and Jens Brandt, editors, TPHOLs 2007,
LNCS 4732, pages 351–366. Springer-Verlag, 2007.
70. Reinhard Wilhelm, Jakob Engblom, Andreas Ermedahl, Niklas Holsti, Stephan
Thesing, David B. Whalley, Guillem Bernat, Christian Ferdinand, Reinhold Heckmann, Tulika Mitra, Frank Mueller, Isabelle Puaut, Peter P. Puschner, Jan Staschulat, and Per Stenström. The worst-case execution-time problem - overview of
methods and survey of tools. ACM Trans. Embedded Comput. Syst., 7(3), 2008.
338
Burkhart Wolff came in 1991 into Bernd-Krieg
Brückners group and worked on various approaches to construct a successor of the ProSpecTraTransformation system developed earlier there. In
1997, he received his Phd and moved to Freiburg,
where he held a research assistent position till 2003
(for which he eventually got his Habilitation degree in
2005). From 2004 till 2006, he was assistant professor (Öberassistent”) in the computer security group
at the ETH Z¸rich, and was invited researcher at Microsoft Research, Redmond, in 2007. Since 2008, he is
full professor at the Universié Paris-Sud in the formal
testing and software-engineering group ForTesSE. His
main work area is the field of environments for verification and validation of software systems.
Author Index
Autexier, Serge
35
Bibel, Wolfgang 15
Birbach, Oliver 88, 103
Catesbeiana, Erwin R. 40
Catesbeiana, Erwin R. jr. 17
Codescu, Mihai 51
de la Cruz Ramos, Pedro
Dietrich, Dominik 35
Dovland, Johan 72
Drechsler Rolf 140
28
Fey, Görschwin 140
Frese, Udo 88, 103, 255
Gersdorf, Bernd 1, 255
Gimblett, Andy 121
Gogolla, Martin 18
Große, Daniel 140
Hoffmann, Berthold
Hutter, Dieter 35
Ibraheem, Modal
1, 152
19
Johnsen, Einar Broch
72
Kahrs, Stefan 168
Kahsai, Temesghen 121
Kohlhase, Andrea 184
Kohlhase, Michael 184, 201
Kreowski, Hans-Jörg 217
Kurlbaum, Jörg 103
Lankenau, Axel 255
Laue, Tim 103
Lemburg, Johannes 201
Li, Lixing 273
Li, Xiaoyu 273
Lu, Ruqian 273
Maeder, Christian 51
Mandel, Christian 255
Mossakowski, Till 1, 51
Mosses, Peter 23
Nake, Frieder
24
O’Reilly, Liam 121
Owe, Olaf 72
Peleska, Jan
Qian, Zhenyu
229
26
Röfer, Thomas 1, 103, 255
Rauer, Sylvie 29
Reichel, Horst 246
Roggenbach, Markus 121
Sannella, Don 30
Schlingloff, Holger 291
Schröder, Lutz 201
Schulz, Ewaryst 201
Shang, Yun 273
Shi, Hui 303
Steffen, Martin 72
Tarlecki, Andrzej
Lüth, Christoph 20
Lüttich, Klaus 22
Langenstein, Bruno 51
32
Werner, Mattias 13, 34
Wolff, Burkhart 314